{"id":195,"date":"2018-08-01T23:26:53","date_gmt":"2018-08-01T22:26:53","guid":{"rendered":"https:\/\/teatronika.org\/?p=195"},"modified":"2018-08-05T10:08:10","modified_gmt":"2018-08-05T09:08:10","slug":"inteligencia-artificial-y-juegos-de-ordenador","status":"publish","type":"post","link":"https:\/\/teatronika.org\/en\/2018\/08\/01\/inteligencia-artificial-y-juegos-de-ordenador\/","title":{"rendered":"Artificial Intelligence and Computer Games"},"content":{"rendered":"<p class=\"p1\">Halfway between synthetic robotic environments, agent simulations and reality we have computer games: a good testbed scenario for artificial intelligences.<\/p>\n<p class=\"p1\">OpenAI is preparing the program that will control up to 5 agents against 5 human professional players in the complex game Dota2, a game in which the objective is to destroy and conquer the opponent team\u2019s castle (<a href=\"https:\/\/blog.openai.com\/openai-five-benchmark\/\">link<\/a>). OpenAI bot has already proven superiority against human players on the 1 versus 1 Dota2 with restricted play (<a href=\"https:\/\/blog.openai.com\/more-on-dota-2\/\">link<\/a>). The addressed multiplayer scenario is non-trivial and this is because when learning from inputs that include other\u2019s agent behaviours we cannot guarantee the convergence of learning algorithms having to deal with non-stationary environments.<span class=\"Apple-converted-space\">\u00a0 \u00a0<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-208 size-full\" src=\"https:\/\/teatronika.org\/wp-content\/uploads\/2018\/08\/AI-Games-Examples.jpg\" alt=\"\" width=\"1756\" height=\"1182\" srcset=\"https:\/\/teatronika.org\/wp-content\/uploads\/2018\/08\/AI-Games-Examples.jpg 1756w, https:\/\/teatronika.org\/wp-content\/uploads\/2018\/08\/AI-Games-Examples-300x202.jpg 300w, https:\/\/teatronika.org\/wp-content\/uploads\/2018\/08\/AI-Games-Examples-768x517.jpg 768w, https:\/\/teatronika.org\/wp-content\/uploads\/2018\/08\/AI-Games-Examples-1024x689.jpg 1024w\" sizes=\"auto, (max-width: 767px) 89vw, (max-width: 1000px) 54vw, (max-width: 1071px) 543px, 580px\" \/><\/p>\n<p><!--more--><\/p>\n<p class=\"p1\">Deep Mind has also addressed the multiplayer game in a recently published a paper (<a href=\"https:\/\/arxiv.org\/abs\/1807.01281\">link<\/a>) in which a program reaches superhuman level play in the game \u201c<span class=\"s1\">Quake III Arena Capture the Flag\u201d a 2 vs. 2 scenario in which teams have to capture the opponent\u2019s flag and bring it back to the base.<\/span><\/p>\n<p class=\"p1\">It is amazing to see how in the 90s a Deep Learning Neural Network was used to learn Backgammon using Reinforcement Learning TD-Gammon (<a href=\"https:\/\/en.wikipedia.org\/wiki\/TD-Gammon\">link<\/a>). 25 years after, classic Atari games from the 80s were addressed with success by Google Deep Mind with its Deep Reinforcement Learning (DQN) algorithm achieving human and superhuman performance in 60% of the game set (<a href=\"https:\/\/www.nature.com\/articles\/nature14236\">link<\/a>). In 15% of the games, the algorithm DQN had serious difficulties mainly because of the involvement of long-term planning, like in the game Montezuma Revenge where DQN achieved 0 points. The main reason for this failure is that distant rewards cannot be achieved by the initial random exploration of Reinforcement Learning algorithms. Recently OpenAI published very good results achieved in Montezuma by learning from human play (see <a href=\"https:\/\/blog.openai.com\/learning-montezumas-revenge-from-a-single-demonstration\/\">link<\/a>).<\/p>\n<p class=\"p1\">Curiously, the non-deterministic version of PacMan (Atari\u00a0Ms Pac Man, the one in which phantoms take a random action when close to bifurcations) has only been solved recently by an approach that decomposes the reward function into different components (see <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/hybrid-reward-architecture-achieving-super-human-ms-pac-man-performance\/\">link<\/a>).<\/p>\n<p class=\"p1\">Deep Mind also has proven superiority of a program, Alpha Go and Alpha Go Zero (learning from scratch without using human gameplay, see <a href=\"http:\/\/www.nature.com\/articles\/nature24270\">link<\/a>) in the ancient game of GO. Alpha Go Zero also learned to play chess better than humans. It&#8217;s interesting to analyze how Alpha Go Zero uses different openings during the course of learning, proving that sometimes the human culture assigns value to openings in a human biased way.<\/p>\n<p class=\"p1\">The scientific achievements of these algorithms are in debate (for a discussion in the field of GO see <a href=\"https:\/\/medium.com\/@josecamachocollados\/is-alphazero-really-a-scientific-breakthrough-in-ai-bf66ae1c84f2\">link<\/a>) and we still need to know the real contributions that they can bring to society. One of the main reasons of this debate is that the computations needed to achieve them increase exponentially (see the debate in OpenAI article <a href=\"https:\/\/blog.openai.com\/ai-and-compute\/\">link<\/a>).<\/p>\n<p class=\"p1\">For a historical review see the Wikipedia page <a href=\"https:\/\/en.wikipedia.org\/wiki\/Artificial_intelligence_in_video_games\">link<\/a> and the book Artificial Intelligence and Games (2018, <a href=\"http:\/\/gameaibook.org\/\">link<\/a>).<\/p>\n<p class=\"p1\">\n","protected":false},"excerpt":{"rendered":"<p>Halfway between synthetic robotic environments, agent simulations and reality we have computer games: a good testbed scenario for artificial intelligences. OpenAI is preparing the program that will control up to 5 agents against 5 human professional players in the complex game Dota2, a game in which the objective is to destroy and conquer the opponent &hellip; <\/p>\n<p class=\"link-more\"><a href=\"https:\/\/teatronika.org\/en\/2018\/08\/01\/inteligencia-artificial-y-juegos-de-ordenador\/\" class=\"more-link\">Continue reading<span class=\"screen-reader-text\"> &#8220;Artificial Intelligence and Computer Games&#8221;<\/span><\/a><\/p>\n","protected":false},"author":9,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[1],"tags":[],"class_list":["post-195","post","type-post","status-publish","format-standard","hentry","category-sin-categoria"],"acf":[],"_links":{"self":[{"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/posts\/195","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/users\/9"}],"replies":[{"embeddable":true,"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/comments?post=195"}],"version-history":[{"count":9,"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/posts\/195\/revisions"}],"predecessor-version":[{"id":210,"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/posts\/195\/revisions\/210"}],"wp:attachment":[{"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/media?parent=195"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/categories?post=195"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/teatronika.org\/en\/wp-json\/wp\/v2\/tags?post=195"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}