HOME





AlphaZero
AlphaZero is a computer program developed by artificial intelligence research company DeepMind to master the games of chess, shogi and Go (game), go. This algorithm uses an approach similar to AlphaGo Zero. On December 5, 2017, the DeepMind team released a preprint paper introducing AlphaZero, which would soon play three games by defeating world-champion chess engines Stockfish (chess), Stockfish, Elmo (shogi engine), Elmo, and the three-day version of AlphaGo Zero. In each case it made use of custom tensor processing units (TPUs) that the Google programs were optimized to use. AlphaZero was trained solely via Self-play (reinforcement learning technique), self-play using 5,000 first-generation TPUs to generate the games and 64 second-generation TPUs to train the neural networks, all in parallel computing, parallel, with no access to Chess opening book (computers), opening books or Endgame tablebase, endgame tables. After four hours of training, DeepMind estimated AlphaZero wa ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Stockfish (chess)
Stockfish is a free and open-source chess engine, available for various desktop and mobile platforms. It can be used in Computer chess, chess software through the Universal Chess Interface. Stockfish has been one of the strongest chess engines in the world for several years; it has won all main events of the Top Chess Engine Championship (TCEC) and the Chess.com Computer Chess Championship (CCC) since 2020 and, , is the strongest CPU chess engine in the world with an estimated Elo rating system, Elo rating of 3644, in a time control of 40/15 (15 minutes to make 40 moves), according to CCRL. The Stockfish engine was developed by Tord Romstad, Marco Costalba, and Joona Kiiski, and was derived from Glaurung, an open-source engine by Tord Romstad released in 2004. It is now being developed and maintained by the Stockfish community. Stockfish historically used only a classical Evaluation function#Handcrafted evaluation functions, hand-crafted function to evaluate board positions, b ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


AlphaGo Zero
AlphaGo Zero is a version of DeepMind's Go software AlphaGo. AlphaGo's team published an article in ''Nature'' in October 2017 introducing AlphaGo Zero, a version created without using data from human games, and stronger than any previous version. By playing games against itself, AlphaGo Zero: surpassed the strength of AlphaGo Lee in three days by winning 100 games to 0; reached the level of AlphaGo Master in 21 days; and exceeded all previous versions in 40 days. Training artificial intelligence (AI) without datasets derived from human experts has significant implications for the development of AI with superhuman skills, as expert data is "often expensive, unreliable, or simply unavailable." Demis Hassabis, the co-founder and CEO of DeepMind, said that AlphaGo Zero was so powerful because it was "no longer constrained by the limits of human knowledge". Furthermore, AlphaGo Zero performed better than standard deep reinforcement learning models (such as Deep Q-Network impleme ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


MuZero
MuZero is a computer program developed by artificial intelligence research company DeepMind to master games without knowing their rules. Its release in 2019 included benchmarks of its performance in go, chess, shogi, and a standard suite of Atari games. The algorithm uses an approach similar to AlphaZero. It matched AlphaZero's performance in chess and shogi, improved on its performance in Go (setting a new world record), and improved on the state of the art in mastering a suite of 57 Atari games (the Arcade Learning Environment), a visually-complex domain. MuZero was trained via self-play, with no access to rules, opening books, or endgame tablebases. The trained algorithm used the same convolutional and residual architecture as AlphaZero, but with 20 percent fewer computation steps per node in the search tree. MuZero’s capacity to plan and learn effectively without explicit rules makes it a groundbreaking achievement in reinforcement learning and AI, pushing the boundarie ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


DeepMind
DeepMind Technologies Limited, trading as Google DeepMind or simply DeepMind, is a British–American artificial intelligence research laboratory which serves as a subsidiary of Alphabet Inc. Founded in the UK in 2010, it was acquired by Google in 2014 and merged with Google AI's Google Brain division to become Google DeepMind in April 2023. The company is headquartered in London, with research centres in the United States, Canada, France, Germany, and Switzerland. DeepMind introduced neural Turing machines (neural networks that can access external memory like a conventional Turing machine), resulting in a computer that loosely resembles short-term memory in the human brain. DeepMind has created neural network models to play video games and board games. It made headlines in 2016 after its AlphaGo program beat a human professional Go player Lee Sedol, a world champion, in a five-game match, which was the subject of a documentary film. A more general program, AlphaZer ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Monte Carlo Tree Search
In computer science, Monte Carlo tree search (MCTS) is a heuristic search algorithm for some kinds of decision processes, most notably those employed in software that plays board games. In that context MCTS is used to solve the game tree. MCTS was combined with neural networks in 2016 and has been used in multiple board games like Chess, Shogi, Checkers, Backgammon, Contract Bridge, Go, Scrabble, and Clobber as well as in turn-based-strategy video games (such as Total War: Rome II's implementation in the high level campaign AI) and applications outside of games. History Monte Carlo method The Monte Carlo method, which uses random sampling for deterministic problems which are difficult or impossible to solve using other approaches, dates back to the 1940s. In his 1987 PhD thesis, Bruce Abramson combined minimax search with an ''expected-outcome model'' based on random game playouts to the end, instead of the usual static evaluation function. Abramson said the expected-out ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Chess
Chess is a board game for two players. It is an abstract strategy game that involves Perfect information, no hidden information and no elements of game of chance, chance. It is played on a square chessboard, board consisting of 64 squares arranged in an 8×8 grid. The players, referred to as White and Black in chess, "White" and "Black", each control sixteen Chess piece, pieces: one king (chess), king, one queen (chess), queen, two rook (chess), rooks, two bishop (chess), bishops, two knight (chess), knights, and eight pawn (chess), pawns, with each type of piece having a different pattern of movement. An enemy piece may be captured (removed from the board) by moving one's own piece onto the square it occupies. The object of the game is to "checkmate" (threaten with inescapable capture) the enemy king. There are also several ways a game can end in a draw (chess), draw. The recorded history of chess goes back to at least the emergence of chaturanga—also thought to be an ancesto ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




Elmo (shogi Engine)
Elmo (stylized as elmo, a blend of ''elastic'' and ''monkey'') is a computer shogi evaluation function and book file ('' joseki'') created by Makoto Takizawa (). It is designed to be used with a third-party shogi alpha–beta search engine. Combined with the ''yaneura ou'' () search, Elmo became the champion of the 27th annual World Computer Shogi Championship () in May 2017. However, in the Den Ō tournament () in November 2017, Elmo was not able to make it to the top five engines losing to (1st), shotgun (2nd), ponanza (3rd), (4th), and Qhapaq_conflated (5th). It won the World Championship again in 2021. In October 2017, DeepMind claimed that its program AlphaZero, after two hours of massively parallel training (700,000 steps or 10,300,000 games), began to exceed Elmo's performance. With a full nine hours of training (24 million games), AlphaZero defeated Elmo in a 100-game match, winning 90, losing 8, and drawing two. Elmo is free software that may be run on shogi engine i ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Shogi
, also known as Japanese chess, is a Strategy game, strategy board game for two players. It is one of the most popular board games in Japan and is in the same family of games as chess, Western chess, chaturanga, xiangqi, Indian chess, and janggi. ''Shōgi'' means general's (''shō'' ) board game (''gi'' ). Shogi was the earliest historical chess-related game to allow captured pieces to be returned to the board by the capturing player. This ''drop rule'' is speculated to have been invented in the 15th century and possibly connected to the practice of 15th-century Mercenary#15th to 18th centuries, mercenaries switching loyalties when captured instead of being killed. The earliest predecessor of the game, chaturanga, originated in India in the 6th century, and the game was likely transmitted to Japan via China or Korea sometime after the Nara period."Shogi". ''Encyclopædia Britannica''. 2002. Shogi in its present form was played as early as the 16th century, while a direct ancesto ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Demis Hassabis
Sir Demis Hassabis (born 27 July 1976) is a British artificial intelligence (AI) researcher, and entrepreneur. He is the chief executive officer and co-founder of Google DeepMind, and Isomorphic Labs, and a UK Government AI Adviser. In 2024, Hassabis and John M. Jumper were jointly awarded the Nobel Prize in Chemistry for their AI research contributions for protein structure prediction. Hassabis is a Fellow of the Royal Society, and has won many prestigious awards for his research work including the Breakthrough Prize, the Canada Gairdner International Award, and the Lasker Award. In 2017 he was appointed a CBE and listed in the Time 100 most influential people list. In 2024 he was knighted for services to AI, and was listed in the Time 100 again in 2025, this time featured in one of the five covers of the printed version. Early life and education Hassabis was born to Costas and Angela Hassabis. His father is Greek Cypriot and his mother is from Singapore. Demis grew up in ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Self-play (reinforcement Learning Technique)
Self-play is a technique for improving the performance of reinforcement learning agents. Intuitively, agents learn to improve their performance by playing "against themselves". Definition and motivation In multi-agent reinforcement learning experiments, researchers try to optimize the performance of a learning agent on a given task, in cooperation or competition with one or more agents. These agents learn by trial-and-error, and researchers may choose to have the learning algorithm play the role of two or more of the different agents. When successfully executed, this technique has a double advantage: # It provides a straightforward way to determine the actions of the other agents, resulting in a meaningful challenge. # It increases the amount of experience that can be used to improve the policy, by a factor of two or more, since the viewpoints of each of the different agents can be used for learning. Czarnecki et al argue that most of the games that people play for fun are "Ga ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Draw (chess)
In chess, there are a number of ways that a game can end in a draw, in which neither player wins. Draws are codified by various rules of chess including stalemate (when the player to move is not in check (chess), check but has no legal move), threefold repetition (when the same position occurs three times with the same player to move), and the fifty-move rule (when the last fifty successive moves made by both players contain no or pawn (chess), pawn move). Under the standard FIDE rules, a draw also occurs in a ''dead position'' (when no sequence of legal moves can lead to checkmate), most commonly when neither player has sufficient to checkmate the opponent. Unless specific tournament rules forbid it, players may draw by agreement, agree to a draw at any time. Ethical considerations may make a draw uncustomary in situations where at least one player has a reasonable chance of winning. For example, a draw could be called after a move or two, but this would likely be thought unsp ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




Reinforcement Learning
Reinforcement learning (RL) is an interdisciplinary area of machine learning and optimal control concerned with how an intelligent agent should take actions in a dynamic environment in order to maximize a reward signal. Reinforcement learning is one of the three basic machine learning paradigms, alongside supervised learning and unsupervised learning. Reinforcement learning differs from supervised learning in not needing labelled input-output pairs to be presented, and in not needing sub-optimal actions to be explicitly corrected. Instead, the focus is on finding a balance between exploration (of uncharted territory) and exploitation (of current knowledge) with the goal of maximizing the cumulative reward (the feedback of which might be incomplete or delayed). The search for this balance is known as the exploration–exploitation dilemma. The environment is typically stated in the form of a Markov decision process (MDP), as many reinforcement learning algorithms use dyn ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]