Alphazero Chess Download

0 views
Skip to first unread message

Dibe Naro

unread,
May 10, 2024, 4:15:44 PM5/10/24
to talihosul

Demis Hassabis, the founder and CEO of DeepMind and an expert chess player himself, presented further details of the system, called Alpha Zero, at an AI conference in California on Thursday. The program often made moves that would seem unthinkable to a human chess player.

alphazero chess download


Downloadhttps://t.co/xfGdmx9pIK



Alpha Zero is a more general version of AlphaGo, the program developed by DeepMind to play the board game Go. In 24 hours, Alpha Zero taught itself to play chess well enough to beat one of the best existing chess programs around.

20 years after DeepBlue defeated Garry Kasparov in a match,chess players have awoken to a new revolution. The AlphaZero algorithm developedby Google and DeepMind took just four hours of playing against itself tosynthesise the chess knowledge of...

Chess engines are powerful tools used on a regular basis by chess professionals and amateurs alike in analysing and understanding individual positions and openings. Recently, neural network chess engines such as AlphaZero, Leela Chess Zero, the Stockfish NNUE and Fat Fritz have emerged as powerful chess engines that are able to challenge more traditional engines which use manually implemented evaluation functions.

Abstract: What is being learned by superhuman neural network agents such as AlphaZero? This question is of both scientific and practical interest. If the representations of strong neural networks bear no resemblance to human concepts, our ability to understand faithful explanations of their decisions will be restricted, ultimately limiting what we can achieve with neural network interpretability. In this work we provide evidence that human knowledge is acquired by the AlphaZero neural network as it trains on the game of chess. By probing for a broad range of human chess concepts we show when and where these concepts are represented in the AlphaZero network. We also provide a behavioural analysis focusing on opening play, including qualitative analysis from chess Grandmaster Vladimir Kramnik. Finally, we carry out a preliminary investigation looking at the low-level details of AlphaZero's representations, and make the resulting behavioural and representational analyses available online.

Anyone who seriously deals with openings cannot avoid the opening encyclopaedia. Whether beginner or grandmaster. The Opening Encyclopaedia is by far the most comprehensive chess theory work: over 1,463(!) theory articles offer a huge fund of ideas!

We analyze the knowledge acquired by AlphaZero, a neural network engine that learns chess solely by playing against itself yet becomes capable of outperforming human chess players. Although the system trains without access to human games or guidance, it appears to learn concepts analogous to those used by human chess players. We provide two lines of evidence. Linear probes applied to AlphaZero's internal state enable us to quantify when and where such concepts are represented in the network. We also describe a behavioral analysis of opening play, including qualitative commentary by a former world chess champion.

First and foremost is that in Go deepmind had no superhuman opponents to challenge. Go engines were not anywhere near the highest level of the top human players. In chess, however, the engines are 500 ELO points stronger than the top human players. This is a massive difference. The amount of work that has gone into contemporary chess engines is staggering. We are talking about millions of hours in programming, hundreds of thousands of iterations. It is a massive body of knowledge and work. To overcome and surpass all of that in 4 hours is staggering.

Secondly it is not so much the result itself which is surprising to chess masters but instead its how AlphaZero plays chess. It's quite ironic that a system which had no human knowledge or expertise plays the most like we do. Engines are notorious for playing ugly looking moves, those lacking harmony etc. Its hard to explain to a non-chess player but there is such a thing as an "Artificial move" like the contemporary engines come up with often. AlphaZero does not play like this at all. It has a very human-like style where it dominates the opponent's pieces with deep strategic play and stunning position sacrifices. AlphaZero plays the way we aspire to, combining deep positional understanding with the precision of an engines calculation.

MCTS for chess had been tried in the literature with little success. It was assumed AlphaGo's approach would never work on chess, maybe in Go but not in chess. Suddenly, Google announced the approach was working and it was beating the World's strongest chess program by a very signficiant margin.

Before Google, all chess programmers were taught crafting heuristics in engine programming was a better strategy than machine learning. No matter how you implemented neural networks, it would have never ran faster than a bunch of 64-bit bitboards instructions. AlphaGo was running quite slow, but it played strongest chess.

The world was appalled when AlphaGo first played Lee Sedol in Go, winning 4 matches to 1. DeepMind subsequently released AlphaGo Zero, an iteration on AlphaGo that beat it 100-1. Going even further, they released AlphaZero, which learned how to play games such as Shogi and Chess. Kevin, an avid chess enthusiast, wanted to discuss what this meant for the Chess world.

In any case, my immediate concern from this development is how it will affect human chess. Over the last decade or two, engines such as stockfish have played a profound role in the development of chess, with current top players such as Magnus Carlsen or Sergey Karjakin having trained extensively with these engines.

DeepMind's AlphaZero was a turning point in AI research, achieving superhuman capabilities through self-play reinforcement learning and mastering chess at a new level. However, difficult chess puzzles still baffle even the strongest chess AI systems, suggesting room for improvement. Researchers at Google DeepMind are now proposing to combine several different AlphaZero agents into an ensemble system, called AZdb, to further improve its capabilities in chess and beyond. AZdb, combines multiple AlphaZero agents into a "league".

Using "behavioral diversity" and "response diversity" techniques, AZdb's agents are trained to play chess in different ways. According to Google Deepmind, behavioral diversity maximizes the difference in average piece positions between agents, while response diversity exposes agents to games against different opponents. In practice, this also means that AZdb's agent will get to see many more different positions, expanding this range of in-distribution data, which should allow the system to better generalize to unseen positions.

As inspiration for this approach, the team cites cases in which clubs used to collaborate and play against each other via correspondence chess, such as "Kasparov versus The World," about which the famous chess player said he had "never expended as much effort into any other game in his life." Chess grandmasters also often prepare for important games with a team of strong players with different styles.

The researchers then examined whether this diversity provides a creative advantage when attempting to solve challenging chess puzzles collected from multiple sources, including puzzles specifically designed to trick chess engines. They found that, given ample time to think, AZdb solved twice as many of these very challenging puzzles compared to individual AlphaZero. This shows that AZdb's diverse team collectively considered more possibilities, they said, with different agents specializing to excel at certain puzzle types. Chess games also showed that the agents specialized in different openings.

One of the most interesting events in the chess programming world occurred in December 2018 when a challenge match of 100 games between AlphaZero and Stockfish version 8 was played. This represented a clash of completely different classes of algorithms to computer chess.

The AlphaZero algorithm was only given the rules of the game of chess and this engine taught itself strategy by playing a very large number of games against itself. This effectively is a reinforcement learning approach that uses neural nets extensively in the implementation. The evaluation strategy checks fewer nodes but each node evaluated has a higher complexity.

One of the most interesting things about these games is that they are effectively much higher level than any human games of chess. Since Kasperov's loss to Deep Blue in 1997 there's been a lot of use of computer analysis in chess coaching and training.

One of the interesting things about the Stockfish vs AlphaZero match is that some of the strategies that come about seem to be very strong but are not necessarily "intuitive". From reviewing these games I found it interesting how the engines like pushing pawns and moving pieced to reduce opponent options even if this leaves the pieces in awkward positions, they also don't seem to mind moving the same piece more than once in the opening if it gains advantage. Perhaps one of the most interesting things is how the notion of what is considered to be "intuitive" to recent players has been shaped so heavily by use of these chess engines for analysis for the last two decades. Many of the existing engines before 2018 were focused on depth-first algorithms to do searches of the game space. This causes those computer engines to play with a certain style which has in turn influenced many of the players who have used these engines to learn.

I remember when I was younger the only way I could beat GNU chess was by playing certain irregular openings that gained positional advantages that were hard for that engine to properly compute, some gambit based openings seemed to do disproportionately well (I wish I remembered all the details, I seem to recall that there was some variations of the Scandinavian Defense that it played outstandingly badly against). I suspect if there was no opening book used by the GNU chess engine a very substantial decline in its effective ELO rating strength would have been apparent when it faced certain openings compared to others but I don't think this would have been a uniform drop across various openings. This experience playing against a specific engine shaped how I played for a while and I still play the Scandinavian Defense somewhat frequently (when I play). Notably I didn't do so well with these strategies vs stronger human players at the chess club.

08ab062aa8
Reply all
Reply to author
Forward
0 new messages