Archive for the ‘Alphazero’ Category

No-Castling Masters: Kramnik and Caruana will play in Dortmund – ChessBase

Press release by Initiative Pro Schach

The field of participants for the NC World Masters, part of the 50th edition of the International Dortmund Chess Days Festival, has been determined. The 14th World Chess Champion, Vladimir Kramnik, and former World Championship challenger Fabiano Caruana will be playing no-castling chess at the Goldsaal of the Dortmund Westfalenhallen from 26 June.

Navigating the Ruy Lopez Vol.1-3

The Ruy Lopez is one of the oldest openings which continues to enjoy high popularity from club level to the absolute world top. In this video series, American super GM Fabiano Caruana, talking to IM Oliver Reeh, presents a complete repertoire for White.

Vladimir Kramnik already played a match of no-castling chess against Viswanathan Anand in the first edition of the event, in 2021. He is a great advocate of the chess variation and researched it early on together with Alpha Zero, the AI engine developed by DeepMind, the world-leading company in this field.

Vladimir Kramnik

Fabiano Caruana is not only a World Championship challenger, but also a three-time winner of the Dortmund super-tournament. He won the event in 2012, 2014 and 2015. His last visit to Dortmund was in 2016, when he finished in third place.

Fabiano Caruana

Last years winner, Dmitrij Kollars, will also return to Dortmund. The German national player was a late replacement at the 2022 NC World Masters and was able to adapt to the special format very quickly. Kollars celebrated the biggest success of his career by winning the tournament ahead of Viswanathan Anand.

Dmitrij Kollars

The fourth player is Pavel Eljanov. The Ukrainian impressively won the grandmaster tournament of the International Dortmund Festival two years in a row.

Master Class Vol.11: Vladimir Kramnik

This DVD allows you to learn from the example of one of the best players in the history of chess and from the explanations of the authors (Pelletier, Marin, Mller and Reeh) how to successfully organise your games strategically, consequently how to keep y

Pavel Eljanov

The organizing association, Initiative pro Schach e.V., has not only put together an absolute top field, but also invited outstanding players from previous tournaments years to the 50th anniversary. This underlines the historical significance of the chess festival for the region and the chess world.

The tournament starts on Monday, 26 June, at the Goldsaal of the Dortmund Westfalenhallen. The players will meet each opponent twice until Sunday, 2 July. Thursday is a rest day. The exact pairings will be published well in advance.

Spectators and participants of the Chess Festival will again have the chance to watch the stars up close in Dortmund. The A-Open will be played in the same room as the NC World Masters, the Goldsaal of the Dortmund Westfalenhallen.

Follow this link:
No-Castling Masters: Kramnik and Caruana will play in Dortmund - ChessBase

AI is teamwork Bits&Chips – Bits&Chips

Albert van Breemen is the CEO of VBTI.

15 March

Like with any tool, its knowing how to use it that makes a deep-learning algorithm useful, observes Albert van Bremen.

Last week, I visited a customer interested to learn more about artificial intelligence and its application in pick-and-place robots. After a quick personal introduction, I started to share some of my learnings while working for more than four years in the field of applying deep learning to high-tech systems. Somewhat proudly I explained that almost all deep-learning algorithms out there are available as open-source implementations. This means, I said, that anybody with some Python programming experience can download deep-learning models from the internet and start training. My customer promptly asked: If everything is open and accessible to any artificial-intelligence company, how do they differentiate between themselves?

The question took me a bit off-guard. After a short hesitation, I replied: In the same way that a hammer and a spade are tools that are available to everybody, not everybody can make beautiful things with them. Data and algorithms are the tools of an AI engineer. Artificial-intelligence companies can set themselves apart with their experience and knowledge of applying these tools to solve engineering problems. While my answer kept the conversation going well at that time, I needed to reflect on it later.

Having access to data and algorithms doesnt give any guarantees that you can make deep learning work. In my company, I introduced the Friday Afternoon Experiments, something I borrowed from Phillips Research when I was working there back in 2001. Everybody in my company can spend the Friday afternoon on a topic theyre interested in and think might be relevant for the company. It encourages knowledge development, innovation and work satisfaction.

I started a Friday Afternoon Experiment myself, repeating a Deepmind project. In 2016, Deepmind created an algorithm called Alphago that was the first to defeat a professional human Go player. In a short time, the algorithm developed into the more generic Alphazero algorithm, which was trained in one day to play Go, Chess and Shogi at world champion level.

The devil of deep-learning technology is in the details

It took me over three months to get my Alphazero to work for the less complex games Connect 4 and Othello. In one day, I can now train a strong Connect 4 or Othello Alphazero player. The project took way longer than I hoped for. It made me realize that the devil of deep-learning technology really is in the details. Deep-learning algorithms learn from data. But to set up the learning process and train it successfully, you must define many so-called hyper-parameters. Small changes matter a lot, and a large part of your time can be spent on finding good hyper-parameter settings. Im lucky to have an experienced team to discuss problems and bottlenecks.

Besides data and algorithms, compute power was a key success factor of Deepmind. To stay with the metaphor of tools, some AI companies have power tools that differentiate them from others. Companies like OpenAI, Deepmind and Meta have huge amounts of compute power available for deep-learning purposes. The AI trinity of dataalgorithmscompute power defines the complexity level of the problems they can solve. If all you have is a spade, you can dig a decent hole in a day. If you have an excavator, you can dig a swimming pool within the same timeframe. Huge compute power is something not all companies have access to and this is where some AI companies can differentiate. Deepmind trained Alphago using thousands of CPUs and hundreds of GPUs. I was limited during my experiment to 64 CPU cores and 1 GPU.

If youre searching for a solution to a standard problem, you can almost go to any artificial-intelligence startup. However, if you have a problem that hasnt been solved before, you need more than just data, algorithms and compute power. An experienced and dedicated team makes the difference. This might seem obvious, but AI techno-babble might easily let you think otherwise. AI is teamwork!

Read more:
AI is teamwork Bits&Chips - Bits&Chips

Resolve Strategic nuclear subs poll (open thread) The Poll Bludger – The Poll Bludger

A detailed poll on the AUKUS nuclear submarines deal finds strong support among Labor and Coalition voters alike.

The Age/Herald published a Resolve Strategic poll on Saturday concerning AUKUS and nuclear submarines, which I held back on doing a post on because I thought voting intention results might follow. That hasnt happened yet, so here goes.

As is perhaps unavoidable with the matter at hand, respondents were given fairly lengthly explanations of the relevant issues before having their opinions gauged on them, such that the results need to be considered carefully alongside what was actually asked. The first outlined the proposed acquisition and pointed out both the expense and the expectation that it would create 20,000 jobs, and found 50% in favour and 17% opposed. Breakdowns by party support found near identical results for Labor and Coalition results, with weaker support among an others category inclusive of both the Greens and minor parties of the right.

The second question asked respondents how they felt specifically about Australian submarines being nuclear-powered, finding 25% actively supportive, 39% considering the notion acceptable, and 17% actively opposed. The third put it to respondents that the federal government has hitherto being committed to spending 2% of GDP on defence, and that Anthony Albanese says he would like to spend more: 39% concurred, 31% said it should remain as is, and 9% felt it should be reduced. Finally, 46% felt large single-party states, like Russia and China were a threat to Australia, but one that could be carefully managed; 36% felt they were a threat that needed to be confronted soon; and 8% felt they were no threat at all.

The sample was conducted last Sunday to Thursday from a sample of 1600.

William Bowe is a Perth-based election analyst and occasional teacher of political science. His blog, The Poll Bludger, has existed in one form or another since 2004, and is one of the most heavily trafficked websites on Australian politics.View all posts by William Bowe

See the article here:
Resolve Strategic nuclear subs poll (open thread) The Poll Bludger - The Poll Bludger

How AlphaZero Learns Chess – Chess.com

AlphaZero's learning process is, to some extent, similar to that of humans. A new paper from DeepMind, which includes a contribution from the 14th world chess champion Vladimir Kramnik, provides strong evidence for the existence of human-understandable concepts in AlphaZero's network, even though AlphaZero has never seen a human game of chess.

How does AlphaZero learn chess? Why does it make certain moves? What values does it give to concepts such as king safety or mobility? How does it learn openings, and how is that different from how humans developed opening theory?

Questions like these are being discussed in a fascinating new paper by DeepMind, titled Acquisition of Chess Knowledge in AlphaZero. It was written by Thomas McGrath, Andrei Kapishnikov, Nenad Tomasev, Adam Pearce, Demis Hassabis, Been Kim, and Ulrich Paquet together with Kramnik. It is the second cooperation between DeepMind and Kramnik, after their research from last year when they used AlphaZero to explore the design of different variants of the game of chess, with different sets of rules.

In their latest paper, the researchers tried a method for encoding human conceptual knowledge, to determine the extent to which the AlphaZero network represents human chess concepts. Examples of such concepts are the bishop pair, material (im)balance, mobility, or king safety. These concepts have in common that they are pre-specified functions that encapsulate a particular piece of domain-specific knowledge.

Some of these concepts were taken from Stockfish 8's evaluation function, such as material, imbalance, mobility, king safety, threats, passed pawns, and space. Stockfish 8 uses these as sub-functions that give individual scores leading to a "total" evaluation that is exported as a continuous value, such as "0.25" (a slight advantage to White) or "-1.48" (a big advantage to Black). Note that more recent versions of Stockfish have developed into Alpha-Zero-like neural networks but were not used for this paper.

The third type of concepts encapsulates more specific lower-level features, such as the existence of forks, pins, or contested files, as well as a range of features regarding pawn structure.

Having established this wide array of human concepts, the next step for the researchers was to try and find them within the AlphaZero network, for which they used a sparse linear regression model. After that, they started visualizing the human concept learning with what they call what-when-where plots: what concept is learned when in training time where in the network.

According to the researchers, AlphaZero indeed develops representations that are closely related to a number of human concepts over the course of training, including high-level evaluation of the position, potential moves and consequences, and specific positional features.

One interesting result was about material imbalance.As was demonstrated in Matthew Sadler and Natasha Regan's award-winning book Game Changer: AlphaZeros Groundbreaking Chess Strategies and the Promise of AI (New In Chess, 2019), AlphaZero seems to view material imbalance differently from Stockfish 8. The paper gives empirical evidence that this is the case at the representational level: AlphaZero initially "follows" Stockfish 8's evaluation of material more and more during its training, but at some point, it turns away from it again.

The next step for the researchers was to relate the human concepts to AlphaZero's value function. One of the first concepts they looked at was piece value, something a beginner will first learn when starting to play chess. The classical values are nine for a queen, five for a rook, three for both the bishop and knight, and one for a pawn. The left figure below (taken from the paper) shows the evolution of piece weights during AlphaZero's training, with piece values converging towards commonly-accepted values.

The image on the right shows that during AlphaZero's training, material becomes more and more important in the early stages of learning chess (consistent to human learning) but it reaches a plateau and at some point, the values of more subtle concepts such as mobility and king safety are becoming more important while material actually decreases in importance.

Another part of the paper is dedicated to comparing AlphaZero's training to the progression of human knowledge over history. The researchers point out that there is a marked difference between AlphaZeros progression of move preferences through its history of training steps, and what is known of the progression of human understanding of chess since the 15th century:

AlphaZero starts with a uniform opening book, allowing it to explore all options equally, and largely narrows down plausible options over time. Recorded human games over the last five centuries point to an opposite pattern: an initial overwhelming preference for 1.e4, with an expansion of plausible options over time.

The researchers compare the games AlphaZero is playing against itself with a large sample taken from the ChessBase Mega Database, starting with games from the year 1475 up till the 21st century.

Humans initially played 1.e4 almost exclusively but 1.d4 was slightly more popular in the early 20th century, soon followed by the increasing popularity of more flexible systems like 1.c4 and 1.Nf3. AlphaZero, on the other hand, tries out a wide array of opening moves in the early stage of its training before starting to value the "main" moves higher.

A more specific example provided is about the Berlin variation of the Ruy Lopez (the move 3...Nf6 after 1.e4 e5 2.Nf3 Nc6 3.Bb5), which only became popular at the top level early 21st century, after Kramnik successfully used it in his world championship match with GM Garry Kasparov in 2000. Before that, it was considered to be somewhat passive and slightly better for White with the move 3...a6 being preferable.

The researchers write:

Looking back in time, it took a while for human chess opening theory to fully appreciate the benefits of Berlin defense and to establish effective ways of playing with Black in this position. On the other hand, AlphaZero develops a preference for this line of play quite rapidly, upon mastering the basic concepts of the game. This already highlights a notable difference in opening play evolution between humans and the machine.

Remarkably, when different versions of AlphaZero are trained from scratch, half of them strongly prefer 3 a6, while the other half strongly prefer 3 Nf6! It is interesting as it means that there is no "unique good chess player. The following table shows the preferences of four different AlphaZero neural networks:

The AlphaZero prior network preferences after 1. e4 e5 2. Nf3 Nc6 3. Bb5, for four different training runs of the system (four different versions of AlphaZero). The prior is given after one million training steps. Sometimes AlphaZero converges to become a player that prefers 3 a6, and sometimes AlphaZero converges to become a player that prefers to respond with 3 Nf6.

In a similar vein, AlphaZero develops its own opening "theory" for a much wider array of openings over the course of its training. At some point, 1.d4 and 1.e4 are discovered to be good opening moves and are rapidly adopted. Similarly, AlphaZero's preferred continuation after 1.e4 e5 is determined in another short temporal window. The figure below illustrates how both 2.d4 and 2.Nf3 are quickly learned as reasonable White moves, but 2.d4 is then dropped almost as quickly in favor of 2.Nf3 as a standard reply.

Kramnik's contribution to the paper is a qualitative assessment, as an attempt to identify themes and differences in the style of play of AlphaZero at different stages of its training. The 14th world champion was provided sample games from four different stages to look at.

According to Kramnik, in the early training stage, AlphaZero has "a crude understanding of material value and fails to accurately assess material in complex positions. This leads to potentially undesirable exchange sequences, and ultimately losing games on material." In the second stage, AlphaZero seemed to have "a solid grasp on material value, thereby being able to capitalize on the material assessment weakness" of the early version.

In the third stage, Kramnik feels that AlphaZero has a better understanding of king safety in imbalanced positions. This manifests in the second version "potentially underestimating the attacks and long-term material sacrifices of the third version, as well as the second version overestimating its own attacks, resulting in losing positions."

In its fourth stage of the training, has a "much deeper understanding" of which attacks will succeed and which would fail. Kramnik notices that it sometimes accepts sacrifices played by the "third version," proceeds to defend well, keep the material advantage, and ultimately converts to a win.

Another point Kramnik makes, which feels similar to how humans learn chess, is that tactical skills appear to precede positional skills as AlphaZero learns. By generating self-play games over separate opening sets (e.g. the Berlin or the Queen's Gambit Declined in the "positional" set and the Najdorf and King's Indian in the "tactical" set), the researchers manage to provide circumstantial evidence but note that further work is needed to understand the order in which skills are acquired.

For a long time, it was believed that machine-learning systems learn uninterpretable representations that have little in common with human understanding of the domain they are trained on. In other words, how and what AI teaches itself is mostly gibberish to humans.

With their latest paper, the researchers have provided strong evidence for the existence of human-understandable concepts in an AI system that wasn't exposed to human-generated data. AlphaZero's network shows the use of human concepts, even though AlphaZero has never seen a human game of chess.

This might have implications outside the chess world. The researchers conclude:

The fact that human concepts can be located even in a superhuman system trained by self-play broadens the range of systems in which we should expect to find human-understandable concepts. We believe that the ability to find human-understandable concepts in the AZ network indicates that a closer examination will reveal more.

Co-author Nenad Tomasev commented to Chess.com that for him personally, he was really curious to consider if there is such a thing as a "natural" progression of chess theory:

Even in the human contextif we were to 'restart' history, go back in timewould the theory of chess have developed in the same way? There were a number of prominent schools of thought in terms of the overall understanding of chess principles and middlegame positions: the importance of dynamism vs. structure, material vs. sacrificial attacks, material imbalance, the importance of space vs. the hypermodern school that invites overextension in order to counterattack, etc. This also informed the openings that were played. Looking at this progression, what remains unclear is whether it would have happened the same way again. Maybe some pieces of chess knowledge and some perspectives are simply easier and more natural for the human mind to grasp and formulate? Maybe the process of refining them and expanding them has a linear trajectory, or not? We can't really restart history, so we can only ever guess what the answer might be.

However, when it comes to AlphaZero, we can retrain it many timesand also compare the findings to what we have previously seen in human play. We can therefore use AlphaZero as a Petri dish for this question, as we look at how it acquires knowledge about the game. As it turns out, there are both similarities and dissimilarities in how it builds its understanding of the game compared to human history. Also, while there is some level of stability (results being in agreement across different training runs), it is by no means absolute (sometimes the training progression looks a little bit different, and different opening lines end up being preferred).

Now, this is by no means a definitive answer to what is, to me personally, a fascinating question. There is still plenty to think about here. Yet, we hope that our results provide an interesting perspective and make it possible for us to start thinking a bit deeper about how we learn, grow, improvethe very nature of intelligence and how it goes all the way from a blank slate to what is a deep understanding of a very complex domain like chess.

Kramnik commented to Chess.com:

"There are two major things which we can try to find out with this work. One is: how does AlphaZero learn chess, how does it improve? That is actually quite important. If we manage one day to understand it fully, then maybe we can interpret it into the human learning process.

Secondly, I believe it is quite fascinating to discover that there are certain patterns that AlphaZero finds meaningful, which actually make little sense for humans. That is my impression. That actually is a subject for further research, in fact, I was thinking that it might easily be that we are missing some very important patterns in chess, because after all, AlphaZero is so strong that if it uses those patterns, I suspect they make sense. That is actually also a very interesting and fascinating subject to understand, if maybe our way of learning chess, of improving in chess, is actually quite limited. We can expand it a bit with the help of AlphaZero, of understanding how it sees chess."

Read this article:
How AlphaZero Learns Chess - Chess.com

AI Topic: AlphaZero, ChatGPT, Bard, Stable Diffusion and more!

I'm in on the Bing AI (aka: ChatGPT).

I decided to have as "natural" of a discussion as I could with the AI. I already know the answers since I've done research in this subject, so I'm pretty aware of mistakes / errors as they come up. Maybe for a better test, I need to use this as a research aid and see if I'm able to pick up on the bullshit on a subject I don't know about...

Well, bam. Already Bing is terrible, unable to answer my question and getting it backwards (giving a list of RP2040 reasons instead of AVR reasons). Its also using a rather out-of-date ATMega328 as a comparison point. So I type up a quick retort to see what it says...

This is... wrong. RP2040 doesn't have enough current to drive a 7-segment LED display. PIO seems like a terrible option as well. MAX7219 is a decent answer, but Google could have given me that much faster (ChatGPT / Bing is rather slow).

"Background Writes" is a software thing. You'd need to combine it with the electrical details (ie: MAX7219).

7-segment displays can't display any animations. The amount of RAM you need to drive it is like... 1 or 2 bytes, the 264kB RAM (though an advantage to the RP2040), is completely wasted in this case.

Fail. RP2040 doesn't have enough current. RP2040 literally cannot do the job as they describe here.

Wow. So apparently its already forgotten what the AVR DD was, despite giving me a paragraph or two just a few questions ago. I thought this thing was supposed to have better memory than that?

I'll try the ATMega328p, which is what it talked about earlier.

Fails to note that ATMega328 has enough current to drive the typical 7-segment display even without a adapter like MAX7219. So despite all this rambling, its come to the wrong conclusion.

------------

So it seems like ChatGPT / Bing AI is about doing a "research", while summarizing pages from the top of the internet for the user? You don't actually know if the information is correct or not however, so that limits its usefulness.

It seems like Bing AI is doing a good job at summarizing the articles that pop up on the internet, and giving citations. But its conclusions and reasoning can be very wrong. It also can have significant blind spots (ie: RP2040 not having enough current to directly drive a 7-segment display. A key bit of information that this chat session was unable to discover, or even figure out it might be a problem).

----------

Anyone have a list of questions they want me to give to ChatGPT?

Another run...

I think I'm beginning to see what this chatbot is designed to do.

1. This thing is decent at summarizing documents. But notice: it pulls the REF1004 as my "5V" voltage reference. Notice anything wrong? https://www.ti.com/lit/ds/sbvs002/sbvs002.pdf . Its a 2.5V reference, seems like ChatGPT pattern-matched on 5V and doesn't realize its a completely different number than 2.5V (or some similar error?)

2. Holy crap its horrible at math. I don't even need a calculator, and the 4.545 kOhm + 100 Ohm trimmer pot across 5V obviously can't reach 1mA, let alone 0.9mA. Also, 0.9mA to 1.1mA is +/- 10%, I was asking for 1.000mA.

-------

Instead, what ChatGPT is "good" at, is summarizing articles that exist inside of the Bing Database. If it can "pull" a fact out of the search engine, it seems to summarize it pretty well. But the moment it tries to "reason" with the knowledge and combine facts together, it gets it horribly, horribly wrong.

Interesting tool. I'll need to play with it more to see how it could possibly ever be useful. But... I'm not liking it right now. Its extremely slow, its wrong in these simple cases. So I'm quite distrustful of it being a useful tool on a subject I know nothing about. I'd have to use this tool on a subject I'm already familiar with, so that I can pick out the bullshit from the good stuff.

See more here:
AI Topic: AlphaZero, ChatGPT, Bard, Stable Diffusion and more!