piotr makowski 27LH 0jXKYI unsplash

IT TOOK ABOUT 50 years for computers to eviscerate humans in the venerable game of chess. A standard smartphone can now play the kind of moves that make a grandmaster’s head spin. But one artificial intelligence program is taking a few steps backward, to appreciate how average humans play—blunders and all.

The AI chess program, known as Maia, uses the kind of cutting-edge AI behind the best superhuman chess-playing programs. But instead of learning how to destroy an opponent on the board, Maia focuses on predicting human moves, including the mistakes they make.

Jon Kleinberg, a professor at Cornell University who led the development of Maia, says this is a first step toward developing AI that better understands human fallibility. The hope is that it may therefore be better at interacting with humans, by teaching or assisting them, for example, or even negotiating with them.

One possible use, Kleinberg says, is health care. A system that anticipates errors might be used to train doctors to read medical images or help them catch errors. “One way to do this is to take problems in which human doctors form diagnoses based on medical images, and to look for images on which the system predicts a high level of disagreement among them,” he says.

“Just imagine you’re preparing for the world championship game against Magnus Carlsen, and you have an engine that plays just like Magnus.”

MATTHEW SADLER, BRITISH CHESS GRANDMASTER, AUTHOR OF GAME CHANGER

Kleinberg says he chose to focus on chess because it is one of the first domains where machine intelligence has triumphed over humans. “It is this sort of ideal system for trying out algorithms,” he says. “Sort of a model for AI dominance.”

In addition, he says, chess has been studied intensely, making it something similar to the fruit fly, or drosophila, in biology. “Chess has the distinction of having been called the drosophila of psychology by Herb Simon and the drosophila of AI by John McCarthy,” Kleinberg says, referring to two giants of their respective fields.

Maia was developed using code adapted from Leela Zero, an open source clone of Alpha Zero, a revolutionary AI program created by the Alphabet subsidiary DeepMind.

Alpha Zero broke from conventional AI chess programs by having computers learn, independent of any human instruction, how to play the game expertly. Within the program, a simulated neural network contains virtual neurons that can be tuned to respond to input. For chess, Alpha Zero is fed board positions and moves generated in practice games, and it tunes its neurons’ firing to favor winning moves, an approach known as reinforcement learning. Alpha Zero can use the same approach to learn to play other board games such as checkers or Go with minimal modification.

The Cornell team modified Leela Zero’s code to create a program that learned by favoring accurate predictions of human moves. Other AI chess players, including Deep Blue, the IBM machine that defeated then world champ Garry Kasparov in 1997, may attempt to look ahead in a game by exploring possible moves. But Maia is unusual in how it focuses on finding the most likely move a human will play.

Maia was taught using data from LiChess, a popular online chess server. The result is a chess program capable of playing in a more human way. Several versions of Maia, tuned to different strengths of play, can now be challenged at LiChess.

More sophisticated forms of AI may eventually outstrip human intelligence in all sorts of domains, from mathematics to literature and beyond. But Kleinberg says “there’ll be a long transition period where AI and humans will be working together, and there’s going to be some communication between them.”

Sadler says Maia could be useful for training and practice, and he says experts have talked about the idea of chess programs that mimic particular players for a while. With a sufficient number of games to learn from, Maia might be trained to predict only the moves of a particular player. “Just imagine you’re preparing for the world championship game against Magnus Carlsen,” Sadler says, referring to the current world champion. “And you have an engine that plays just like Magnus.”

This might extend beyond chess. Imagine playing a video game against an AI player trained to replicate a top esports star. Beyond games, AI programs that understand human behavior could help companies preempt tactics in negotiations, or produce software programs and robots that can anticipate what human coworkers are about to do.

“The questions they are asking—how well can the approach predict human performance—are interesting,” says Julie Shah, an associate professor at MIT who studies human-machine interaction and collaboration. She notes that the technical side of the work would need to be validated, but she suggests it would be interesting to examine whether the approach can produce a superior way for humans and machines to collaborate on chess.

Matthias Söllner, a professor at Germany’s University of Kassel who studies how AI can help office workers, says it will be crucial for such systems to behave in more humanlike ways; but he says it may be even more important for people to understand how such AI systems work. If a system performs poorly, and it’s not clear why, he says, “it might actually hurt acceptance of such AI systems.”

Will Knight