In a hotel ballroom in Seoul, South Korea, early in 2016, a centuries-old strategy game offered a glimpse into the fantastic future of computing.
The computer program AlphaGo bested a world champion player at the Chinese board game Go, four games to one (SN Online: 3/15/16). The victory shocked Go players and computer gurus alike. “It happened much faster than people expected,” says Stuart Russell, a computer scientist at the University of California, Berkeley. “A year before the match, people were saying that it would take another 10 years for us to reach this point.”
The match was a powerful demonstration of the potential of computers that can learn from experience. Elements of artificial intelligence are already a reality, from medical diagnostics to self-driving cars (SN Online: 6/23/16), and computer programs can even find the fastest routes through the London Underground. “We don’t know what the limits are,” Russell says. “I’d say there’s at least a decade of work just finding out the things we can do with this technology.”
AlphaGo’s design mimics the way human brains tackle problems and allows the program to fine-tune itself based on new experiences. The system was trained using 30 million positions from 160,000 games of Go played by human experts. AlphaGo’s creators at Google DeepMind honed the software even further by having it play games against slightly altered versions of itself, a sort of digital “survival of the fittest.”
These learning experiences allowed AlphaGo to more efficiently sweat over its next move. Programs aimed at simpler games play out every single hypothetical game that could result from each available choice in a branching pattern — a brute-force approach to computing. But this technique becomes impractical for more complex games such as chess, so many chess-playing programs sample only a smaller subset of possible outcomes. That was true of Deep Blue, the computer that beat chess master Garry Kasparov in 1997.
But Go offers players many more choices than chess does. A full-sized Go board includes 361 playing spaces (compared with chess’ 64), often has various “battles” taking place across the board simultaneously and can last for more moves.
AlphaGo overcomes Go’s sheer complexity by drawing on its own developing knowledge to choose which moves to evaluate. This intelligent selection led to the program’s surprising triumph, says computer scientist Jonathan Schaeffer of the University of Alberta in Canada. “A lot of people have put enormous effort into making small, incremental progress,” says Schaeffer, who led the development of the first computer program to achieve perfect play of checkers. “Then the AlphaGo team came along and, incremental progress be damned, made this giant leap forward.”
Real-world problems have complexities far exceeding those of chess or Go, but the winning strategies demonstrated in 2016 could be game changers.
F.-Y. Wang et al. Where does AlphaGo go: from church-turing thesis to AlphaGo thesis and beyond. IEEE/CAA Journal of Automatica Sinica. Vol. 3, April 2016, p. 113. doi: 10.1109/JAS.2016.7471613.
D. Silver et al. Mastering the game of Go with deep neural networks and tree search. Nature. Vol. 529, January 28, 2016, p. 484. doi: 10.1038/nature16961.
B. Bouzy and T. Cazenave. Computer Go: An AI oriented survey. Artificial Intelligence. Vol. 132, October 2001, p. 39.
L. Hamers. AI system learns like a human, stores info like a computer. Science News Online, October 14, 2016.
B. Bower. Moral dilemma could put brakes on driverless cars. Science News Online, June 23, 2016.
M. Rosen. Computer program bests world champion 4-1 in strategy game Go. Science News Online, March 15, 2016.
M. Rosen. Machine trumps man in strategy game Go. Science News Online, January 27, 2016.
L. Guterman. Artificial intelligence conquers Space Invaders, Pong, Q*bert. Science News Online, February 25, 2015.