The match was a five-game match with one million US dollars as the grand prize, The match was played at the
Four Seasons Hotel in
Seoul, South Korea in March 2016 and was video-streamed live with commentary; the English language commentary was done by
Michael Redmond (9-dan professional) and Chris Garlock.
Aja Huang, a DeepMind team member and amateur 6-dan Go player, placed stones on the
Go board for AlphaGo, which ran through the
Google Cloud Platform with its server located in the United States.
Summary Game 1 AlphaGo (white) won the first game. Lee appeared to be in control throughout the match, but AlphaGo gained the advantage in the final 20 minutes, and Lee resigned. David Ormerod, commenting on the game at Go Game Guru, described Lee's seventh stone as "a strange move to test AlphaGo's strength in the opening", characterising the move as a mistake and AlphaGo's response as "accurate and efficient". He described AlphaGo's position as favourable in the first part of the game, considering that Lee started to come back with move 81 before making "questionable" moves at 119 and 123, followed by a "losing" move at 129. Professional Go player
Cho Hanseung commented that AlphaGo's game had greatly improved from when it beat
Fan Hui in October 2015. According to 9-dan Go grandmaster Kim Seong-ryong, Lee seemed stunned by AlphaGo's strong play on the 102nd stone. After watching AlphaGo make the game's 102nd move, Lee mulled over his options for more than 10 minutes. "from very beginning of the game I did not feel like there was a point that I was leading". One of the creators of AlphaGo, Demis Hassabis, said that the system was confident of victory from the midway point of the game, even though the professional commentators could not tell which player was ahead. AlphaGo showed anomalies and moves from a broader perspective, which professional Go players described as looking like mistakes at first sight but an intentional strategy in hindsight. As one of the creators of the system explained, AlphaGo does not attempt to maximize its points or its margin of victory, but tries to maximize its probability of winning. After the second game, players still had doubts about whether AlphaGo was truly a strong player in the sense that a human might be. The third game was described as removing that doubt, with analysts commenting that: According to An Younggil (8p) and David Ormerod, the game showed that "AlphaGo is simply stronger than any known human Go player." AlphaGo was seen to capably navigate tricky situations known as
ko that did not come up in the previous two matches. An and Ormerod consider move 148 to be particularly notable: in the middle of a complex
ko fight, AlphaGo displayed sufficient "confidence" that it was winning the game to play a significant move elsewhere. AlphaGo had gained control of the game by move 48, and forced Lee onto the defensive. Lee counterattacked at moves 77/79, but AlphaGo's response was effective, and its move 90 succeeded in simplifying the position. It then gained a large area of control at the bottom of the board, strengthening its position with moves from 102 to 112 described by An as "sophisticated". By doing so, his apparent aim was to force an "all or nothing" style of situation – a possible weakness for an opponent strong at negotiation types of play, and one which might make AlphaGo's capability of deciding slim advantages largely irrelevant.
Gu Li (9p) described it as a "
divine move" and stated that the move had been completely unforeseen by him. AlphaGo responded poorly on move 79, at which time it estimated it had a 70% chance to win the game. Lee followed up with a strong move at white 82. provoking it to make a series of very bad moves from black 87 to 101. David Ormerod characterised moves 87 to 101 as typical of Monte Carlo-based program mistakes. For this reason, he requested that he play black in the fifth game, which is considered more risky. David Ormerod of Go Game Guru stated that although an analysis of AlphaGo's play around 79–87 was not yet available, he believed it resulted from a known weakness in play algorithms that use
Monte Carlo tree search. In essence, the search attempts to prune less relevant sequences. In some cases, a play can lead to a particular line of play which is significant but which is overlooked when the tree is pruned, and this outcome is therefore "off the search radar".
Game 5 AlphaGo (white) won the fifth game. Lee, playing black, opened similarly to the first game and began to stake out territory in the right and top left corners – a similar strategy to the one he employed successfully in game 4 – while AlphaGo gained influence in the centre of the board. The game remained even until white moves 48 to 58, which AlphaGo played in the bottom right. These moves unnecessarily lost ko threats and aji, allowing Lee to take the lead. == Coverage ==