Talk:1002: Game AIs
Mornington Crescent would be impossible for a computer to play, let alone win... -- 184.108.40.206 (talk) (please sign your comments with ~~~~) It is unclear which side of the line jeopard fall upon. Why so close to the line I wonder. DruidDriver (talk) 01:04, 16 January 2013 (UTC)
- Because of Watson (computer). (Anon) 13 August 2013 220.127.116.11 (talk) (please sign your comments with ~~~~)
- I agree, this is far more likely. 18.104.22.168 10:21, 11 September 2013 (UTC)
On the old blog version of this article, a comment mentioned Ken tweeting his method right after this comic was posted. He joked that they would asphyxiate themselves to actually see heaven for seven minutes. I don't know how to search for tweets, or if they even save them after so much time, but I thought it should be noted. 22.214.171.124 07:11, 27 October 2014 (UTC)
I disagree about the poker part. Reading someone's physical tells is just a small part of the game. Theoretically there is a Nash equilibrium for the game, the reason why it hasn't been found is that the amount of ways a deck can be shuffled is astronomical (even if you just count the cards that you use) and you also have to take into account the various betsizes. A near perfect solution for 2 player limit poker has been found by the Cepheus Poker Project: http://poker.srv.ualberta.ca/.
~ Could the description of tic-tac-toe link to xkcd 832 which explains the strategy? 126.96.36.199 13:13, 27 January 2016 (UTC)
Saying that computers are very close to beating top humans as of January 2016 is misleading at best. There is not enough details in the BBC article, but it sounds like the Facebook program has about a 50% chance of beating 5-dan amateurs. In other words, it needs a 4-stone handicap (read: 4 free moves) to have a 50% chance to win against top-level amateurs, to say nothing about professionals. If a robotic team could have a 50% chance to beating Duke University at football (a skilled amateur team), would you say they were very close to being able to consistently beat the Patriots (a top-level professional)? If anything that underestimates the skill difference in Go, but the general point stands. 188.8.131.52 (talk) (please sign your comments with ~~~~)
- How about bearing one of the top players five times in a row and being scheduled to play against the world champion in March? http://www.engadget.com/2016/01/27/google-s-ai-is-the-first-to-defeat-a-go-champion/ Mikemk (talk) 06:18, 28 January 2016 (UTC)
- However DeepMind ranked AlphaGo close to Fan Hui 2P and the distributed version has being at the upper tier of Fan's level. http://www.nature.com/nature/journal/v529/n7587/fig_tab/nature16961_F4.html
- The official games were 5-0 however the unofficial were 3-2. Averaging to 8-2 in favor of AlphaGo.
- Looking at http://www.goratings.org/ Fan Hui is ranked 631, while Lee Sedol 9P, whom is playing in March, is in the top 184.108.40.206.47 06:12 5 February 2016 (UTC)
- Original poster here (sorry, not sure how to sign). Okay, you all are right. Go AI has advanced a lot more than I had understood. I'm still curious how the game against Lee Sedol will go, but that that is even an interesting question shows how much Go AI has improved. 220.127.116.11 (talk) (please sign your comments with ~~~~)
Is the transcript (currently in table format) accessible for blind users? Should it be? 18.104.22.168 10:48, 19 February 2017 (UTC)
At the very least the transcript needs to be fixed so that it factually represents the comic. Jeopardy is in the wrong spot with just a quick glance which is all I have time for here at work. 22.214.171.124 16:58, 24 August 2017 (UTC)
AlphaStar beat Mana pretty decisively, but it was cheating and Mana won the game where it wasn't, and it could only play on a certain map in Protoss vs Protoss. However, that was a while ago. Google dropped AlphaStar on ladder under barcode usernames, and it's been doing rather well... but Serral (one of the world's best players) recently beat it pretty decisively. 126.96.36.199 01:49, 19 September 2019 (UTC)
Who is Ken Jennings?
I feel like it should be relatively easy to make a computer program that can learn the rules of Mao without knowing them to begin with. There has to be some feedback: a player gets penalties if he breaks the rules. This can be used to write a self-learning algorithm.
- In case anyone is checking up on the AlphaStar thing: AlphaStar definitely plays at a high human level in Starcraft 2 now (2020), without doing much that seems 'humanly impossible' (ie, like cheating, as it did during the MaNa matchup), but it's in relatively limited maps, and it not only loses fairly regularly if not most of the time to top-ranked humans like Serral, it also loses to essentially random grab-bags of very good players like Lowko on occasion. Like, Lowko's much better than me but he's not tournament-level good and he beat it.
- Also, technically, AlphaStar isn't even *one* program. It's an ensemble of many programs, each one specific to a different SC2 race and specializing in different strategies. Maybe if there were a 'seamless' amalgam where it were 'choosing' a strategy it could be arguably one program, but it's literally a totally separately trained neural network for each 'agent'.
- Furthermore, when you watch it play sometimes it does extremely stupid things like trap its tanks in its own base. SC2 is, at least this year, still a human endeavor at high tiers.