I also found this one - http://senseis.xmp.net/?AlphaGo Second headline talks about the algorithm, soooooooo... Alpha Go, you are busted
Joking aside, since AlphaGo was built on the NEAT principle with neural networks and what not, it proves its superiority to chess - in which chess programs merely store billions of games and just put the move with the highest chance of success.
I suppose what they did was feed the programs millions of pro games and its gradually developed a style to beat them, as @dankenzon suggested.
Stefany93 wrote:I suppose what they did was feed the programs millions of pro games and its gradually developed a style to beat them, as @dankenzon suggested.
Actually not, as far as we're aware. Prior to the original Nature paper they only trained it using KGS games.
Although they may have added pro games I doubt whether there are even one million of them available - GoGod doesn't even have 100,000.
Even if they have a large number it's a good question as to whether the number they have will be significant compared to the total KGS and other online games they may have acquired.
I think the type of games should not matter after a few generations of the network. They games are just a source for the (random) patterns. Even in a KGS game there is a good local move from time to time . The network "learns" how to apply the available patterns optimal. If the number of available pro games would be big enough to use them exclusivly for training the network would just learn a little bit faster.
mumps wrote: Actually not, as far as we're aware. Prior to the original Nature paper they only trained it using KGS games.
Although they may have added pro games I doubt whether there are even one million of them available - GoGod doesn't even have 100,000.
Even if they have a large number it's a good question as to whether the number they have will be significant compared to the total KGS and other online games they may have acquired.
Jon
Haven't I read somewhere that AlphaGo ran 24/7 for some time already? It played against itself many times, perhaps indeed millions of games.
Stefany93 wrote:I suppose what they did was feed the programs millions of pro games and its gradually developed a style to beat them, as @dankenzon suggested.
Actually not, as far as we're aware. Prior to the original Nature paper they only trained it using KGS games.
Although they may have added pro games I doubt whether there are even one million of them available - GoGod doesn't even have 100,000.
Even if they have a large number it's a good question as to whether the number they have will be significant compared to the total KGS and other online games they may have acquired.
Jon
I guess they had to be pro games, even in KGS in order for the AlphaGo to get stronger.
mumps wrote: Actually not, as far as we're aware. Prior to the original Nature paper they only trained it using KGS games.
Although they may have added pro games I doubt whether there are even one million of them available - GoGod doesn't even have 100,000.
Even if they have a large number it's a good question as to whether the number they have will be significant compared to the total KGS and other online games they may have acquired.
Jon
Haven't I read somewhere that AlphaGo ran 24/7 for some time already? It played against itself many times, perhaps indeed millions of games.
That's what I was thinking - it will need millions of games to develop its neural networks to a point where it will be able to beat all the pros.
Even if it generates them itself, or gets them online.
My assumption is that it got markedly stronger over the year through self-play, not through training its policy network with more pro games. For one thing, a lot of its best moves were unusual from a pro perspective.
dfan wrote:My assumption is that it got markedly stronger over the year through self-play, not through training its policy network with more pro games. For one thing, a lot of its best moves were unusual from a pro perspective.
That the thing - in order for it to play unusual moves, it has to learn what "usual" moves the pros play to surprise them.
It doesn't seem to me that that has to be the case. If it's playing the best moves, some of them will happen to be unusual. I don't think that playing unusual moves was a goal in itself.
I also found this one - http://senseis.xmp.net/?AlphaGo Second headline talks about the algorithm, soooooooo... Alpha Go, you are busted
Joking aside, since AlphaGo was built on the NEAT principle with neural networks and what not, it proves its superiority to chess - in which chess programs merely store billions of games and just put the move with the highest chance of success.
I suppose what they did was feed the programs millions of pro games and its gradually developed a style to beat them, as @dankenzon suggested.
Millions of pro games? I will be surprised if at all the history there are 500 thousand! By the way, AlphaGo learning experience was feeded by amateur games instead. It was used as a reference model to learn the game and find its own solutions.
So, he didn't developed a style to beat them. It has the goal of finding the moves that give the higher % of a sure win: it doesn't care about big wins but for a 100% sure win even if it means winning by only half point.
mumps wrote: Actually not, as far as we're aware. Prior to the original Nature paper they only trained it using KGS games.
Although they may have added pro games I doubt whether there are even one million of them available - GoGod doesn't even have 100,000.
Even if they have a large number it's a good question as to whether the number they have will be significant compared to the total KGS and other online games they may have acquired.
Jon
Haven't I read somewhere that AlphaGo ran 24/7 for some time already? It played against itself many times, perhaps indeed millions of games.
dfan wrote:My assumption is that it got markedly stronger over the year through self-play, not through training its policy network with more pro games. For one thing, a lot of its best moves were unusual from a pro perspective.
That the thing - in order for it to play unusual moves, it has to learn what "usual" moves the pros play to surprise them.
Again: he didn't learn nothing from pro games. It was feeded by amateur games. It just practiced and practiced and practiced and compared results until it came to the conclusion that some moves are more prone to have a higher probability of winning that others.
It doesn't care about usual or not. It simply evaluates and practices possibilities without never getting tired and with unlimited chance to make tests.