Has anyone compiled or trained weights with the characteristics of classic players? Games trained to follow the style of the Yasui school, for example; or Go Seigen.
Take care
Flavoured weights
-
Ferran
- Lives in gote
- Posts: 664
- Joined: Mon Apr 22, 2019 1:04 am
- Rank: OGS ddk
- GD Posts: 0
- KGS: Ferran
- IGS: Ferran
- OGS: Ferran
- Has thanked: 177 times
- Been thanked: 121 times
Re: Flavoured weights
Someone is already doing it with chess. I've only skimmed the abstracts, so far, but about 2000 games to train. We have as many games of some of the great players...
viewtopic.php?f=8&t=18003&view=unread#unread
Take care
viewtopic.php?f=8&t=18003&view=unread#unread
Take care
一碁一会
- MikeKyle
- Lives with ko
- Posts: 205
- Joined: Wed Jul 26, 2017 2:27 am
- Rank: EGF 2k
- GD Posts: 0
- KGS: MKyle
- Has thanked: 49 times
- Been thanked: 36 times
Re: Flavoured weights
I always assumed that even the most frequently playing pro with the longest career would not play enough games for an ai built using current techniques to learn from. Pleased to hear that I may be wrong.
I've often wondered about making a takemiya-style centre-oriented bot by training using modified rules. Perhaps the rules could give an extra point to the owner of tengen? Or apply bonuses to a wider range of central points? Maybe the player with the most stones and or territory above the 4th line gets a bonus?
I would be interested to see how much each incentive in the rules would make the bot play differently?
I've often wondered about making a takemiya-style centre-oriented bot by training using modified rules. Perhaps the rules could give an extra point to the owner of tengen? Or apply bonuses to a wider range of central points? Maybe the player with the most stones and or territory above the 4th line gets a bonus?
I would be interested to see how much each incentive in the rules would make the bot play differently?
-
Bill Spight
- Honinbo
- Posts: 10905
- Joined: Wed Apr 21, 2010 1:24 pm
- Has thanked: 3651 times
- Been thanked: 3373 times
Re: Flavoured weights
Learning by self play introduces path dependency. I would expect that today's AI bots could learn how to predict the plays of a specific player in short order, such that the predictions would generalize so that predictions in positions that the player never met are not just random. OC, the level of play of the bot at that point would be rather weak. Starting from that point the bot could be trained by self play to reach superhuman strength. Because of path dependency and, I believe, the likelihood that in most go positions there is more than one optimal play, I imagine that certain recognizable aspects of the player's style would be preserved in the bot's play.MikeKyle wrote:I always assumed that even the most frequently playing pro with the longest career would not play enough games for an ai built using current techniques to learn from. Pleased to hear that I may be wrong.
The Adkins Principle:
At some point, doesn't thinking have to go on?
— Winona Adkins
Visualize whirled peas.
Everything with love. Stay safe.
At some point, doesn't thinking have to go on?
— Winona Adkins
Visualize whirled peas.
Everything with love. Stay safe.
-
Ferran
- Lives in gote
- Posts: 664
- Joined: Mon Apr 22, 2019 1:04 am
- Rank: OGS ddk
- GD Posts: 0
- KGS: Ferran
- IGS: Ferran
- OGS: Ferran
- Has thanked: 177 times
- Been thanked: 121 times
Re: Flavoured weights
It might be interesting to see what happens with players known to have changed styles. Would AI find a common theme?
Take care.
Take care.
一碁一会
-
hakuseki
- Dies with sente
- Posts: 100
- Joined: Thu Oct 29, 2020 6:18 am
- Rank: KGS 2 dan
- GD Posts: 0
- KGS: hakuseki
- Been thanked: 15 times
Re: Flavoured weights
Training an AI to play like a (pre-AI era) 9-dan pro seems like a hard problem. If the AI is a pure policy with no search, then it is likely to misread many situations that the pro would read accurately. But the more you use search to correct the reading errors, the more it will tend to filter out the human-style moves that the value model judges suboptimal.
Maybe an existing AI like KataGo could be run for a fixed number of iterations (e.g. 1000 visits) and its move evaluations could be used as an input feature for a new model trained to predict the human moves.
Maybe an existing AI like KataGo could be run for a fixed number of iterations (e.g. 1000 visits) and its move evaluations could be used as an input feature for a new model trained to predict the human moves.