Flavoured weights

For discussing go computing, software announcements, etc.
Post Reply
Ferran
Lives in gote
Posts: 664
Joined: Mon Apr 22, 2019 1:04 am
Rank: OGS ddk
GD Posts: 0
KGS: Ferran
IGS: Ferran
OGS: Ferran
Has thanked: 177 times
Been thanked: 121 times

Flavoured weights

Post by Ferran »

Has anyone compiled or trained weights with the characteristics of classic players? Games trained to follow the style of the Yasui school, for example; or Go Seigen.

Take care
一碁一会
Ferran
Lives in gote
Posts: 664
Joined: Mon Apr 22, 2019 1:04 am
Rank: OGS ddk
GD Posts: 0
KGS: Ferran
IGS: Ferran
OGS: Ferran
Has thanked: 177 times
Been thanked: 121 times

Re: Flavoured weights

Post by Ferran »

Someone is already doing it with chess. I've only skimmed the abstracts, so far, but about 2000 games to train. We have as many games of some of the great players...

viewtopic.php?f=8&t=18003&view=unread#unread

Take care
一碁一会
User avatar
MikeKyle
Lives with ko
Posts: 205
Joined: Wed Jul 26, 2017 2:27 am
Rank: EGF 2k
GD Posts: 0
KGS: MKyle
Has thanked: 49 times
Been thanked: 36 times

Re: Flavoured weights

Post by MikeKyle »

I always assumed that even the most frequently playing pro with the longest career would not play enough games for an ai built using current techniques to learn from. Pleased to hear that I may be wrong.

I've often wondered about making a takemiya-style centre-oriented bot by training using modified rules. Perhaps the rules could give an extra point to the owner of tengen? Or apply bonuses to a wider range of central points? Maybe the player with the most stones and or territory above the 4th line gets a bonus?
I would be interested to see how much each incentive in the rules would make the bot play differently?
Bill Spight
Honinbo
Posts: 10905
Joined: Wed Apr 21, 2010 1:24 pm
Has thanked: 3651 times
Been thanked: 3373 times

Re: Flavoured weights

Post by Bill Spight »

MikeKyle wrote:I always assumed that even the most frequently playing pro with the longest career would not play enough games for an ai built using current techniques to learn from. Pleased to hear that I may be wrong.
Learning by self play introduces path dependency. I would expect that today's AI bots could learn how to predict the plays of a specific player in short order, such that the predictions would generalize so that predictions in positions that the player never met are not just random. OC, the level of play of the bot at that point would be rather weak. Starting from that point the bot could be trained by self play to reach superhuman strength. Because of path dependency and, I believe, the likelihood that in most go positions there is more than one optimal play, I imagine that certain recognizable aspects of the player's style would be preserved in the bot's play.
The Adkins Principle:
At some point, doesn't thinking have to go on?
— Winona Adkins

Visualize whirled peas.

Everything with love. Stay safe.
Ferran
Lives in gote
Posts: 664
Joined: Mon Apr 22, 2019 1:04 am
Rank: OGS ddk
GD Posts: 0
KGS: Ferran
IGS: Ferran
OGS: Ferran
Has thanked: 177 times
Been thanked: 121 times

Re: Flavoured weights

Post by Ferran »

It might be interesting to see what happens with players known to have changed styles. Would AI find a common theme?
Take care.
一碁一会
hakuseki
Dies with sente
Posts: 100
Joined: Thu Oct 29, 2020 6:18 am
Rank: KGS 2 dan
GD Posts: 0
KGS: hakuseki
Been thanked: 15 times

Re: Flavoured weights

Post by hakuseki »

Training an AI to play like a (pre-AI era) 9-dan pro seems like a hard problem. If the AI is a pure policy with no search, then it is likely to misread many situations that the pro would read accurately. But the more you use search to correct the reading errors, the more it will tend to filter out the human-style moves that the value model judges suboptimal.

Maybe an existing AI like KataGo could be run for a fixed number of iterations (e.g. 1000 visits) and its move evaluations could be used as an input feature for a new model trained to predict the human moves.
Post Reply