Hello list,
I am a Master's student currently working on his thesis about certain
aspects of Monte-Carlo Go. I would like to pose a question concerning
the literature - I hope some of you can help me out!
My problem is that I can't find many papers about learning of MC playout
policies, in particular patterns. A lot of programs seem to be using
Mogo's 3x3 patterns, which have been handcoded, or some variation
thereof. A lot of people have tried some form of pattern learning, but
mostly to directly predict expert moves it seems, not explicitly
optimizing the patterns for their function in an MC playout policy.
Actually, I am only aware of "Computing Elo Ratings of Move Patterns in
the Game of Go", where patterns have been learned from pro moves, but
then also successfully used in an MC playout policy; and "Monte Carlo
Simulation Balancing".
Considering the huge impact local patterns have had on the success of MC
programs, I would have expected more attention towards automatically
learning and weighting them specifically for MC playouts. There is no
reason why patterns which are good for predicting experts should also be
good for guaranteeing diverse, balanced playout distributions. Have I
missed something?
Or how did your program come to its patterns? I'd be interested. Did you
maybe even try learning something else than patterns for your playout
policy?
cheers,
Hendrik
_______________________________________________
Computer-go mailing list
Compu...@dvandva.org
http://dvandva.org/cgi-bin/mailman/listinfo/computer-go
--
You received this message because you are subscribed to the Google Groups "Computer Go Archive" group.
To post to this group, send email to
computer-...@googlegroups.com.
To unsubscribe from this group, send email to
computer-go-arc...@googlegroups.com.
For more options, visit this group at
http://groups.google.com/group/computer-go-archive?hl=en.