[Computer-go] Computer-go Digest, Vol 12, Issue 79
Aja
ajahuang at gmail.com
Wed Jan 26 07:42:04 PST 2011
Hi Hendrik,
> that's a good question. At least for the LGR policy without forgetting
> (https://webdisk.lclark.edu/drake/publications/drake-icga-2009.pdf), only
> using the first appearance of a reply did not significantly differ in
> performance.
Thanks for your explanation. Yes, my experiment indicates that the
playing strengh is almost the same.
> It's only a few lines of code, test it and see if it makes a difference
> for your playout policy and program architecture. Stronger playout
> policies than Orego's will have different interactions with LGRF. You
> could even try saving several sets of replies per intersection, for the
> first, second, third appearance of the previous move in a playout, in the
> hope of capturing certain tactical situations with sacrifices. But I don't
> expect much.
Indeed. My plan is to generalize this scheme with more strict conditions.
Maybe we even can combine LGRF with the information of RAVE (inspired from
Arpad Rimmel's works). If the learning works well, it should fix a lot of
errors in my rules of the playout features. This might be a way to make the
playouts to learn how to play correct semeai moves.
Aja
More information about the Computer-go
mailing list