[Computer-go] Computer-go Digest, Vol 12, Issue 79

Aja ajahuang at gmail.com
Wed Jan 26 07:42:04 PST 2011


   Hi Hendrik,

> that's a good question. At least for the LGR policy without forgetting 
> (https://webdisk.lclark.edu/drake/publications/drake-icga-2009.pdf), only 
> using the first appearance of a reply did not significantly differ in 
> performance.

   Thanks for your explanation. Yes, my experiment indicates that the 
playing strengh is almost the same.

> It's only a few lines of code, test it and see if it makes a difference 
> for your playout policy and program architecture. Stronger playout 
> policies than Orego's will have different interactions with LGRF. You 
> could even try saving several sets of replies per intersection, for the 
> first, second, third appearance of the previous move in a playout, in the 
> hope of capturing certain tactical situations with sacrifices. But I don't 
> expect much.

   Indeed. My plan is to generalize this scheme with more strict conditions. 
Maybe we even can combine LGRF with the information of RAVE (inspired from 
Arpad Rimmel's works). If the learning works well, it should fix a lot of 
errors in my rules of the playout features. This might be a way to make the 
playouts to learn how to play correct semeai moves.

  Aja




More information about the Computer-go mailing list