[Computer-go] mini-max with Policy and Value network
Erik van der Werf
erikvanderwerf at gmail.com
Mon May 22 06:46:37 PDT 2017
On Mon, May 22, 2017 at 11:27 AM, Erik van der Werf <
erikvanderwerf at gmail.com> wrote:
> On Mon, May 22, 2017 at 10:08 AM, Gian-Carlo Pascutto <gcp at sjeng.org>
>> ... This heavy pruning
>> by the policy network OTOH seems to be an issue for me. My program has
>> big tactical holes.
> Do you do any hard pruning? My engines (Steenvreter,Magog) always had a
> move predictor (a.k.a. policy net), but I never saw the need to do hard
> pruning. Steenvreter uses the predictions to set priors, and it is very
> selective, but with infinite simulations eventually all potentially
> relevant moves will get sampled.
Oh, haha, after reading Brian's post I guess I misunderstood :-)
Anyway, LMR seems like a good idea, but last time I tried it (in Migos) it
did not help. In Magog I had some good results with fractional depth
reductions (like in Realization Probability Search), but it's a long time
ago and the engines were much weaker then...
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Computer-go