[Computer-go] Crazy Stone is back
valkyria at phmp.se
valkyria at phmp.se
Tue Mar 6 01:48:33 PST 2018
I think you misunderstood what I wrote,
if perfect play on 9x9 is 6000 Elo, then if the value function is 3000
Elo and MC eval is 2000 Elo with 1 second thinking time then it might
be that the combination of a value function and MC eval ends up being
2700 Elo. It could also be that it ends up at 3200 Elo.
I personally believe MC eval alone can be very strong so it might be
that the capacity of a neural network is not enough to replace MC eval.
When I wrote "converge to perfection" it was not to claim that the Zero
approach reaches perfection, just that it get stronger over time.
The interesting question is if old school MC evaluation can fill up the
knowledge gaps of the value function.
For my own Odin project I am not working on the MC evaluation currently,
since only when I a have final neural network solution can I see which
weaknesses I need to fix.
On 2018-03-05 21:12, Hideki Kato wrote:
> DCNNs are not magic but just non-linear continuous function
> approximators with finite freedom and we can provide up to
> 10^8 samples (board positions) in practice.
> Why do most people believe VN can approximate (perfect or
> near perfect) value function? What do they estimate the
> complexity of the value function for 19x19 Go?
> valkyria at phmp.se: <aa3cca138c40cbd620700cc36e950aed at phmp.se>:
>> My guess is that there is some kind of threshold depending on the
>> relative strength of MC eval and the value function of the NN.
>> If the value function is stronger than MC eval I would guess MCEval
>> turns into a bad noisy feature with little benefit.
>> Depending on how strong MC eval is this threshold is probably very
>> different between engines. Also i can imagine that NN value function
>> have some gaping holes in its knowledge that even simple MC eval can
>> patch up. Probably true for supervised learning where training data
>> probably has a lot of holes since bad moves are not in the data.
>> The Zero approach is different because it should converge to
>> in the limit, thus overcome any weaknesses of the value function early
>> on. At least in theory.
>> On 2018-03-05 14:04, Gian-Carlo Pascutto wrote:
>>> On 5/03/2018 12:28, valkyria at phmp.se wrote:
>>>> Remi twittered more details here (see the discussion with gghideki:
>>> Thank you. So Remi gave up on rollouts as well. Interesting
>>> of opinion" there with Zen.
>>> Last time I tested this in regular Leela, playouts were beneficial,
>>> this was before combined value+policy nets and much more training
>>> was available. I do not know what the current status would be.
>> Computer-go mailing list
>> Computer-go at computer-go.org
More information about the Computer-go