[Computer-go] Source code (Was: Reducing network size? (Was: AlphaGo Zero))
sligocki at gmail.com
Thu Oct 26 15:33:51 PDT 2017
On Thu, Oct 26, 2017 at 2:02 PM, Gian-Carlo Pascutto <gcp at sjeng.org> wrote:
> On 26-10-17 15:55, Roel van Engelen wrote:
> > @Gian-Carlo Pascutto
> > Since training uses a ridiculous amount of computing power i wonder
> > if it would be useful to make certain changes for future research,
> > like training the value head with multiple komi values
> > <https://arxiv.org/pdf/1705.10701.pdf>
> Given that the game data will be available, it will be trivial for
> anyone to train a different network architecture on the result and see
> if they get better results, or a program that handles multiple komi
> values, etc.
> The problem is getting the *data*, not the training.
But the data should be different for different komi values, right?
Iteratively producing self-play games and training with the goal of
optimizing for komi 7 should converge to a different optimal player than
optimizing for komi 5. But maybe having high quality data for komi 7 will
still save a lot of the work for training a komi 5 (or komi agnostic)
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Computer-go