←back to thread

365 points lawrenceyan | 5 comments | | HN request time: 0.62s | source
Show context
chvid ◴[] No.41877343[source]
But the gigantic synthetic dataset that is used for training is created with plenty of traditional search. So it is all a bit silly but I guess cool none the less ...
replies(3): >>41877359 #>>41877644 #>>41877711 #
1. amunozo ◴[] No.41877644[source]
It's a knowledge distillation. You can then use this smaller, more efficient models instead of the larger one.
replies(2): >>41878341 #>>41883075 #
2. chvid ◴[] No.41878341[source]
Or maybe it is just memorizing a very large number of games.
replies(2): >>41880894 #>>41882390 #
3. azakai ◴[] No.41880894[source]
They address the possibility of memorization in the PDF:

> This effect cannot be explained by memorization since < 1.41% of the initial puzzle board states appear in our training set.

4. tech_ken ◴[] No.41882390[source]
Seems more like a 'compression' of the large number of games, or even like an approximate 'index' of the database
5. alkonaut ◴[] No.41883075[source]
Is this network smaller than stockfish and by what metric is that?