←back to thread

365 points lawrenceyan | 9 comments | | HN request time: 0.895s | source | bottom
1. chvid ◴[] No.41877343[source]
But the gigantic synthetic dataset that is used for training is created with plenty of traditional search. So it is all a bit silly but I guess cool none the less ...
replies(3): >>41877359 #>>41877644 #>>41877711 #
2. chvid ◴[] No.41877359[source]
If anything it demonstrates the limits of NN. A human brain can learn based on far fewer examples.
replies(1): >>41882113 #
3. amunozo ◴[] No.41877644[source]
It's a knowledge distillation. You can then use this smaller, more efficient models instead of the larger one.
replies(2): >>41878341 #>>41883075 #
4. msoad ◴[] No.41877711[source]
Searched only once. If this can be applied to other knowledge with this efficiency we're onto something
5. chvid ◴[] No.41878341[source]
Or maybe it is just memorizing a very large number of games.
replies(2): >>41880894 #>>41882390 #
6. azakai ◴[] No.41880894{3}[source]
They address the possibility of memorization in the PDF:

> This effect cannot be explained by memorization since < 1.41% of the initial puzzle board states appear in our training set.

7. jxy ◴[] No.41882113[source]
Nature's evolution algorithm took millions of years to find the architecture and the base model, which then takes decades to be fine tuned to be able to form this opinion.
8. tech_ken ◴[] No.41882390{3}[source]
Seems more like a 'compression' of the large number of games, or even like an approximate 'index' of the database
9. alkonaut ◴[] No.41883075[source]
Is this network smaller than stockfish and by what metric is that?