←back to thread

235 points tosh | 3 comments | | HN request time: 0.605s | source
Show context
xanderlewis ◴[] No.40214349[source]
> Stripped of anything else, neural networks are compositions of differentiable primitives

I’m a sucker for statements like this. It almost feels philosophical, and makes the whole subject so much more comprehensible in only a single sentence.

I think François Chollet says something similar in his book on deep learning: one shouldn’t fall into the trap of anthropomorphising and mysticising models based on the ‘neural’ name; deep learning is simply the application of sequences of operations that are nonlinear (and hence capable of encoding arbitrary complexity) but nonetheless differentiable and so efficiently optimisable.

replies(12): >>40214569 #>>40214829 #>>40215168 #>>40215198 #>>40215245 #>>40215592 #>>40215628 #>>40216343 #>>40216719 #>>40216975 #>>40219489 #>>40219752 #
andoando ◴[] No.40214569[source]
What does "differentiable primitives" mean here?
replies(4): >>40214623 #>>40214658 #>>40215206 #>>40215221 #
1. CobrastanJorji ◴[] No.40214623[source]
Continuous mathematical functions which have derivatives.
replies(1): >>40214677 #
2. ◴[] No.40214677[source]
replies(1): >>40215010 #
3. ◴[] No.40215010[source]