←back to thread

251 points slyall | 3 comments | | HN request time: 0.21s | source
1. hollerith ◴[] No.42063807[source]
The deep learning boom caught deep-learning researchers by surprise because deep-learning researchers don't understand their craft well enough to predict essential properties of their creations.

A model is grown, not crafted like a computer program, which makes it hard to predict. (More precisely, a big growth phase follows the crafting phase.)

replies(2): >>42063891 #>>42065828 #
2. lynndotpy ◴[] No.42063891[source]
I was a deep learning researcher. The problem is that accuracy (+ related metrics) were prioritized in research and funding. Factors like interpretability, extrapolation, efficiency, or consistency were not prioritized, but were clearly important before being implemented.

Dall-E was the only big surprising consumer model-- 2022 saw a sudden huge leap from "txt2img is kind of funny" to "txt2img is actually interesting". I would have assumed such a thing could only come in 2030 or earlier. But deep learning is full of counterintuitive results (like the NFL theorem not mattering, or ReLU being better than sigmoid).

But in hindsight, it was naive to think "this does not work yet" would get in the way of the products being sold and monetized.

3. nxobject ◴[] No.42065828[source]
I'm still very taken aback by how far we've been able to take prompting as somehow our universal language to communicate with AI of choice.