←back to thread

295 points rttti | 1 comments | | HN request time: 0.208s | source
Show context
CGMthrowaway ◴[] No.45119871[source]
Honest feedback - I was really excited when I read the opening. However, I did not come away from this without a greater understanding than I already had.

For reference, my initial understanding was somewhat low: basically I know a) what embedding is basically b) transformers work by matrix multiplication, and c) it's something like a multi-threaded Markov chain generator with the benefit of prior-trained embeddings

replies(8): >>45120114 #>>45120200 #>>45122565 #>>45123711 #>>45125243 #>>45128482 #>>45129469 #>>45134872 #
onename ◴[] No.45120200[source]
Have you checked out this video from 3Blue1Brown that talks bit about transformers?

https://youtu.be/wjZofJX0v4M

replies(3): >>45121968 #>>45125756 #>>45125969 #
1. rhdunn ◴[] No.45125969[source]
There's also various videos by Welch Labs that are very good. -- https://www.youtube.com/@WelchLabsVideo/videos