←back to thread

568 points PaulHoule | 1 comments | | HN request time: 0.22s | source
1. awaymazdacx5 ◴[] No.44490962[source]
Having token embeddings with diffusion models, for 16x16 transformer encoding. Image is tokenized before transformers compile it. If decomposed virtualization modulates according to a diffusion model.