←back to thread

566 points PaulHoule | 1 comments | | HN request time: 0.243s | source
1. awaymazdacx5 ◴[] No.44490962[source]
Having token embeddings with diffusion models, for 16x16 transformer encoding. Image is tokenized before transformers compile it. If decomposed virtualization modulates according to a diffusion model.