> I don’t know. After the model has been created (trained), I’m pretty sure that generating embeddings is much less computationally intensive than generating text.
An embedding is generated after a single pass through the model, so functionally it's the equivalent of generating a single token from an text generation model.
replies(2):