←back to thread

426 points benchmarkist | 1 comments | | HN request time: 0.207s | source
Show context
danpalmer ◴[] No.42179527[source]
I'm not sure if they're comparing apples to apples on the latency here. There are roughly three parts to the latency: the throughput of the context/prompt, the time spent queueing for hardware access, and the other standard API overheads (network, etc).

From what I understand, several, maybe all, of the comparison services are not based on provisioned capacity, which means that the measurements include the queue time. For LLMs this can be significant. The Cerebras number on the other hand almost certainly doesn't have some unbounded amount of queue time included, as I expect they had guaranteed hardware access.

The throughput here is amazing, but to get that throughput at a good latency for end-users means over-provisioning, and it's unclear what queueing will do to this. Additionally, does that latency depend on the machine being ready with the model, or does that include loading the model if necessary? If using a fine-tuned model does this change the latency?

I'm sure it's a clear win for batch workloads where you can keep Cerebras machines running at 100% utilisation and get 1k tokens/s constantly.

replies(1): >>42179788 #
qeternity ◴[] No.42179788[source]
Everyone presumes this is under ideal conditions...and it's incredible.

It's bs=1. At 1,000 t/s. Of a 405B parameter model. Wild.

replies(3): >>42179907 #>>42180069 #>>42191042 #
1. ryao ◴[] No.42191042[source]
They claim it is with between 8 and 20 users:

https://x.com/draecomino/status/1858998347090325846

That said, they appear to be giving the per user performance.