This is what local LLMs need—being treated like first-class citizens by the companies that make them.
That said, the first graph is misleading about the number of H100s required to run DeepSeek r1 at FP16. The model is FP8.
replies(2):
That said, the first graph is misleading about the number of H100s required to run DeepSeek r1 at FP16. The model is FP8.