I feel the article presents the data selectively in some places. Two examples:
* The article compares Gemini 2.5 Pro Experimental to DeepSeek-R1 in accuracy benchmarks. Then, when the comparison becomes about cost, it compares Gemini 2.0 Flash to DeepSeek-R1.
* In throughput numbers, DeepSeek-R1 is quoted at 24 tok/s. There are half a dozen providers, who give you easily 100+ tok/s and at scale.
There's no doubt that Gemini 2.5 Pro Experimental is a state of the art model. I just think it's very hard to win on every AI front these days.
replies(2):