←back to thread

129 points xnx | 1 comments | | HN request time: 0.202s | source
Show context
xnx ◴[] No.45159332[source]
Interestingly, per the recent Google antitrust ruling documents, AI mode is extra fast because of a special FastSearch index: https://x.com/Marie_Haynes/status/1963031598829314161
replies(1): >>45159523 #
cj ◴[] No.45159523[source]
Gemini in general is extremely fast, compared to ChatGPT 5 Thinking.

It also seems to excel at things ChatGPT 5 Thinking isn't good at. Simple things like "Here's a screenshot of text, please transcribe it" - ChatGPT 5 Thinking will spend 2 minutes and still get the results wrong, while Gemini Pro will spend 20-30 seconds and transcribe everything perfectly.

Obviously that's just 1 use case, but as someone who previously used ChatGPT exclusively, I'm increasingly impressed by Gemini the more I use it. Mainly due to the much faster thinking times that seem to provide equal or better results than GTP 5 Thinking.

replies(3): >>45162490 #>>45165112 #>>45171443 #
GlitchInstitute ◴[] No.45162490[source]
Gemini is very fast because it runs on TPUsV7 mostly
replies(1): >>45165376 #
1. sigmoid10 ◴[] No.45165376[source]
It is definitely because it's a smaller model. TPUv7 has ~10% lower flops at FP8 and 33% lower memory bandwidth than Nvidia Blackwell cards. Add CUDA to the comparison and they'll probably be even worse at real world utilization. Grok is already running on Blackwell cards and although there's little info on GPT5, I doubt they are behind.