←back to thread

1303 points serjester | 1 comments | | HN request time: 0.217s | source
Show context
Havoc ◴[] No.42953438[source]
Been toying with the flash model. Not the top model, but think it'll see plenty use due to the details. Wins on things other than top of benchmark logs

* Generous free tier

* Huge context window

* Lite version feels basically instant

However

* Lite model seems more prone to repeating itself / looping

* Very confusing naming e.g. {model}-latest worked for 1.5 but now its {model}-001? The lite has a date appended, the non-lite does not. Then there is exp and thinking exp...which has a date. wut?

replies(1): >>42953462 #
ai-christianson ◴[] No.42953462[source]
> * Huge context window

But how well does it actually handle that context window? E.g. a lot of models support 200K context, but the LLM can only really work with ~80K or so of it before it starts to get confused.

replies(5): >>42953514 #>>42953536 #>>42953554 #>>42953762 #>>42955202 #
1. summerlight ◴[] No.42953514[source]
My experience is that Gemini works relatively well on larger contexts. Not perfect, but more reliable.