←back to thread

1303 points serjester | 2 comments | | HN request time: 0.442s | source
Show context
Havoc ◴[] No.42953438[source]
Been toying with the flash model. Not the top model, but think it'll see plenty use due to the details. Wins on things other than top of benchmark logs

* Generous free tier

* Huge context window

* Lite version feels basically instant

However

* Lite model seems more prone to repeating itself / looping

* Very confusing naming e.g. {model}-latest worked for 1.5 but now its {model}-001? The lite has a date appended, the non-lite does not. Then there is exp and thinking exp...which has a date. wut?

replies(1): >>42953462 #
ai-christianson ◴[] No.42953462[source]
> * Huge context window

But how well does it actually handle that context window? E.g. a lot of models support 200K context, but the LLM can only really work with ~80K or so of it before it starts to get confused.

replies(5): >>42953514 #>>42953536 #>>42953554 #>>42953762 #>>42955202 #
1. asadm ◴[] No.42953554[source]
it works REALLY well. I have used it to dump many references codes and then help me write a new modules etc. I have gone up to 200k tokens I think with no problems in recall.
replies(1): >>42953640 #
2. ai-christianson ◴[] No.42953640[source]
Awesome. Models that can usefully leverage such large context windows are rare at this point.

Something like this opens up a lot of use cases.