←back to thread

612 points meetpateltech | 2 comments | | HN request time: 0s | source
Show context
leonidasv ◴[] No.42952286[source]
That 1M tokens context window alone is going to kill a lot of RAG use cases. Crazy to see how we went from 4K tokens context windows (2023 ChatGPT-3.5) to 1M in less than 2 years.
replies(6): >>42952393 #>>42952519 #>>42952569 #>>42954277 #>>42958220 #>>42975332 #
1. torginus ◴[] No.42954277[source]
That's not really my experience. Error rate goes up the more stuff you cram into the context, and processing gets both slower and more expensive with the amount of input tokens.

I'd say it makes sense to do RAG even if your stuff fits into context comfortably.

replies(1): >>42954311 #
2. lamuswawir ◴[] No.42954311[source]
Try exp-1206. That thing works on large context.