←back to thread

1303 points serjester | 1 comments | | HN request time: 0.451s | source
Show context
gapeslape ◴[] No.42956276[source]
In my mind, Gemini 2.0 changes everything because of the incredibly long context (2M tokens on some models), while having strong reasoning capabilities.

We are working on compliance solution (https://fx-lex.com) and RAG just doesn’t cut it for our use case. Legislation cannot be chunked if you want the model to reason well about it.

It’s magical to be able to just throw everything into the model. And the best thing is that we automatically benefit from future model improvements along all performance axes.

replies(2): >>42957222 #>>42959708 #
manmal ◴[] No.42957222[source]
Maybe a dumb question, have you tried fine tuning on the corpus, and then adding a reasoning process (like all those R1 distillations)?
replies(1): >>42960569 #
1. gapeslape ◴[] No.42960569[source]
We haven't tried that, we might do that in the future.

My intuition - not based on any research - is that recall should be a lot better from in context data vs. weights in the model. For our use case, precise recall is paramount.