←back to thread

548 points tifa2up | 1 comments | | HN request time: 0s | source
Show context
jascha_eng ◴[] No.45645905[source]
I have a RAG setup that doesn't work on documents but other data points that we use for generation (the original data is call recordings but it is heavily processed to just a few text chunks). Instead of a reranker model we do vector search and then simply ask GPT-5 in an extra call which of the results is the most relevant to the input question. Is there an advantage to actual reranker models rather than using a generic LLM?
replies(2): >>45645956 #>>45649058 #
tifa2up ◴[] No.45645956[source]
OP here. rerankers are finetuned small models, they're cheap and very fast compared to an additional GPT-5 call.
replies(1): >>45646428 #
1. jascha_eng ◴[] No.45646428[source]
It's an async process in my case (custom deep research like) so speed is not that critical