←back to thread

548 points tifa2up | 1 comments | | HN request time: 0s | source
Show context
jascha_eng ◴[] No.45645905[source]
I have a RAG setup that doesn't work on documents but other data points that we use for generation (the original data is call recordings but it is heavily processed to just a few text chunks). Instead of a reranker model we do vector search and then simply ask GPT-5 in an extra call which of the results is the most relevant to the input question. Is there an advantage to actual reranker models rather than using a generic LLM?
replies(2): >>45645956 #>>45649058 #
1. alansaber ◴[] No.45649058[source]
I think you should do both in parallel, rather than sequentially. Main reason is vector scoring could cut off something that an LLM will score as relevant