←back to thread

469 points samuelstros | 7 comments | | HN request time: 1.003s | source | bottom
1. ebzlo ◴[] No.44998638[source]
Yes technically it is RAG, but a lot of the community is associating RAG with vector search specifically.
replies(1): >>44998673 #
2. BoorishBears ◴[] No.44998656[source]
If you want to be really stringent, RAG originally referred to going from user query to retrieving information directly based on the query then passing it to an LLM: With CC the LLM is taking the raw user query then crafting its own searches

But realistically lots of RAG systems have LLM calls interleaved for various reasons, so what they probably mean it not doing the usual chunking + embeddings thing.

replies(1): >>44998746 #
3. dingnuts ◴[] No.44998673[source]
it does? why? the term RAG as I understand it leaves the methodology for retrieval vague so that different techniques can be used depending on the, er, context.. which makes a lot more sense to me
replies(1): >>44998795 #
4. theptip ◴[] No.44998746[source]
Yeah, TFA clearly explains their point. They mean RAG=vector search, and contrast this with tool calling (eg Grep).
5. koakuma-chan ◴[] No.44998795{3}[source]
> why?

Hype. There's nothing wrong with using, e.g., full-text search for RAG.

6. nuwandavek ◴[] No.44999808[source]
(blogpost author here) You're right! I did make the distinction in an earlier draft, but decided to use "RAG" interchangeably with vector search, as it is popularly known today in code-gen systems. I'd probably go back to the previous version too.

But I do think there is a qualitative different between getting candidates and adding them to context before generating (retrieval augmented generation) vs the LLM searching for context till it is satisfied.

7. faangguyindia ◴[] No.45002603[source]
It doesn't use RAG is the most obvious way, like taking whole text/code and generating embeddings and performing vector search on it.