March 7, 2024, 10:32 p.m. | /u/Jeff_1987

Natural Language Processing www.reddit.com

I think I fundamentally misunderstand how RAGs generate output. Everything I've read online seems to suggest that relevant documents are retrieved from a database (with relevance to the prompt calculated based on their embeddings), after which the text of the relevant documents are appended to the prompt's text before being pushed to the LLM for generation.


For some reason, I had it in my mind that the embeddings of the retrieved documents, along with that of the prompt, could be …

database documents embeddings everything generate languagetechnology llm prompt rags text the prompt think work

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US