March 7, 2024, 10:32 p.m. | /u/Jeff_1987

Natural Language Processing www.reddit.com

I think I fundamentally misunderstand how RAGs generate output. Everything I've read online seems to suggest that relevant documents are retrieved from a database (with relevance to the prompt calculated based on their embeddings), after which the text of the relevant documents are appended to the prompt's text before being pushed to the LLM for generation.


For some reason, I had it in my mind that the embeddings of the retrieved documents, along with that of the prompt, could be …

database documents embeddings everything generate languagetechnology llm prompt rags text the prompt think work

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US