Nov. 6, 2023, 12:48 p.m. | Andrej Baranovskij

Andrej Baranovskij www.youtube.com

I explain how to get structured JSON output from LLM RAG running using Haystack API on top of Llama.cpp. Vector embeddings are stored in Weaviate database, the same as in my previous video. When extracting data, a structured JSON response is preferred because we are not interested in additional descriptions.

Invoice Data Processing with Llama2 13B LLM RAG on Local CPU [Weaviate, Llama.cpp, Haystack]:
https://www.youtube.com/watch?v=XuvdgCuydsM

GitHub repo:
https://github.com/katanaml/llm-rag-invoice-cpu

0:00 Intro
0:55 Prompts
5:18 Summary

CONNECT:
- Subscribe to this YouTube …

api cpp cpu data database embeddings haystack json llama llm llm rag rag running vector vector embeddings video weaviate

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Robotics Technician - 3rd Shift

@ GXO Logistics | Perris, CA, US, 92571