Oct. 22, 2023, 1:49 p.m. | Andrej Baranovskij

Andrej Baranovskij www.youtube.com

I explained how to set up local LLM RAG to process invoice data with Llama2 13B. Based on my experiments, Llama2 13B works better with tabular data compared to Mistral 7B model. This example presents a production LLM RAG setup with Weaviate database for vector embeddings, Haystack for LLM API, and Llama.cpp to run Llama2 13b on a local CPU.

GitHub repo:
https://github.com/katanaml/llm-rag-invoice-cpu/

0:00 Intro
1:10 Examples
6:35 Setup
8:20 Config
9:45 Weaviate Docker
10:05 Data Ingest Code
10:30 Inference …

13b cpp cpu data database data processing embeddings example explained haystack invoice llama llama2 llm llm rag mistral mistral 7b process processing production rag set setup tabular tabular data vector vector embeddings weaviate

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US