Nov. 27, 2023, 12:42 p.m. | Andrej Baranovskij

Andrej Baranovskij www.youtube.com

I explain the implementation of the pipeline to process invoice data from PDF documents. The data is loaded into Chroma DB's vector store. Through LangChain API, the data from the vector store is ready to be consumed by LLM as part of the RAG infrastructure.

GitHub repo:
https://github.com/katanaml/llm-ollama-invoice-cpu

0:00 Intro
1:19 Libs
1:54 Ingest data with ChromaDB
6:17 Main script
6:59 Pipeline with LangChain
9:00 Testing and Summary

CONNECT:
- Subscribe to this YouTube channel
- Twitter: https://twitter.com/andrejusb
- LinkedIn: …

api chroma chromadb data documents easy github github repo implementation infrastructure intro invoice invoice processing langchain llm part pdf pipeline process processing rag store through tutorial vector vector store

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US