Jan. 8, 2024, 8:19 a.m. | Andrej Baranovskij

Andrej Baranovskij www.youtube.com

This is Sparrow, our open-source solution for document processing with local LLMs. I'm running local Starling LLM with Ollama. I explain how to get structured JSON output with LlamaIndex and dynamic Pydantic class. This helps to implement the use case of data extraction from invoice documents. The solution runs on the local machine, thanks to Ollama. I'm using a MacBook Air M1 with 8GB RAM.

Sparrow GitHub repo:
https://github.com/katanaml/sparrow

0:00 Intro
0:42 Example
2:29 Config
3:08 RAG with Sparrow and …

case class data data extraction document document processing documents dynamic extraction invoice json llamaindex llm llms ollama processing pydantic running solution

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

RL Analytics - Content, Data Science Manager

@ Meta | Burlingame, CA

Research Engineer

@ BASF | Houston, TX, US, 77079