Sept. 8, 2023, 5:27 p.m. | 1littlecoder

1littlecoder www.youtube.com

Learn how to use Llama 2 GGML (Quantized Model) on CPU with Langchain featuring Streaming Output, Prompt Template and LLM Chain

Ctransformers + Langchain Documentation - https://python.langchain.com/docs/integrations/llms/ctransformers
Colab code used in thevideo https://colab.research.google.com/drive/16Ebw0NSMY-bVUJtK2Pmz5SEC9mhqeRGR?usp=sharing

Llama 2 GGML Models - https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML

❤️ If you want to support the channel ❤️
Support here:
Patreon - https://www.patreon.com/1littlecoder/
Ko-Fi - https://ko-fi.com/1littlecoder

🧭 Follow me on 🧭
Twitter - https://twitter.com/1littlecoder
Linkedin - https://www.linkedin.com/in/amrrs/

cpu gpu langchain learn llama llama 2 llm prompt streaming support template

Senior AI/ML Developer

@ Lemon.io | Remote

Senior Applied Scientist

@ Tractable | London, UK

Senior Data Scientist, Product (Pro Growth)

@ Thumbtack | Remote, Ontario

Specialist Solutions Architect - Data Science / Machine Learning

@ Databricks | United States

Specialist Solutions Architect - Data Engineering (Financial Services)

@ Databricks | United States

Data Engineer I (R-15080)

@ Dun & Bradstreet | Hyderabad - India