Sept. 9, 2023, noon | code_your_own_AI

code_your_own_AI www.youtube.com

LLM Quantization: GPTQ - AutoGPTQ
llama.cpp - ggml.c - GGUL - C++
Compare to HF transformers in 4-bit quantization.

Download Web UI wrappers for your heavily quantized LLM to your local machine (PC, Linux, Apple).
LLM on Apple Hardware, w/ M1, M2 or M3 chip.
Run inference of your LLMs on your local PC, with heavy quantization applied.

Plus: 8 Web UI for GTPQ, llama.cpp or AutoGPTQ, exLLama or GGUF.c
koboldcpp
oobabooga text-generation-webui
ctransformers

https://lmstudio.ai/
https://github.com/marella/ctransformers
https://github.com/ggerganov/ggml
https://github.com/rustformers/llm/blob/main/crates/ggml/README.md
https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/README.md
https://github.com/PanQiWei/AutoGPTQ …

apple chip cpp download hardware inference linux llama llama 2 llm machine quantization transformers tutorial web

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Machine Learning Engineer

@ Samsara | Canada - Remote