Oct. 27, 2023, 11:15 a.m. | Prompt Engineering

Prompt Engineering www.youtube.com

In this video, I will show you how to run MemGPT with Local LLMs by serving them through an API server (Textgen WebUI). I will walk you through a step by step process.

Commands:

Running WebUI API:

python server.py --api --api-blocking-port 5050 \
--model airoboros-l2-70b-3.1.2.Q4_K_M.gguf\
--loader llama.cpp --n-gpu-layers 1 --n_ctx 4096 \
--threads 8 --threads-batch 8 --n_batch 512


Let's Connect:
☕ Buy me a Coffee: https://ko-fi.com/promptengineering
|🔴 Support my work on Patreon: Patreon.com/PromptEngineering
🦾 Discord: https://discord.com/invite/t4eYQRUcXB
💼Consulting: https://calendly.com/engineerprompt/consulting-call


Links:
MemGPT …

api blocking cpp gpu llama llms openai process python running server show them through video

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Robotics Technician - 3rd Shift

@ GXO Logistics | Perris, CA, US, 92571