Oct. 27, 2023, 11:15 a.m. | Prompt Engineering

Prompt Engineering www.youtube.com

In this video, I will show you how to run MemGPT with Local LLMs by serving them through an API server (Textgen WebUI). I will walk you through a step by step process.

Commands:

Running WebUI API:

python server.py --api --api-blocking-port 5050 \
--model airoboros-l2-70b-3.1.2.Q4_K_M.gguf\
--loader llama.cpp --n-gpu-layers 1 --n_ctx 4096 \
--threads 8 --threads-batch 8 --n_batch 512


Let's Connect:
☕ Buy me a Coffee: https://ko-fi.com/promptengineering
|🔴 Support my work on Patreon: Patreon.com/PromptEngineering
🦾 Discord: https://discord.com/invite/t4eYQRUcXB
💼Consulting: https://calendly.com/engineerprompt/consulting-call


Links:
MemGPT …

api blocking cpp gpu llama llms openai process python running server show them through video

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US