May 8, 2024, 3 p.m. | Jules Damji

InfoQ - AI, ML & Data Engineering www.infoq.com

Jules Damji discusses which infrastructure should be used for distributed fine-tuning and training, how to scale ML workloads, how to accommodate large models, and how can CPUs and GPUs be utilized?

By Jules Damji

ai artificial intelligence compute cpus distributed fine-tuning gpus infrastructure large models llm machine learning ml & data engineering modern performance & scalability presentation qcon san francisco 2023 scale scaling stack training transcripts workloads

More from www.infoq.com / InfoQ - AI, ML & Data Engineering

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US