May 5, 2024, 7:25 a.m. | /u/jonathan-lei

Machine Learning www.reddit.com

[https://tensordock.com/benchmarks](https://tensordock.com/benchmarks)

Spent the past few hours putting together some data on vLLM (for both Llama 7B and OPT-125M) and Resnet-50 training performance on the TensorDock cloud.

vLLM data is 100% out of the box, with 2048 batch sizes from [this repository](https://github.com/vllm-project/vllm/tree/main/benchmarks).

My learnings:

* H100 and A100 performance is unbeatable, but the price-to-performance of lower-end RTX cards is pretty darn good. Even the L40 and RTX 6000 Ada outperform the A100 at some tasks, as they are 1 generation newer …

a100 ada cards cloud data good h100 llama machinelearning opt-125m performance price resnet resnet-50 rtx rtx 6000 tasks together training

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US