all AI news
[D] Speed Up in FP32 vs FP16
Jan. 29, 2024, 9:33 p.m. | /u/MaintenanceNo5993
Machine Learning www.reddit.com
- Model: CLIP ViT-B-32
- Dataset: MSCOCO Captions
- Number of Workers: 4
- Batch Size: 240 in case of FP16 and 160 in case of FP32
For both FP32 and FP16, each epoch is taking around 6 mins.
One of reason I consider is that *majority of time might constitute of data movement* rather GPU processing, as in case of FP32 there's hardly a moment when GPU utilization …
captions case clip dataset fp16 gpus machinelearning node reason speed training vit workers
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne