April 2, 2024, 7:43 p.m. | Michael Hassid, Tal Remez, Jonas Gehring, Roy Schwartz, Yossi Adi

cs.LG updates on arXiv.org arxiv.org

arXiv:2404.00725v1 Announce Type: cross
Abstract: It is a common belief that large language models (LLMs) are better than smaller-sized ones. However, larger models also require significantly more time and compute during inference. This begs the question: what happens when both models operate under the same budget? (e.g., compute, run-time). To address this question, we analyze code generation LLMs of various sizes and make comparisons such as running a 70B model once vs. generating five outputs from a 13B model and …

abstract arxiv belief budget code compute cs.ai cs.cl cs.lg cs.se however inference language language models large language large language models larger models llm llms question type via

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US