all AI news
The Larger the Better? Improved LLM Code-Generation via Budget Reallocation
April 2, 2024, 7:43 p.m. | Michael Hassid, Tal Remez, Jonas Gehring, Roy Schwartz, Yossi Adi
cs.LG updates on arXiv.org arxiv.org
Abstract: It is a common belief that large language models (LLMs) are better than smaller-sized ones. However, larger models also require significantly more time and compute during inference. This begs the question: what happens when both models operate under the same budget? (e.g., compute, run-time). To address this question, we analyze code generation LLMs of various sizes and make comparisons such as running a 70B model once vs. generating five outputs from a 13B model and …
abstract arxiv belief budget code compute cs.ai cs.cl cs.lg cs.se however inference language language models large language large language models larger models llm llms question type via
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Business Data Analyst
@ Alstom | Johannesburg, GT, ZA