all AI news
ComplexityNet: Increasing LLM Inference Efficiency by Learning Task Complexity
April 2, 2024, 7:45 p.m. | Henry Bae, Aghyad Deeb, Alex Fleury, Kehang Zhu
cs.LG updates on arXiv.org arxiv.org
Abstract: We present ComplexityNet, a streamlined language model designed for assessing task complexity. This model predicts the likelihood of accurate output by various language models, each with different capabilities. Our initial application of ComplexityNet involves the Mostly Basic Python Problems (MBPP) dataset. We pioneered the creation of the first set of labels to define task complexity. ComplexityNet achieved a notable 79% accuracy in determining task complexity, a significant improvement over the 34% accuracy of the original, …
abstract application arxiv basic capabilities complexity cs.ai cs.cl cs.lg dataset efficiency inference language language model language models likelihood llm python type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Data Engineer
@ Kaseya | Bengaluru, Karnataka, India