Feb. 26, 2024, 5:42 a.m. | Jiayi Liu, Tinghan Yang, Jennifer Neville

cs.LG updates on arXiv.org arxiv.org

arXiv:2402.14833v1 Announce Type: cross
Abstract: Large language models (LLMs) have become pivotal in recent research. However, during the inference process, LLMs still require substantial resources. In this paper, we propose CliqueParcel, a method designed to improve the efficiency of LLMs via prompt batching. Existing strategies to optimize inference efficiency often compromise on output quality, leading to a discounted output problem. This issue might result in reduced accuracy or outputs that are less detailed. CliqueParcel is our answer to this challenge. …

abstract arxiv batching become cs.ai cs.cl cs.lg efficiency inference language language models large language large language models llm llms paper pivotal process prompt prompts research resources strategies type via

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Robotics Technician - 3rd Shift

@ GXO Logistics | Perris, CA, US, 92571