March 1, 2024, 5:49 a.m. | Chia-Hsuan Lee, Hao Cheng, Mari Ostendorf

cs.CL updates on arXiv.org arxiv.org

arXiv:2311.09758v2 Announce Type: replace
Abstract: Large language models (LLMs) have revolutionized the landscape of Natural Language Processing systems, but are computationally expensive. To reduce the cost without sacrificing performance, previous studies have explored various approaches to harness the potential of Small Language Models (SLMs) as cost-effective alternatives to their larger counterparts. Driven by findings that SLMs and LLMs exhibit complementary strengths in a structured knowledge extraction task, this work presents a novel SLM/LLM routing framework designed to improve computational efficiency …

abstract arxiv cost cs.cl dialogue harness landscape language language models language processing large language large language models llms natural natural language natural language processing orchestration performance processing reduce slms small small language models state studies systems tracking type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Sr. VBI Developer II

@ Atos | Texas, US, 75093

Wealth Management - Data Analytics Intern/Co-op Fall 2024

@ Scotiabank | Toronto, ON, CA