March 26, 2024, 4:51 a.m. | Dongjun Jang, Sungjoo Byun, Hyemi Jo, Hyopil Shin

cs.CL updates on arXiv.org arxiv.org

arXiv:2403.16444v1 Announce Type: new
Abstract: Instruction Tuning on Large Language Models is an essential process for model to function well and achieve high performance in specific tasks. Accordingly, in mainstream languages such as English, instruction-based datasets are being constructed and made publicly available. In the case of Korean, publicly available models and datasets all rely on using the output of ChatGPT or translating datasets built in English. In this paper, We introduce \textit{KIT-19} as an instruction dataset for the development …

abstract arxiv cs.cl datasets english fine-tuning function language language models languages large language large language models performance process specific tasks tasks toolkit type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Data Scientist, gTech Ads

@ Google | Mexico City, CDMX, Mexico

Lead, Data Analytics Operations

@ Zocdoc | Pune, Maharashtra, India