all AI news
Large Language Models Are State-of-the-Art Evaluator for Grammatical Error Correction
March 27, 2024, 4:48 a.m. | Masamune Kobayashi, Masato Mita, Mamoru Komachi
cs.CL updates on arXiv.org arxiv.org
Abstract: Large Language Models (LLMs) have been reported to outperform existing automatic evaluation metrics in some tasks, such as text summarization and machine translation. However, there has been a lack of research on LLMs as evaluators in grammatical error correction (GEC). In this study, we investigate the performance of LLMs in GEC evaluation by employing prompts designed to incorporate various evaluation criteria inspired by previous research. Our extensive experimental results demonstrate that GPT-4 achieved Kendall's rank …
abstract art arxiv cs.cl error error correction evaluation evaluation metrics gec however language language models large language large language models llms machine machine translation metrics research state study summarization tasks text text summarization translation type
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Business Data Analyst
@ Alstom | Johannesburg, GT, ZA