April 2, 2024, 7:52 p.m. | Zilong Wang, Xufang Luo, Xinyang Jiang, Dongsheng Li, Lili Qiu

cs.CL updates on arXiv.org arxiv.org

arXiv:2404.00998v1 Announce Type: new
Abstract: Evaluating generated radiology reports is crucial for the development of radiology AI, but existing metrics fail to reflect the task's clinical requirements. This study proposes a novel evaluation framework using large language models (LLMs) to compare radiology reports for assessment. We compare the performance of various LLMs and demonstrate that, when using GPT-4, our proposed metric achieves evaluation consistency close to that of radiologists. Furthermore, to reduce costs and improve accessibility, making this method practical, …

abstract arxiv assessment clinical cs.ai cs.cl development evaluation framework generated language language models large language large language models llm llms metrics novel performance radiologist radiology ray report reports requirements study type x-ray

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

#13721 - Data Engineer - AI Model Testing

@ Qualitest | Miami, Florida, United States

Elasticsearch Administrator

@ ManTech | 201BF - Customer Site, Chantilly, VA