Jan. 31, 2024, 3:42 p.m. | Angels Balaguer Vinamra Benara Renato Luiz de Freitas Cunha Roberto de M. Estev\~ao Filho Todd Hendry Daniel H

cs.CL updates on arXiv.org arxiv.org

There are two common ways in which developers are incorporating proprietary and domain-specific data when building applications of Large Language Models (LLMs): Retrieval-Augmented Generation (RAG) and Fine-Tuning. RAG augments the prompt with the external data, while fine-Tuning incorporates the additional knowledge into the model itself. However, the pros and cons of both approaches are not well understood. In this paper, we propose a pipeline for fine-tuning and RAG, and present the tradeoffs of both for multiple popular LLMs, including Llama2-13B, …

agriculture applications building case case study cs.cl cs.lg data developers domain external data fine-tuning knowledge language language models large language large language models llms pipelines prompt proprietary pros rag retrieval retrieval-augmented study the prompt

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Research Scientist, Demography and Survey Science, University Grad

@ Meta | Menlo Park, CA | New York City

Computer Vision Engineer, XR

@ Meta | Burlingame, CA