Feb. 6, 2024, 5:54 a.m. | Ahmed Abdelali Hamdy Mubarak Shammur Absar Chowdhury Maram Hasanain Basel Mousi Sabri Boughorbel Yassi

cs.CL updates on arXiv.org arxiv.org

Recent advancements in Large Language Models (LLMs) have significantly influenced the landscape of language and speech research. Despite this progress, these models lack specific benchmarking against state-of-the-art (SOTA) models tailored to particular languages and tasks. LAraBench addresses this gap for Arabic Natural Language Processing (NLP) and Speech Processing tasks, including sequence tagging and content classification across different domains. We utilized models such as GPT-3.5-turbo, GPT-4, BLOOMZ, Jais-13b-chat, Whisper, and USM, employing zero and few-shot learning techniques to tackle 33 distinct …

arabic art benchmarking cs.ai cs.cl gap landscape language language models language processing languages large language large language models llms natural natural language natural language processing nlp processing progress research sota speech speech processing state tagging tasks

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote