March 18, 2024, 4:47 a.m. | Thales Sales Almeida, Hugo Abonizio, Rodrigo Nogueira, Ramon Pires

cs.CL updates on arXiv.org arxiv.org

arXiv:2403.09887v1 Announce Type: new
Abstract: We introduce Sabi\'a-2, a family of large language models trained on Portuguese texts. The models are evaluated on a diverse range of exams, including entry-level tests for Brazilian universities, professional certification exams, and graduate-level exams for various disciplines such as accounting, economics, engineering, law and medicine. Our results reveal that our best model so far, Sabi\'a-2 Medium, matches or surpasses GPT-4's performance in 23 out of 64 exams and outperforms GPT-3.5 in 58 out of …

abstract accounting arxiv certification cs.ai cs.cl diverse economics engineering exams family graduate language language models large language large language models law medicine professional tests type universities

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote