June 30, 2023, 3 p.m. | Venelin Valkov

Venelin Valkov www.youtube.com

Are Open LLMs any good when it comes to longer texts?

In this video, we dive into the world of Long Sequence Modeling, exploring a 7B LLM, XGen-7B. With an impressive 8K input sequence length and fine-tuning on public-domain instructional data, XGen-7B promises a competition against state-of-the-art LLMs. We'll look at performance on standard NLP benchmarks, long sequence modeling tasks, and code generation.

I'll take you through the process of loading the instruction model in a Google Colab Notebook and …

code colab competition data dataset fine-tuning good google llm llms modeling overview public tokens video world

More from www.youtube.com / Venelin Valkov

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Machine Learning Engineer

@ Samsara | Canada - Remote