all AI news
XGen-7B: Long Sequence Modeling with (up to) 8K Tokens. Overview, Dataset & Google Colab Code.
June 30, 2023, 3 p.m. | Venelin Valkov
Venelin Valkov www.youtube.com
In this video, we dive into the world of Long Sequence Modeling, exploring a 7B LLM, XGen-7B. With an impressive 8K input sequence length and fine-tuning on public-domain instructional data, XGen-7B promises a competition against state-of-the-art LLMs. We'll look at performance on standard NLP benchmarks, long sequence modeling tasks, and code generation.
I'll take you through the process of loading the instruction model in a Google Colab Notebook and …
code colab competition data dataset fine-tuning good google llm llms modeling overview public tokens video world
More from www.youtube.com / Venelin Valkov
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Machine Learning Engineer
@ Samsara | Canada - Remote