April 16, 2024, 10:24 p.m. | Mike Young

DEV Community dev.to

This is a Plain English Papers summary of a research paper called H2O-Danube-1.8B Technical Report. If you like these kinds of analysis, you should subscribe to the AImodels.fyi newsletter or follow me on Twitter.





Overview



  • Presents H2O-Danube, a series of small 1.8B language models

  • H2O-Danube-1.8B is trained on 1T tokens, and H2O-Danube2-1.8B is trained on an additional 2T tokens

  • Models exhibit highly competitive metrics across multiple benchmarks

  • H2O-Danube2-1.8B achieves top ranking on Open LLM Leaderboard for models below …

ai aimodels analysis beginners danube datascience english h2o h2o-danube-1.8b language language models machinelearning newsletter overview paper papers plain english papers report research research paper series small summary technical tokens twitter

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Data Scientist

@ ITE Management | New York City, United States