Machine Learning

H2O.ai releases new language model H2O-Danube-1.8B for mobile

H2O-Danube-1.8B super tiny LLM model designed to run on smartphones, laptops, desktops and IoT devices, spurring growth in natural language applications and further democratizing AI

H2O.ai, the open source leader in Generative AI and machine learning and maker behind Enterprise h2oGPTe, is announcing the release of H2O-Danube-1.8B – an open source natural language model with 1.8 billion parameters. Despite being trained on significantly less data than comparable models, benchmark results show H2O-Danube-1.8B achieves highly competitive performance across a wide range of natural language tasks.

“We are excited to release H2O-Danube-1.8B as a portable LLM on small devices like your smartphone, something that Anthropic is not offering today. The proliferation of smaller, lower-cost hardware and more efficient training now allows modestly-sized models to be accessible to a wider audience. With an Apache 2.0 license for commercial use and versatile capabilities, we believe H2O-Danube-1.8B will be a game changer for mobile offline applications,” said Sri Ambati, CEO and co-founder of H2O.ai.

As detailed in the arXiv technical report, H2O-Danube-1.8B was trained on 1 trillion tokens collected from diverse web sources, with techniques refined from models like LLama 2 and Mistral. Despite the relatively limited training data, benchmark results show H2O-Danube-1.8B performs on par or better than other models in the 1-2 billion parameter size class across tasks like common sense reasoning, reading comprehension, summarization and translation.

H2O.ai also announced the release of H2O-Danube-1.8B-Chat, a version of the model fine-tuned specifically for conversational applications. Building on the base H2O-Danube-1.8B model, the chat version was tuned using supervised learning on dialog datasets followed by reinforcement learning using human preferences. Initial benchmark results show state-of-the-art performance compared to existing chat models with less than 2 billion parameters.

Both the base H2O-Danube-1.8B model and chat-tuned version are available immediately from Hugging Face. H2O.ai will be releasing additional tools to simplify using the models in applications, as well as exploring potential future model scaling.

“We are committed to advancing AI for responsible progress. H2O-Danube-1.8B raises the bar through its impressive performance combined with an open license enabling broad access,” said Sri Ambati, CEO and co-founder of H2O.ai.

Visit AITechPark for cutting-edge Tech Trends around AI, ML, Cybersecurity, along with AITech News, and timely updates from industry professionals!

Related posts

Roostify Named Google Partner Award Recipient for AI and ML

PR Newswire

QAD Brings Power of Process Intelligence to the Adaptive Enterprise

Business Wire

Riverain to Showcase AI Innovations and Initiatives at RSNA 2020

PR Newswire