A real-time foundation model that brings human-like digital presence to customer conversations, virtual assistants, training, and interactive experiences

Today, we are publicly releasing Higgs Audio 3.0, a state-of-the-art Speech-to-Text (STT / ASR) foundation model. It supports 94 languages with sophisticated language detection, advanced sentiment and semantic understanding, and outperforms whisper-v3-large by a large margin on key languages.

After a successful event in Toronto at the MScAC headquarters last October (200 participants), we are bringing the same energy to California with the 2026 Bay Area edition of our Higgs Audio Hackathon series in partnership with Eigen AI.

Today, we are proud to launch Higgs Audio 2.5, the latest iteration of Boson AI’s audio model, designed to bring high-fidelity generation into production environments. Building on Higgs Audio 2, this release combines improved efficiency with the stability required for real-world deployment.

From October 24th through October 26th, 2025, Boson AI hosted its first large-scale Higgs Audio Hackathon in partnership with the Master of Science in Applied Computing (MScAC) program at the University of Toronto. With the recent release of our open-sourced Higgs Audio 2 text-to-speech (TTS) and automatic speech recognition (ASR) models, we opened the doors to over 200 innovators and developers, eager to explore, build, and most importantly, break our models.

Announcing Version 2 of Higgs Audio Generation, our latest advancement in audio generation technology with enhanced multi-speaker and dialog capabilities. Now open source.