Get the latest tech news

Moshi: A speech-text foundation model for real time dialogue


Contribute to kyutai-labs/moshi development by creating an account on GitHub.

Mimi builds on previous neural audio codecs such as SoundStream and EnCodec, adding a Transformer both in the encoder and decoder, and adapting the strides to match an overall frame rate of 12.5 Hz. This allows Mimi to get closer to the average frame rate of text tokens (~3-4 Hz), and limit the number of autoregressive steps in Moshi. Finally, and similarly to EBEN, Mimi uses only an adversarial training loss, along with feature matching, showing strong improvements in terms of subjective quality despite its low bitrate.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Moshi

Moshi

Photo of real time dialogue

real time dialogue

Related news:

News photo

Moshi: A voice AI that can display emotions