Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audio codec. Mimi processes 24 kHz audio, down to a 12.5 Hz representation with a bandwidth of 1.1 kbps, in a fully streaming manner (latency of 80ms, the frame size), yet performs better than existing, non-streaming, codecs like SpeechTokenizer (50 Hz, 4kbps), or SemantiCodec (50 Hz, 1.3kbps). Moshi models two streams of audio: one corresponds to Moshi, and the other one to the user. At inference, the stream from the user is taken from the audio input, and the one for Moshi is sampled from the model's output. Along these two audio streams, Moshi predicts text tokens corresponding to its own speech, its inner monologue, which greatly improves the quality of its generation. A small Depth Transformer models inter codebook dependencies for a given time step, while a large, 7B parameter Temporal Transformer models the temporal dependencies.

Features

  • Converts JSON to Java and Kotlin objects
  • Annotation-based mapping for flexibility
  • Supports custom adapters for complex models
  • Lightweight with minimal dependencies
  • JSON serialization and deserialization

Project Samples

Project Activity

See All Activity >

Categories

Speech

License

Apache License V2.0

Follow Moshi

Moshi Web Site

Other Useful Business Software
The ultimate digital workspace. Icon
The ultimate digital workspace.

Axero Intranet is an award-winning intranet and employee experience platform.

Hundreds of companies and millions of employees use Axero’s intranet software to communicate, collaborate, manage tasks and events, organize content, and develop their company culture.
Learn More
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Moshi!

Additional Project Details

Operating Systems

Linux, Mac, Windows

Programming Language

Python

Related Categories

Python Speech Software

Registered

2024-11-04