VoxCPM2 logo

VoxCPM2

Open-source 48kHz TTS with voice design and cloning

2026-04-13

Product Introduction

  1. Definition: VoxCPM2 is a 2-billion parameter (2B), open-source Text-to-Speech (TTS) model and speech generation framework. It utilizes a tokenizer-free, end-to-end diffusion autoregressive architecture to convert text into high-fidelity, continuous speech representations without the artifacts often associated with discrete tokenization.

  2. Core Value Proposition: VoxCPM2 exists to provide a production-ready, multilingual speech synthesis solution that bridges the gap between synthetic robotic voices and human-like expressive audio. By integrating advanced "Voice Design" capabilities and controllable voice cloning into a 48kHz studio-quality pipeline, it empowers developers and creators to generate personalized, high-performance vocal content at scale. Primary keywords include zero-shot TTS, multilingual speech generation, real-time voice cloning, and tokenizer-free architecture.

Main Features

  1. Tokenizer-Free Diffusion Autoregressive Architecture: Unlike traditional TTS systems that rely on discrete audio tokens (which can lose nuances), VoxCPM2 operates entirely in the continuous latent space of AudioVAE V2. It employs a four-stage pipeline—LocEnc (Local Encoder), TSLM (Text-to-Speech Language Model), RALM (Reference-Aware Language Model), and LocDiT (Localized Diffusion Transformer). This approach allows the model to capture fine-grained prosody, rhythm, and emotional inflections.

  2. Creative Voice Design from Natural Language: This feature allows users to generate a completely unique synthetic voice based solely on a text description. By inputting parameters such as gender, age, tone, and emotion (e.g., "A young woman with a gentle, sweet voice"), the model synthesizes speech without requiring any reference audio clip. This is powered by the model's deep understanding of vocal characteristics embedded in its 2B parameter backbone.

  3. Hybrid Voice Cloning (Controllable & Ultimate): VoxCPM2 offers two distinct cloning modes. "Controllable Cloning" allows for the replication of a speaker's timbre from a short clip while using text prompts to steer the emotion or pace. "Ultimate Cloning" (audio-continuation) uses both a reference clip and its transcript to faithfully reproduce every nuance of the original speaker, including specific speech habits and stylistic quirks, ensuring maximum similarity.

  4. 48kHz Studio-Quality Audio Output: The system features an asymmetric AudioVAE V2 design that accepts 16kHz reference audio but directly generates 48kHz high-resolution output. This built-in super-resolution capability eliminates the need for external neural upsamplers or vocoders, reducing pipeline complexity while maintaining professional audio standards.

  5. 30-Language Multilingual Support: The model is pre-trained on over 2 million hours of speech data, supporting 30 languages including English, Chinese (and multiple dialects like Cantonese and Sichuanese), Japanese, Korean, Spanish, French, Arabic, and more. It features automatic language inference, meaning users do not need to manually provide language tags for synthesis.

Problems Solved

  1. Pain Point: Robotic and Monotonous Synthetic Speech. Traditional TTS models often struggle with "uncanny valley" effects and lack emotional depth. VoxCPM2's context-aware synthesis automatically infers appropriate prosody from the text content, making the output sound natural and human-like.

  2. Pain Point: High Latency in Production Environments. High-parameter models are often too slow for live applications. VoxCPM2 addresses this with a Real-Time Factor (RTF) of ~0.3 on standard hardware (RTX 4090) and as low as ~0.13 when accelerated by Nano-vLLM, making it suitable for real-time streaming and interactive AI agents.

  3. Target Audience: This product is designed for AI Research Engineers, Game Developers (for NPC dialogue), Content Creators (for podcasts and YouTube narration), Localization Specialists, and Enterprise Developers building customer service voicebots or accessibility tools.

  4. Use Cases: Essential for generating high-quality localized marketing content across dozens of languages, creating unique brand voices through Voice Design, developing immersive gaming experiences with diverse character voices, and providing real-time, low-latency voice responses for AI assistants.

Unique Advantages

  1. Differentiation: Unlike competitors that require massive datasets for fine-tuning, VoxCPM2 excels at zero-shot synthesis. Its "tokenizer-free" approach sets it apart from models like VITS or early GPT-based TTS by providing smoother transitions and higher emotional fidelity. Furthermore, it combines the stability of autoregressive models with the generative quality of diffusion models.

  2. Key Innovation: The integration of the MiniCPM-4 backbone provides a superior linguistic understanding, allowing the model to interpret the emotional "intent" behind the text. Additionally, the Apache-2.0 license provides a significant commercial advantage over proprietary models (like ElevenLabs) or more restrictively licensed open-source models, allowing for unrestricted commercial deployment and local hosting.

  3. Ecosystem Versatility: VoxCPM2 is built for the community, supporting various deployment backends including GGML/GGUF (for CPU inference), ONNX, Apple Neural Engine (ANE), and ComfyUI integration, ensuring it can run on everything from high-end servers to local edge devices.

Frequently Asked Questions (FAQ)

  1. What makes VoxCPM2 "tokenizer-free" and why does it matter? VoxCPM2 bypasses the traditional step of converting audio into discrete digital "tokens" (like words in an LLM). Instead, it generates continuous speech representations. This matters because it preserves the subtle, fluid nuances of human speech—such as breathing, pitch slides, and emotional micro-expressions—that are often lost in tokenized systems.

  2. Can VoxCPM2 be used for commercial projects? Yes. Both the model weights and the underlying code are released under the Apache-2.0 license. This allows businesses to use, modify, and distribute the model for commercial purposes without paying royalty fees, provided they comply with the standard license terms.

  3. How much VRAM is required to run VoxCPM2? For standard inference, VoxCPM2 requires approximately 8GB of VRAM, making it compatible with consumer-grade GPUs like the NVIDIA RTX 3060 or 4060. For high-throughput production serving using Nano-vLLM, similar memory profiles apply while significantly increasing the generation speed.

  4. How do I create a specific voice without having a recording? You can use the "Voice Design" feature. By placing a description in parentheses at the start of your text prompt—such as "(A middle-aged man with a deep, authoritative voice)"—the model uses its internal 2B-parameter knowledge to synthesize a new voice that matches those specific characteristics without needing any external audio input.

Subscribe to Our Newsletter

Get weekly curated tool recommendations and stay updated with the latest product news