<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://tts.wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ttswikiadmin</id>
	<title>TTS Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://tts.wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ttswikiadmin"/>
	<link rel="alternate" type="text/html" href="https://tts.wiki/index.php/Special:Contributions/Ttswikiadmin"/>
	<updated>2026-04-03T17:29:00Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.41.5</generator>
	<entry>
		<id>https://tts.wiki/index.php?title=SNAC&amp;diff=59</id>
		<title>SNAC</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=SNAC&amp;diff=59"/>
		<updated>2025-12-23T03:44:01Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;&amp;#039;&amp;#039;&amp;#039;SNAC&amp;#039;&amp;#039;&amp;#039; (Multi-Scale Neural Audio Codec) is a neural audio codec that introduces multi-scale temporal quantization for efficient audio compression. It was presented at the NeurIPS 2024 Workshop on AI-Driven Speech, Music, and Sound Generation by researchers from Papla Media and ETH Zurich.  === Overview === Neural audio codecs have recently gained popularity because they can represent audio signals with high fidelity at very low bitrates, making it feasible to use...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;SNAC&#039;&#039;&#039; (Multi-Scale Neural Audio Codec) is a neural audio codec that introduces multi-scale temporal quantization for efficient audio compression. It was presented at the NeurIPS 2024 Workshop on AI-Driven Speech, Music, and Sound Generation by researchers from [[Papla Media]] and ETH Zurich.&lt;br /&gt;
&lt;br /&gt;
=== Overview ===&lt;br /&gt;
Neural audio codecs have recently gained popularity because they can represent audio signals with high fidelity at very low bitrates, making it feasible to use language modeling approaches for audio generation and understanding. While Residual Vector Quantization (RVQ) has become the standard technique for neural audio compression using a cascade of VQ codebooks, SNAC proposes a simple extension of RVQ where the quantizers can operate at different temporal resolutions.&lt;br /&gt;
&lt;br /&gt;
=== Architecture ===&lt;br /&gt;
SNAC encodes audio into hierarchical tokens similarly to SoundStream, EnCodec, and [[DAC]]. However, SNAC introduces a simple change where coarse tokens are sampled less frequently, covering a broader time span.&lt;br /&gt;
&lt;br /&gt;
The architecture includes several key innovations:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Multi-Scale Quantization&#039;&#039;&#039;: By applying a hierarchy of quantizers at variable frame rates, the codec adapts to the audio structure across multiple timescales.&lt;br /&gt;
* &#039;&#039;&#039;Noise Blocks&#039;&#039;&#039;: Noise blocks that inject input-dependent Gaussian noise for enhanced expressiveness&lt;br /&gt;
* &#039;&#039;&#039;Depthwise Convolutions&#039;&#039;&#039;: Depthwise convolutions for efficient computation and training stability&lt;br /&gt;
* &#039;&#039;&#039;Local Windowed Attention&#039;&#039;&#039;: Local windowed attention layers at the lowest temporal resolution to capture contextual relationships&lt;br /&gt;
&lt;br /&gt;
=== Model Variants ===&lt;br /&gt;
SNAC offers several pretrained models optimized for different use cases:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Model&lt;br /&gt;
!Sample Rate&lt;br /&gt;
!Bitrate&lt;br /&gt;
!RVQ Levels&lt;br /&gt;
!Token Rates&lt;br /&gt;
!Parameters&lt;br /&gt;
!Use Case&lt;br /&gt;
|-&lt;br /&gt;
|snac_24khz&lt;br /&gt;
|24 kHz&lt;br /&gt;
|0.98 kbps&lt;br /&gt;
|3&lt;br /&gt;
|12, 23, and 47 Hz&lt;br /&gt;
|~20M&lt;br /&gt;
|Speech&lt;br /&gt;
|-&lt;br /&gt;
|snac_32khz&lt;br /&gt;
|32 kHz&lt;br /&gt;
|1.9 kbps&lt;br /&gt;
|4&lt;br /&gt;
|10, 21, 42, and 83 Hz&lt;br /&gt;
|~55M&lt;br /&gt;
|General audio&lt;br /&gt;
|-&lt;br /&gt;
|snac_44khz&lt;br /&gt;
|44 kHz&lt;br /&gt;
|2.6 kbps&lt;br /&gt;
|4&lt;br /&gt;
|14, 29, 57, and 115 Hz&lt;br /&gt;
|~55M&lt;br /&gt;
|Music/SFX&lt;br /&gt;
|}&lt;br /&gt;
Each codebook holds 4096 entries (12-bit). The general audio model consists of 16M parameters in the encoder and 38.3M in the decoder, totaling 54.5 M parameters.&lt;br /&gt;
&lt;br /&gt;
=== Performance ===&lt;br /&gt;
For speech, SNAC consistently outperforms all other codecs. Notably, even at bitrates below 1 kbit/s, SNAC maintains audio quality that closely approaches the reference signal. In evaluations, SNAC outperformed competing codecs like Encodec and DAC at comparable bitrates, even matching the quality of systems operating at twice its bitrate.&lt;br /&gt;
&lt;br /&gt;
=== Applications ===&lt;br /&gt;
SNAC has been adopted in several text-to-speech systems:&lt;br /&gt;
&lt;br /&gt;
* [[Orpheus TTS|&#039;&#039;&#039;Orpheus TTS&#039;&#039;&#039;]]: Orpheus uses SNAC, which creates tokens at four levels of hierarchy. The SNAC model is relatively lightweight and fast, making it suitable for real-time decoding.&lt;br /&gt;
&lt;br /&gt;
With coarse tokens of ~10 Hz and a context window of 2048 you can effectively model a consistent structure of an audio track for ~3 minutes.&lt;br /&gt;
&lt;br /&gt;
=== Comparison with Other Codecs ===&lt;br /&gt;
SNAC from Orpheus does 83 tokens per second, compared to 50 t/s for [[X-Codec|X-Codec 2.0]] and 25 t/s for [[CosyVoice]]&#039;s codec. SNAC uses one codebook but tokens are created for each level of downsampling, in contrast to codecs like [[Mimi]] which use multiple separate codebooks.&lt;br /&gt;
&lt;br /&gt;
[[Category:Neural audio codecs]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=X-Codec&amp;diff=58</id>
		<title>X-Codec</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=X-Codec&amp;diff=58"/>
		<updated>2025-12-23T02:33:59Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;X-Codec&#039;&#039;&#039; is a neural audio codec designed to enhance semantic understanding in audio language models (LLMs). It was introduced in the paper &amp;quot;Codec Does Matter: Exploring the Semantic Shortcoming of Codec for Audio Language Model,&amp;quot; published at AAAI 2025.&lt;br /&gt;
&lt;br /&gt;
=== Background ===&lt;br /&gt;
Traditional audio codecs like EnCodec were originally designed for audio compression, which may lead to suboptimal performance in the context of audio LLM. Research found that methods like VALL-E, which condition acoustic token generation on text transcriptions, often suffer from content inaccuracies and elevated word error rates (WER) due to semantic misinterpretations of acoustic tokens, resulting in word skipping and errors.&lt;br /&gt;
&lt;br /&gt;
=== Architecture ===&lt;br /&gt;
X-Codec addresses these limitations through a dual-encoder design that incorporates semantic features from a pre-trained semantic encoder before the Residual Vector Quantization (RVQ) stage and introduces a semantic reconstruction loss after RVQ.&lt;br /&gt;
&lt;br /&gt;
The architecture consists of:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Acoustic Encoder/Decoder&#039;&#039;&#039;: Convolutional encoder and decoder with a Residual Vector Quantizer (RVQ)&lt;br /&gt;
* &#039;&#039;&#039;Semantic Module&#039;&#039;&#039;: A pre-trained self-supervised model such as HuBERT or WavLM&lt;br /&gt;
* &#039;&#039;&#039;Projectors&#039;&#039;&#039;: Linear layers that combine and process the acoustic and semantic features&lt;br /&gt;
&lt;br /&gt;
The acoustic and semantic features are concatenated, transformed, and then quantized together. After quantization, separate post-processing layers reconstruct both semantic and acoustic representations.&lt;br /&gt;
&lt;br /&gt;
=== Applications ===&lt;br /&gt;
X-Codec demonstrated improvements across multiple audio generation tasks including text-to-speech synthesis, music continuation, and general audio classification tasks. By enhancing the semantic ability of the codec, X-Codec significantly reduces WER in speech synthesis tasks and extends these benefits to non-speech applications, including music and sound generation.&lt;br /&gt;
&lt;br /&gt;
== X-Codec 2.0 ==&lt;br /&gt;
&#039;&#039;&#039;X-Codec 2.0&#039;&#039;&#039; (also written as XCodec2) is a successor to X-Codec, introduced alongside the LLaSA text-to-speech system in the paper &amp;quot;LLaSA: Scaling Train-time and Inference-time Compute for LLaMA-based Speech Synthesis.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
=== Key Differences from X-Codec ===&lt;br /&gt;
X-Codec2 extends the original X-Codec by refining how semantic and acoustic information is integrated and tokenized. Major architectural changes include:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Unified Semantic-Acoustic Tokenization&#039;&#039;&#039;: X-Codec2 fuses outputs from a semantic encoder (e.g., Wav2Vec2-BERT) and an acoustic encoder into a single embedding, capturing both high-level meaning (e.g., text content, emotion) and low-level audio details (e.g., timbre).&lt;br /&gt;
* &#039;&#039;&#039;Single-Stage Vector Quantization&#039;&#039;&#039;: Unlike the multi-layer residual VQ in most approaches (e.g., X-Codec, DAC, EnCodec), X-Codec2 uses a single-layer Feature-Space Quantization (FSQ) for stability and compatibility with causal language models.&lt;br /&gt;
* &#039;&#039;&#039;Large Codebook&#039;&#039;&#039;: 65,536 codebook size using Finite Scalar Quantization achieving 99% codebook usage, which is comparable to text tokenizers (LLaMA3 uses 128,256).&lt;br /&gt;
&lt;br /&gt;
=== Technical Specifications ===&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Semantic Encoder&#039;&#039;&#039;: Wav2Vec2-BERT, a semantic encoder pre-trained on 4.5M hours of unlabeled audio data covering more than 143 languages.&lt;br /&gt;
* &#039;&#039;&#039;Training Data&#039;&#039;&#039;: Codec trained on 150k hours of multilingual speech data, including Emilia (En/Zh/De/Fr/Ja/Ko) and MLS (En/Fr/De/Nl/Es/It/Pt/Pl).&lt;br /&gt;
* &#039;&#039;&#039;Quantization&#039;&#039;&#039;: Finite Scalar Quantization (FSQ), which does not require an explicit VQ objective term (e.g., codebook commitment loss), simplifying optimization during training.&lt;br /&gt;
&lt;br /&gt;
=== Derivatives ===&lt;br /&gt;
X-Codec 2.0 has been extended in several ways:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;NeuCodec&#039;&#039;&#039;: Neuphonic&#039;s codec for on-device TTS, which is largely based on extending X-Codec 2.0&lt;br /&gt;
* &#039;&#039;&#039;XCodec2-Streaming&#039;&#039;&#039;: A streaming variant that adopts a causal decoder to focus solely on historical context, enabling streaming waveform reconstruction.&lt;br /&gt;
&lt;br /&gt;
=== Availability ===&lt;br /&gt;
X-Codec is available on GitHub and integrated into Hugging Face&#039;s Transformers library. X-Codec 2.0 is available via the &amp;lt;code&amp;gt;xcodec2&amp;lt;/code&amp;gt; Python package and on Hugging Face.&lt;br /&gt;
[[Category:Neural audio codecs]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=NeuCodec&amp;diff=57</id>
		<title>NeuCodec</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=NeuCodec&amp;diff=57"/>
		<updated>2025-12-23T02:33:27Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;&amp;#039;&amp;#039;&amp;#039;NeuCodec&amp;#039;&amp;#039;&amp;#039; is a neural audio codec developed by Neuphonic, designed for efficient speech tokenization and high-quality audio compression at relatively low bitrates.  === Technical Specifications ===  * &amp;#039;&amp;#039;&amp;#039;Bitrate:&amp;#039;&amp;#039;&amp;#039; 0.8 kbps * &amp;#039;&amp;#039;&amp;#039;Output sample rate:&amp;#039;&amp;#039;&amp;#039; 24 kHz * &amp;#039;&amp;#039;&amp;#039;Frame rate:&amp;#039;&amp;#039;&amp;#039; 50 Hz * &amp;#039;&amp;#039;&amp;#039;Quantization:&amp;#039;&amp;#039;&amp;#039; Finite Scalar Quantization (FSQ) with a single codebook  === Architecture === NeuCodec is largely based on extending the work of X-Codec 2.0. It e...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;NeuCodec&#039;&#039;&#039; is a neural audio codec developed by [[Neuphonic]], designed for efficient speech tokenization and high-quality audio compression at relatively low bitrates.&lt;br /&gt;
&lt;br /&gt;
=== Technical Specifications ===&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Bitrate:&#039;&#039;&#039; 0.8 kbps&lt;br /&gt;
* &#039;&#039;&#039;Output sample rate:&#039;&#039;&#039; 24 kHz&lt;br /&gt;
* &#039;&#039;&#039;Frame rate:&#039;&#039;&#039; 50 Hz&lt;br /&gt;
* &#039;&#039;&#039;Quantization:&#039;&#039;&#039; Finite Scalar Quantization (FSQ) with a single codebook&lt;br /&gt;
&lt;br /&gt;
=== Architecture ===&lt;br /&gt;
NeuCodec is largely based on extending the work of [[X-Codec|X-Codec 2.0]]. It employs a dual-encoder approach, using both audio ([[BigCodec]]) and semantic (Wav2Vec2-BERT) encoders. The FSQ-based design produces a single quantized vector output, making it well-suited for downstream Speech Language Model (SpeechLM) training.&lt;br /&gt;
&lt;br /&gt;
=== Features ===&lt;br /&gt;
&lt;br /&gt;
* Compresses and reconstructs audio with near-inaudible reconstruction loss&lt;br /&gt;
* Upsamples from 16 kHz to 24 kHz&lt;br /&gt;
* Commercial use permitted&lt;br /&gt;
* Pre-encoded datasets available (Emilia-YODAS compressed from 1.7 TB to 41 GB)&lt;br /&gt;
&lt;br /&gt;
=== Applications ===&lt;br /&gt;
NeuCodec serves as the audio codec for [[NeuTTS Air]], Neuphonic&#039;s on-device text-to-speech model with voice cloning capabilities. It&#039;s intended for researchers and developers building text-to-speech systems who need efficient speech tokenization without developing their own codec.&lt;br /&gt;
&lt;br /&gt;
=== Availability ===&lt;br /&gt;
Available on Hugging Face and GitHub under the &amp;lt;code&amp;gt;neuphonic/neucodec&amp;lt;/code&amp;gt; repository, installable via pip.&lt;br /&gt;
&lt;br /&gt;
[[Category:Neural audio codecs]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=X-Codec&amp;diff=56</id>
		<title>X-Codec</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=X-Codec&amp;diff=56"/>
		<updated>2025-12-23T02:30:16Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;&amp;#039;&amp;#039;&amp;#039;NeuCodec&amp;#039;&amp;#039;&amp;#039; is a neural audio codec developed by Neuphonic, designed for efficient speech tokenization and high-quality audio compression at relatively low bitrates.  === Technical Specifications ===  * &amp;#039;&amp;#039;&amp;#039;Bitrate:&amp;#039;&amp;#039;&amp;#039; 0.8 kbps * &amp;#039;&amp;#039;&amp;#039;Output sample rate:&amp;#039;&amp;#039;&amp;#039; 24 kHz * &amp;#039;&amp;#039;&amp;#039;Frame rate:&amp;#039;&amp;#039;&amp;#039; 50 Hz * &amp;#039;&amp;#039;&amp;#039;Quantization:&amp;#039;&amp;#039;&amp;#039; Finite Scalar Quantization (FSQ) with a single codebook  === Architecture === NeuCodec is largely based on extending the work of X-Codec 2.0. It e...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;NeuCodec&#039;&#039;&#039; is a neural audio codec developed by [[Neuphonic]], designed for efficient speech tokenization and high-quality audio compression at relatively low bitrates.&lt;br /&gt;
&lt;br /&gt;
=== Technical Specifications ===&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Bitrate:&#039;&#039;&#039; 0.8 kbps&lt;br /&gt;
* &#039;&#039;&#039;Output sample rate:&#039;&#039;&#039; 24 kHz&lt;br /&gt;
* &#039;&#039;&#039;Frame rate:&#039;&#039;&#039; 50 Hz&lt;br /&gt;
* &#039;&#039;&#039;Quantization:&#039;&#039;&#039; Finite Scalar Quantization (FSQ) with a single codebook&lt;br /&gt;
&lt;br /&gt;
=== Architecture ===&lt;br /&gt;
NeuCodec is largely based on extending the work of [[X-Codec|X-Codec 2.0]]. It employs a dual-encoder approach, using both audio ([[BigCodec]]) and semantic (Wav2Vec2-BERT) encoders. The FSQ-based design produces a single quantized vector output, making it well-suited for downstream Speech Language Model (SpeechLM) training.&lt;br /&gt;
&lt;br /&gt;
=== Features ===&lt;br /&gt;
&lt;br /&gt;
* Compresses and reconstructs audio with near-inaudible reconstruction loss&lt;br /&gt;
* Upsamples from 16 kHz to 24 kHz&lt;br /&gt;
* Commercial use permitted&lt;br /&gt;
* Pre-encoded datasets available (Emilia-YODAS compressed from 1.7 TB to 41 GB)&lt;br /&gt;
&lt;br /&gt;
=== Applications ===&lt;br /&gt;
NeuCodec serves as the audio codec for [[NeuTTS Air]], Neuphonic&#039;s on-device text-to-speech model with voice cloning capabilities. It&#039;s intended for researchers and developers building text-to-speech systems who need efficient speech tokenization without developing their own codec.&lt;br /&gt;
&lt;br /&gt;
=== Availability ===&lt;br /&gt;
Available on Hugging Face and GitHub under the &amp;lt;code&amp;gt;neuphonic/neucodec&amp;lt;/code&amp;gt; repository, installable via pip.&lt;br /&gt;
&lt;br /&gt;
[[Category:Neural audio codecs]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=ElevenLabs&amp;diff=54</id>
		<title>ElevenLabs</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=ElevenLabs&amp;diff=54"/>
		<updated>2025-09-25T03:01:23Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox TTS model&lt;br /&gt;
| name = ElevenLabs&lt;br /&gt;
| developer = ElevenLabs Inc.&lt;br /&gt;
| release_date = January 2023 (beta)&lt;br /&gt;
| latest_version = Eleven v3 (alpha)&lt;br /&gt;
| languages = 32+ languages&lt;br /&gt;
| voices = 1000+ voices&lt;br /&gt;
| voice_cloning = Yes (professional &amp;amp; instant)&lt;br /&gt;
| emotion_control = Yes (via audio tags)&lt;br /&gt;
| streaming = Yes&lt;br /&gt;
| latency = ~135ms (Flash models)&lt;br /&gt;
| open_source = No&lt;br /&gt;
| website = https://elevenlabs.io&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;ElevenLabs&#039;&#039;&#039; is a commercial artificial intelligence company specializing in text-to-speech synthesis and voice cloning technology. Founded in 2022 by Piotr Dąbkowski and Mateusz Staniszewski, the company has gained prominence for its AI-generated voices that can replicate human speech patterns, emotions, and intonation across multiple languages.&lt;br /&gt;
&lt;br /&gt;
== History and Founding ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs was co-founded in 2022 by Piotr Dąbkowski, a former Google machine learning engineer, and Mateusz Staniszewski, an ex-Palantir deployment strategist. Both founders, originally from Poland, reportedly drew inspiration from the poor quality of film dubbing they experienced while watching American movies in their home country.&amp;lt;ref&amp;gt;https://venturebeat.com/ai/now-hear-this-voice-cloning-ai-startup-elevenlabs-nabs-19m-from-a16z-and-other-heavy-hitters&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The founders first met as teenagers at Copernicus High School in Warsaw before pursuing separate academic paths—Dąbkowski studying at Oxford and Cambridge, while Staniszewski studied mathematics in London. Their shared vision of making quality content accessible across all languages led to the creation of ElevenLabs as a research-first company.&amp;lt;ref&amp;gt;https://research.contrary.com/company/elevenlabs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The company launched its beta platform in January 2023, quickly gaining traction with over one million users within five months. This rapid adoption demonstrated market demand for high-quality AI voice synthesis technology.&amp;lt;ref&amp;gt;https://research.contrary.com/company/elevenlabs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Funding and Valuation ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs has experienced rapid growth in both user adoption and valuation:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Pre-seed (January 2023)&#039;&#039;&#039;: $2 million led by Credo Ventures and Concept Ventures&lt;br /&gt;
* &#039;&#039;&#039;Series A (June 2023)&#039;&#039;&#039;: $19 million at $100 million valuation, co-led by Andreessen Horowitz, Nat Friedman, and Daniel Gross&lt;br /&gt;
* &#039;&#039;&#039;Series B (January 2024)&#039;&#039;&#039;: $80 million at $1.1 billion valuation, achieving unicorn status&lt;br /&gt;
* &#039;&#039;&#039;Series C (January 2025)&#039;&#039;&#039;: $180 million at $3.3 billion valuation, led by Andreessen Horowitz and ICONIQ Growth&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/ElevenLabs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The company reportedly achieved $200 million in annual recurring revenue (ARR) by August 2025, demonstrating significant commercial traction.&amp;lt;ref&amp;gt;https://sacra.com/c/elevenlabs/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technology and Products ==&lt;br /&gt;
&lt;br /&gt;
=== Core Technology ===&lt;br /&gt;
&lt;br /&gt;
ElevenLabs&#039;s architecture is proprietary and remains undisclosed, with little information about it being publicly available. Some have speculated that early versions of ElevenLabs were based off of Tortoise TTS; however, these rumors remain unverified.&amp;lt;ref&amp;gt;https://github.com/neonbjb/tortoise-tts/discussions/277&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Product Portfolio ===&lt;br /&gt;
&lt;br /&gt;
==== Text-to-Speech Models ====&lt;br /&gt;
&lt;br /&gt;
ElevenLabs offers several model variants optimized for different use cases:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Multilingual v2&#039;&#039;&#039;: High-quality model supporting 29+ languages, optimized for audiobooks and professional content&lt;br /&gt;
* &#039;&#039;&#039;Flash v2.5&#039;&#039;&#039;: Ultra-low latency model (75ms) designed for real-time conversational applications&lt;br /&gt;
* &#039;&#039;&#039;Turbo v2.5&#039;&#039;&#039;: Balanced quality and speed model for general-purpose applications&lt;br /&gt;
* &#039;&#039;&#039;Eleven v3 (alpha)&#039;&#039;&#039;: Latest model featuring advanced emotion control via audio tags&lt;br /&gt;
* &#039;&#039;&#039;Eleven Scribe v1&#039;&#039;&#039;: SoTA automatic speech recognition model&lt;br /&gt;
* &#039;&#039;&#039;Eleven Music v1&#039;&#039;&#039;: Text-to-music model trained on licensed data&amp;lt;ref&amp;gt;https://elevenlabs.io/docs/models&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://elevenlabs.io/music&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==== Voice Cloning ====&lt;br /&gt;
&lt;br /&gt;
The platform provides two voice cloning approaches:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Instant Voice Cloning&#039;&#039;&#039;: Creates voice replicas from short audio samples (1-5 minutes)&lt;br /&gt;
* &#039;&#039;&#039;Professional Voice Cloning&#039;&#039;&#039;: Higher-fidelity, fine-tuning-based cloning requiring longer training samples&lt;br /&gt;
&lt;br /&gt;
==== Additional Features ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;AI Dubbing&#039;&#039;&#039;: Translates and dubs content while preserving original voice characteristics and emotions&lt;br /&gt;
* &#039;&#039;&#039;Voice Design&#039;&#039;&#039;: Tool for creating entirely synthetic voices from text descriptions&lt;br /&gt;
* &#039;&#039;&#039;Speech Classifier&#039;&#039;&#039;: Detection tool to identify AI-generated audio from ElevenLabs&#039; technology&lt;br /&gt;
* &#039;&#039;&#039;Projects&#039;&#039;&#039;: Long-form content creation tool for audiobooks and extended narration&lt;br /&gt;
&lt;br /&gt;
== Business Model and Pricing ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs operates on a freemium subscription model with usage-based pricing:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Free Tier&#039;&#039;&#039;: 10,000 characters per month with basic voices&lt;br /&gt;
* &#039;&#039;&#039;Starter&#039;&#039;&#039;: $5/month with commercial licensing&lt;br /&gt;
* &#039;&#039;&#039;Creator&#039;&#039;&#039;: $11/month with enhanced features&lt;br /&gt;
* &#039;&#039;&#039;Pro&#039;&#039;&#039;: $99/month for professional use&lt;br /&gt;
* &#039;&#039;&#039;Enterprise&#039;&#039;&#039;: Custom pricing with SLAs and dedicated support&lt;br /&gt;
&lt;br /&gt;
The company has evolved its pricing structure multiple times, transitioning from simple character-based billing to more complex model-aware systems and back to unified credit systems as it scaled.&amp;lt;ref&amp;gt;https://flexprice.io/blog/elevenlabs-pricing-breakdown&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance and Benchmarks ==&lt;br /&gt;
&lt;br /&gt;
Independent evaluations have provided mixed results regarding ElevenLabs&#039; performance relative to competitors:&lt;br /&gt;
&lt;br /&gt;
=== Competitive Analysis ===&lt;br /&gt;
&lt;br /&gt;
According to third-party benchmarks:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Voice Quality&#039;&#039;&#039;: ElevenLabs demonstrates superior Mean Opinion Scores (MOS) compared to Google Cloud Text-to-Speech across fiction, non-fiction, and conversational content&amp;lt;ref&amp;gt;https://unrealspeech.com/compare/elevenlabs-vs-google-text-to-speech&amp;lt;/ref&amp;gt;&lt;br /&gt;
* &#039;&#039;&#039;Latency&#039;&#039;&#039;: Flash models achieve approximately 135ms Time to First Audio (TTFA), competitive with major cloud providers&amp;lt;ref&amp;gt;https://cartesia.ai/vs/elevenlabs-vs-microsoft-azure-text-to-speech&amp;lt;/ref&amp;gt;&lt;br /&gt;
* &#039;&#039;&#039;Accuracy&#039;&#039;&#039;: Word Error Rates vary but generally maintain competitive performance with established providers&lt;br /&gt;
&lt;br /&gt;
However, these evaluations should be interpreted cautiously as they often come from companies with commercial interests in the TTS space, and standardized, independent benchmarking in the industry remains limited.&lt;br /&gt;
&lt;br /&gt;
== Controversies and Ethical Concerns ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs has faced significant criticism regarding the misuse of its technology:&lt;br /&gt;
&lt;br /&gt;
=== Early Misuse Incidents ===&lt;br /&gt;
&lt;br /&gt;
Shortly after the beta launch in January 2023, the platform was exploited by users on 4chan and other forums to create fake audio content. Notable incidents included:&lt;br /&gt;
&lt;br /&gt;
* Creation of celebrity deepfakes, including voices of Emma Watson, Alexandria Ocasio-Cortez, and Ben Shapiro making statements they never made&lt;br /&gt;
* Generation of racist, sexist, and homophobic content using cloned voices&amp;lt;ref&amp;gt;https://www.vice.com/en/article/ai-voice-firm-4chan-celebrity-voices-emma-watson-joe-rogan-elevenlabs/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Political Deepfakes ===&lt;br /&gt;
&lt;br /&gt;
In January 2024, ElevenLabs&#039; technology was used to create a robocall impersonating President Joe Biden, urging New Hampshire voters not to participate in the Democratic primary. The incident prompted investigation by the New Hampshire Attorney General&#039;s office and led to the suspension of the responsible user account.&amp;lt;ref&amp;gt;https://www.bloomberg.com/news/articles/2024-01-26/ai-startup-elevenlabs-bans-account-blamed-for-biden-audio-deepfake&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Legal Challenges ===&lt;br /&gt;
&lt;br /&gt;
The company faces ongoing legal challenges, including:&lt;br /&gt;
&lt;br /&gt;
* A lawsuit from voice actors Mark Boyett and Karissa Vacker, alleging unauthorized use of their voices to create the &amp;quot;Adam&amp;quot; and &amp;quot;Bella&amp;quot; default voices&lt;br /&gt;
* Claims of copyright infringement related to the use of audiobook recordings for training&amp;lt;ref&amp;gt;https://www.thevoicerealm.com/blog/a-look-into-the-elevenlabs-lawsuit/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Safety Measures ===&lt;br /&gt;
&lt;br /&gt;
In response to misuse concerns, ElevenLabs has implemented several safeguards:&lt;br /&gt;
&lt;br /&gt;
* Verification requirements for voice cloning features&lt;br /&gt;
* AI Speech Classifier for detecting ElevenLabs-generated content&lt;br /&gt;
* Partnership with Reality Defender for deepfake detection&lt;br /&gt;
* Mandatory credit card information for certain features&amp;lt;ref&amp;gt;https://www.bloomberg.com/news/articles/2024-07-18/elevenlabs-partners-with-reality-defender-to-combat-deepfake-audio&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Applications and Use Cases ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs technology is utilized across various industries:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Media and Entertainment&#039;&#039;&#039;: Audiobook production, podcast creation, film dubbing&lt;br /&gt;
* &#039;&#039;&#039;Gaming&#039;&#039;&#039;: Character voice generation for video games&lt;br /&gt;
* &#039;&#039;&#039;Education&#039;&#039;&#039;: Educational content narration and language learning&lt;br /&gt;
* &#039;&#039;&#039;Enterprise&#039;&#039;&#039;: Customer service automation, training materials&lt;br /&gt;
* &#039;&#039;&#039;Accessibility&#039;&#039;&#039;: Tools for visually impaired users&lt;br /&gt;
&lt;br /&gt;
The company reports that 41% of Fortune 500 companies use its platform, with notable customers including The Washington Post, TIME magazine, and HarperCollins Publishers.&amp;lt;ref&amp;gt;https://sacra.com/c/elevenlabs/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://elevenlabs.io/ Official ElevenLabs website]&lt;br /&gt;
* [https://elevenlabs.io/docs/ ElevenLabs Documentation]&lt;br /&gt;
* [https://elevenlabs.io/text-to-speech Text-to-Speech Demo]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=VibeVoice&amp;diff=53</id>
		<title>VibeVoice</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=VibeVoice&amp;diff=53"/>
		<updated>2025-09-23T02:53:54Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add note about fine-tuning&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox TTS model&lt;br /&gt;
| name = VibeVoice&lt;br /&gt;
| developer = [[Microsoft Research]]&lt;br /&gt;
| release_date = August 26, 2025&lt;br /&gt;
| latest_version = 7B&lt;br /&gt;
| architecture = [[Qwen]] 2.5 + Diffusion&lt;br /&gt;
| parameters = 1.5B / 7B&lt;br /&gt;
| training_data = Proprietary dataset&lt;br /&gt;
| languages = English, Chinese&lt;br /&gt;
| voices = 4 speakers maximum&lt;br /&gt;
| voice_cloning = Yes&lt;br /&gt;
| emotion_control = Limited&lt;br /&gt;
| streaming = Yes&lt;br /&gt;
| license = MIT&lt;br /&gt;
| open_source = Limited (code removed)&lt;br /&gt;
| code_repository = [https://github.com/vibevoice-community/VibeVoice Community fork]&lt;br /&gt;
| model_weights = [https://huggingface.co/vibevoice Community backup]&lt;br /&gt;
| website = [https://aka.ms/VibeVoice Microsoft]&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;VibeVoice&#039;&#039;&#039; is an experimental [[text-to-speech]] (TTS) framework developed by [[Microsoft Research]] for generating long-form, multi-speaker conversational audio. It was released in August 2025 and is designed to synthesize long-form speech content such as podcasts and audiobooks with up to 4 speakers and with support for voice cloning.&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Development and Release ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice was developed by a team at Microsoft Research led by Zhiliang Peng, Jianwei Yu, and others, with the technical report published on [[arXiv]] in August 2025.&amp;lt;ref&amp;gt;https://arxiv.org/abs/2508.19205&amp;lt;/ref&amp;gt; The project was initially released as open-source software on GitHub and [[Hugging Face]], with model weights made publicly available under the MIT license.&lt;br /&gt;
&lt;br /&gt;
However, the release was disrupted in September 2025 when Microsoft removed the official repository and model weights from public access. According to Microsoft&#039;s statement, this action was taken after discovering &amp;quot;instances where the tool was used in ways inconsistent with the stated intent&amp;quot; and concerns about responsible AI use.&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt; The repository was later restored without implementation code, while community-maintained forks preserved the original materials. The 1.5B pretrained model remains available on Microsoft&#039;s Hugging Face page, but the 7B model was taken down.&lt;br /&gt;
&lt;br /&gt;
Community forks have preserved backups of the code and model weights, including both the 1.5B and 7B models.&amp;lt;ref&amp;gt;https://github.com/vibevoice-community/VibeVoice&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://huggingface.co/aoi-ot/VibeVoice-Large&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice uses a hybrid architecture combining large language models with diffusion-based audio generation. The system uses two specialized tokenizers operating at an ultra-low 7.5 Hz frame rate:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Acoustic Tokenizer&#039;&#039;&#039;: A [[variational autoencoder]] (VAE) based encoder-decoder that compresses audio signals while preserving fidelity&lt;br /&gt;
* &#039;&#039;&#039;Semantic Tokenizer&#039;&#039;&#039;: A content-focused encoder trained using [[automatic speech recognition]] as a proxy task&lt;br /&gt;
&lt;br /&gt;
The core model utilizes [[Qwen2.5]] as its base large language model (available in 1.5B and 7B parameter variants), integrated with a lightweight diffusion head for generating acoustic features. This design achieves what the researchers claim is an 80-fold improvement in data compression compared to the [[Encodec]] model while maintaining audio quality.&amp;lt;ref&amp;gt;https://arxiv.org/abs/2508.19205&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Capabilities and Limitations ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice can generate speech sequences up to 90 minutes in length with support for up to four distinct speakers. The model demonstrates several emergent capabilities not explicitly trained for, including:&lt;br /&gt;
&lt;br /&gt;
* Cross-lingual speech synthesis&lt;br /&gt;
* Spontaneous singing (though often off-key)&lt;br /&gt;
* Contextual background music generation&lt;br /&gt;
* Voice cloning from short prompts&lt;br /&gt;
&lt;br /&gt;
However, the system has notable limitations:&lt;br /&gt;
&lt;br /&gt;
* Language support restricted to English and Chinese&lt;br /&gt;
* No explicit modeling of overlapping speech&lt;br /&gt;
* Occasional instability, particularly with Chinese text synthesis&lt;br /&gt;
* Uncontrolled generation of background sounds and music&lt;br /&gt;
* Limited commercial viability due to various technical constraints&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
In comparative evaluations against contemporary TTS systems including [[ElevenLabs]], [[Google]]&#039;s Gemini 2.5 Pro TTS, and others, VibeVoice reportedly achieved superior scores in subjective metrics of realism, richness, and user preference. The model also demonstrated competitive [[word error rate]]s when evaluated using speech recognition systems.&amp;lt;ref&amp;gt;https://huggingface.co/microsoft/VibeVoice-1.5B&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, these evaluations were conducted on a limited test set of eight conversational transcripts totaling approximately one hour of audio, raising questions about the generalizability of the results to broader use cases.&lt;br /&gt;
&lt;br /&gt;
== Controversies and Concerns ==&lt;br /&gt;
&lt;br /&gt;
The temporary removal of VibeVoice from public access highlighted ongoing concerns about the potential misuse of high-quality synthetic speech technology. Microsoft explicitly warned about the potential for creating [[deepfake]] audio content for impersonation, fraud, or disinformation purposes.&lt;br /&gt;
&lt;br /&gt;
The model&#039;s ability to generate convincing speech from minimal voice prompts, combined with its long-form generation capabilities, raised particular concerns among AI safety researchers about potential misuse for creating fake audio content at scale.&lt;br /&gt;
&lt;br /&gt;
== Fine-Tuning ==&lt;br /&gt;
Following the release of the community backup, a member of the community has released fine-tuning scripts on GitHub. LoRA fine-tuning is fully supported; however, full-fine-tuning is not yet supported.&lt;br /&gt;
&lt;br /&gt;
Members of the Discord community have reported success both extending VibeVoice to new languages and training on specific voices for better voice cloning; however, there are few LoRA adapters publicly available.&lt;br /&gt;
&lt;br /&gt;
== Community Response ==&lt;br /&gt;
&lt;br /&gt;
Following Microsoft&#039;s temporary withdrawal of the official release, the open-source community created several preservation efforts:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;[https://github.com/vibevoice-community/VibeVoice vibevoice-community/VibeVoice]&#039;&#039;&#039;: A community-maintained fork preserving the original codebase and model weights&lt;br /&gt;
* &#039;&#039;&#039;[https://github.com/voicepowered-ai/VibeVoice-finetuning VibeVoice-finetuning]&#039;&#039;&#039;: Unofficial tools for fine-tuning the models using [[Low-Rank Adaptation]] (LoRA) techniques&lt;br /&gt;
&lt;br /&gt;
These community efforts have enabled continued research and development despite the official restrictions, though they operate independently of Microsoft&#039;s oversight.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/vibevoice-community/VibeVoice Community-maintained VibeVoice repository]&lt;br /&gt;
* [https://discord.com/invite/ZDEYTTRxWG VibeVoice Discord server (unofficial)]&lt;br /&gt;
* [https://arxiv.org/abs/2508.19205 Original technical report on arXiv]&lt;br /&gt;
&lt;br /&gt;
[[Category:Artificial intelligence]]&lt;br /&gt;
[[Category:Speech synthesis]]&lt;br /&gt;
[[Category:Microsoft Research]]&lt;br /&gt;
[[Category:Open-source software]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Mean_Opinion_Score&amp;diff=52</id>
		<title>Mean Opinion Score</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Mean_Opinion_Score&amp;diff=52"/>
		<updated>2025-09-22T02:52:24Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: fix list styling&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Mean Opinion Score&#039;&#039;&#039; (&#039;&#039;&#039;MOS&#039;&#039;&#039;) is a numerical measure used in telecommunications and multimedia engineering to represent the overall quality of a stimulus or system as perceived by human evaluators. It is calculated as the arithmetic mean of individual ratings given by test subjects on a predefined scale, typically ranging from 1 (lowest perceived quality) to 5 (highest perceived quality). MOS is widely used for evaluating voice, video, and audiovisual quality in applications ranging from traditional telephony to modern text-to-speech systems and streaming media.&lt;br /&gt;
&lt;br /&gt;
The methodology originated in the telecommunications industry for assessing telephone call quality and was formally standardized by the International Telecommunication Union (ITU-T) in Recommendation P.800 in 1996. Since then, it has become the gold standard for subjective quality assessment across various domains where human perception of quality is critical.&lt;br /&gt;
&lt;br /&gt;
== History and Development ==&lt;br /&gt;
&lt;br /&gt;
=== Early Origins ===&lt;br /&gt;
&lt;br /&gt;
The concept of Mean Opinion Score emerged in the telecommunications industry during the 1970s as telephone networks became more complex and digital transmission methods were introduced. Initially developed by the ITU, MOS provided a standardized way to assess voice transmission quality over telephone networks by aggregating human judgments of call quality.&amp;lt;ref&amp;gt;https://www.telecomtrainer.com/mos-mean-opinion-score/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The early methodology involved having listeners sit in controlled &amp;quot;quiet rooms&amp;quot; and score telephone call quality as they perceived it. This subjective testing approach had been in use in the telephony industry for decades before formal standardization, reflecting the industry&#039;s recognition that technical measurements alone could not capture the human experience of communication quality.&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/Mean_opinion_score&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Standardization ===&lt;br /&gt;
&lt;br /&gt;
The methodology was formally standardized in ITU-T Recommendation P.800, &amp;quot;Methods for subjective determination of transmission quality,&amp;quot; approved on August 30, 1996.&amp;lt;ref&amp;gt;https://www.itu.int/rec/T-REC-P.800-199608-I/en&amp;lt;/ref&amp;gt; This recommendation established rigorous protocols for conducting subjective quality tests, including specific requirements for test environments and procedures.&lt;br /&gt;
&lt;br /&gt;
The standardization specified that test subjects should be seated in quiet rooms with volumes between 30 and 120 cubic meters, reverberation times less than 500 milliseconds (preferably 200-300 ms), and room noise levels below 30 dBA with no dominant spectral peaks. These environmental controls ensured consistency and reliability in MOS evaluations across different testing facilities and organizations.&lt;br /&gt;
&lt;br /&gt;
=== Evolution and Extensions ===&lt;br /&gt;
&lt;br /&gt;
Following the success of P.800, the ITU-T developed additional recommendations to clarify and extend MOS methodology:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.800.1&#039;&#039;&#039; (2003, updated 2016): Established terminology for different types of MOS scores, distinguishing between listening quality subjective (MOS-LQS), listening quality objective (MOS-LQO), and listening quality estimated (MOS-LQE) to avoid confusion about the source and nature of scores.&amp;lt;ref&amp;gt;https://studylib.net/doc/8277727/itu-t-rec.-p.800.1--03-2003--mean-opinion-score--mos--ter&amp;lt;/ref&amp;gt;&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.800.2&#039;&#039;&#039;: Prescribed how MOS values should be reported, emphasizing that MOS scores from separate experiments cannot be directly compared unless explicitly designed for comparison and statistically validated.&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.808&#039;&#039;&#039; (2021): Addressed crowdsourcing methods for conducting subjective evaluations, recognizing the need for scalable approaches to MOS testing in the digital age.&amp;lt;ref&amp;gt;https://www.itu.int/rec/dologin_pub.asp?lang=e&amp;amp;id=T-REC-P.808-202106-I!!PDF-E&amp;amp;type=items&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Methodology ==&lt;br /&gt;
&lt;br /&gt;
=== Rating Scales ===&lt;br /&gt;
&lt;br /&gt;
The most commonly used rating scale is the &#039;&#039;&#039;Absolute Category Rating&#039;&#039;&#039; (ACR) scale, which maps subjective quality ratings to numerical values:&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Score !! Quality Level !! Description&lt;br /&gt;
|-&lt;br /&gt;
| 5 || Excellent || Completely natural speech; imperceptible artifacts&lt;br /&gt;
|-&lt;br /&gt;
| 4 || Good || Mostly natural speech; just perceptible but not annoying&lt;br /&gt;
|-&lt;br /&gt;
| 3 || Fair || Equally natural and unnatural; perceptible and slightly annoying&lt;br /&gt;
|-&lt;br /&gt;
| 2 || Poor || Mostly unnatural speech; annoying but not objectionable&lt;br /&gt;
|-&lt;br /&gt;
| 1 || Bad || Completely unnatural speech; very annoying and objectionable&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Alternative scales may use different ranges (e.g., 1-100) or different qualitative descriptors, depending on the specific application and testing requirements.&amp;lt;ref&amp;gt;https://www.twilio.com/docs/glossary/what-is-mean-opinion-score-mos&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Testing Procedures ===&lt;br /&gt;
&lt;br /&gt;
Traditional MOS testing involves several key steps:&lt;br /&gt;
&lt;br /&gt;
# &#039;&#039;&#039;Subject Selection&#039;&#039;&#039;: Recruiting appropriate test participants, typically naive listeners without specialized training in audio quality assessment&lt;br /&gt;
# &#039;&#039;&#039;Environment Control&#039;&#039;&#039;: Conducting tests in acoustically controlled environments meeting ITU-T specifications&lt;br /&gt;
# &#039;&#039;&#039;Stimulus Presentation&#039;&#039;&#039;: Playing audio samples to subjects in randomized order to minimize bias&lt;br /&gt;
# &#039;&#039;&#039;Rating Collection&#039;&#039;&#039;: Having subjects rate each stimulus on the chosen scale&lt;br /&gt;
# &#039;&#039;&#039;Statistical Analysis&#039;&#039;&#039;: Calculating the arithmetic mean and confidence intervals for each stimulus&lt;br /&gt;
&lt;br /&gt;
Modern extensions include comparative methods such as:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Degradation Category Rating&#039;&#039;&#039; (DCR): Subjects compare processed audio to a reference&lt;br /&gt;
* &#039;&#039;&#039;Comparison Category Rating&#039;&#039;&#039; (CCR): Direct comparison between two stimuli&lt;br /&gt;
&lt;br /&gt;
=== Objective Estimation ===&lt;br /&gt;
&lt;br /&gt;
While traditional MOS relies on human evaluation, objective models have been developed to predict MOS scores automatically. Key standardized methods include:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;PESQ&#039;&#039;&#039; (ITU-T P.862): Perceptual Evaluation of Speech Quality, introduced in 200&lt;br /&gt;
* &#039;&#039;&#039;POLQA&#039;&#039;&#039; (ITU-T P.863): Perceptual Objective Listening Quality Assessment, approved in 201&lt;br /&gt;
* &#039;&#039;&#039;PSQM&#039;&#039;&#039; (ITU-T P.861): Perceptual Speech Quality Measure, the first standardized method from 1997&lt;br /&gt;
&lt;br /&gt;
These algorithms analyze acoustic properties of audio signals to estimate human perceptual quality, enabling automated quality monitoring and real-time assessment.&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/Perceptual_Evaluation_of_Speech_Quality&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Applications ==&lt;br /&gt;
&lt;br /&gt;
=== Telecommunications ===&lt;br /&gt;
&lt;br /&gt;
MOS remains fundamental in telecommunications for evaluating voice and video call quality. In Voice over IP (VoIP) systems, MOS scores help assess the impact of network impairments such as packet loss, jitter, and latency on user experience. The G.711 codec, commonly used in VoIP, has a maximum theoretical MOS of 4.4, serving as a benchmark for quality comparisons.&amp;lt;ref&amp;gt;https://obkio.com/blog/measuring-voip-quality-with-mos-score-mean-opinion-score/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Telecommunications companies use MOS for:&lt;br /&gt;
&lt;br /&gt;
* Network planning and optimization&lt;br /&gt;
* Codec evaluation and selection&lt;br /&gt;
* Service level agreement monitoring&lt;br /&gt;
* Competitive benchmarking&lt;br /&gt;
&lt;br /&gt;
=== Speech Synthesis and AI ===&lt;br /&gt;
&lt;br /&gt;
In modern artificial intelligence applications, MOS has become critical for evaluating text-to-speech (TTS) systems. As synthetic speech quality has improved dramatically with neural approaches like WaveNet, Tacotron, and VITS, MOS remains the primary method for assessing how natural and human-like synthesized speech sounds to listeners.&lt;br /&gt;
&lt;br /&gt;
However, recent research has highlighted limitations of MOS for evaluating state-of-the-art speech synthesis systems. Studies have shown that as synthetic speech approaches human quality, MOS becomes less sensitive to remaining differences, leading researchers to explore complementary evaluation methods.&amp;lt;ref&amp;gt;https://www.sciencedirect.com/science/article/abs/pii/S0885230823000967&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Multimedia and Streaming ===&lt;br /&gt;
&lt;br /&gt;
MOS is extensively used in multimedia applications for evaluating:&lt;br /&gt;
&lt;br /&gt;
* Video streaming quality&lt;br /&gt;
* Audio codec performance&lt;br /&gt;
* Compression artifact assessment&lt;br /&gt;
* Real-time communication platforms&lt;br /&gt;
&lt;br /&gt;
Streaming services use MOS to optimize their delivery pipelines, balancing bandwidth efficiency with perceptual quality to ensure user satisfaction across diverse network conditions.&lt;br /&gt;
&lt;br /&gt;
== Modern Challenges and Limitations ==&lt;br /&gt;
&lt;br /&gt;
=== Statistical and Methodological Issues ===&lt;br /&gt;
&lt;br /&gt;
MOS faces several inherent limitations that researchers and practitioners must consider:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Ordinal Scale Problems&#039;&#039;&#039;: MOS ratings are based on ordinal scales where the ranking of items is known but intervals between ratings are not necessarily equal. Mathematically, calculating an arithmetic mean from ordinal data is problematic, and median values would be more appropriate. However, the practice of using arithmetic means is widely accepted and standardized.&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/Mean_opinion_score&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Range-Equalization Bias&#039;&#039;&#039;: Test subjects tend to use the full rating scale during an experiment, making scores relative to the range of quality present in the test rather than absolute measures of quality. This prevents direct comparison of MOS scores from different experiments.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contextual Dependence&#039;&#039;&#039;: MOS values are influenced by the testing context, participant demographics, and the presence of anchor stimuli (very high or low quality samples that influence perception of other stimuli).&lt;br /&gt;
&lt;br /&gt;
=== Scalability and Cost ===&lt;br /&gt;
&lt;br /&gt;
Traditional MOS testing is time-consuming and expensive, requiring recruitment of human evaluators and controlled testing environments. This has led to increased interest in:&lt;br /&gt;
&lt;br /&gt;
* Crowdsourcing platforms for distributed evaluation&lt;br /&gt;
* Objective quality models that predict MOS&lt;br /&gt;
* Automated evaluation metrics that correlate with human perception&lt;br /&gt;
&lt;br /&gt;
=== Limitations in Advanced Applications ===&lt;br /&gt;
&lt;br /&gt;
As technology advances, particularly in AI-generated content, traditional MOS evaluation faces new challenges:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Ceiling Effects&#039;&#039;&#039;: When synthetic speech approaches human quality, MOS becomes less discriminative, with most systems scoring in the 4.0-4.5 range where small differences may not be statistically significant.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Missing Dimensions&#039;&#039;&#039;: MOS provides only an overall quality rating and may miss specific aspects like speaker similarity, emotional expression, or intelligibility of specific linguistic phenomena.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Cultural and Linguistic Bias&#039;&#039;&#039;: MOS scores can vary based on evaluator demographics, language background, and cultural factors, potentially limiting generalizability across diverse user populations.&lt;br /&gt;
&lt;br /&gt;
== Contemporary Developments ==&lt;br /&gt;
&lt;br /&gt;
=== Crowdsourcing and Remote Testing ===&lt;br /&gt;
&lt;br /&gt;
ITU-T P.808 (2021) established guidelines for conducting MOS evaluations using crowdsourcing platforms, recognizing the need for scalable testing methods. This approach enables larger-scale evaluations but introduces new challenges in quality control and participant screening.&amp;lt;ref&amp;gt;https://www.itu.int/rec/dologin_pub.asp?lang=e&amp;amp;id=T-REC-P.808-202106-I!!PDF-E&amp;amp;type=items&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Deep Learning Integration ===&lt;br /&gt;
&lt;br /&gt;
Recent research explores using deep learning models for automatic MOS prediction, potentially enabling real-time quality assessment. Some approaches integrate MOS predictions into other tasks, such as fake audio detection, where predicted MOS scores help identify synthetic speech.&amp;lt;ref&amp;gt;https://arxiv.org/html/2401.13249v2&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Multi-Modal Assessment ===&lt;br /&gt;
&lt;br /&gt;
Modern applications increasingly require evaluation beyond audio quality alone. Research is extending MOS concepts to multi-modal scenarios, including audio-visual quality assessment and the evaluation of text-to-speech avatars that combine voice and visual synthesis.&lt;br /&gt;
&lt;br /&gt;
== Related Standards and Metrics ==&lt;br /&gt;
&lt;br /&gt;
Several ITU-T recommendations work in conjunction with MOS:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;ITU-T G.107&#039;&#039;&#039;: The E-model for objective quality assessment that can be mapped to MOS scales&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.910&#039;&#039;&#039;: Subjective video quality assessment methods&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.863&#039;&#039;&#039;: POLQA objective quality measurement&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.862&#039;&#039;&#039;: PESQ objective speech quality assessment&lt;br /&gt;
&lt;br /&gt;
Other quality metrics used alongside MOS include technical measurements like signal-to-noise ratio, mean squared error, and mel-cepstral distortion, though these objective measures often correlate poorly with human perception.&lt;br /&gt;
&lt;br /&gt;
== Criticism and Future Directions ==&lt;br /&gt;
&lt;br /&gt;
The speech synthesis research community has increasingly questioned whether MOS alone is sufficient for evaluating modern high-quality systems. Critics argue that the field may have reached &amp;quot;the end of a cul-de-sac by only evaluating the overall quality with MOS&amp;quot; and advocate for developing new evaluation protocols better suited to analyzing advanced speech synthesis technologies.&amp;lt;ref&amp;gt;https://www.research.ed.ac.uk/en/publications/the-limits-of-the-mean-opinion-score-for-speech-synthesis-evaluat&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Proposed alternatives and extensions include:&lt;br /&gt;
&lt;br /&gt;
* Fine-grained evaluation of specific quality dimensions&lt;br /&gt;
* Task-specific intelligibility testing&lt;br /&gt;
* Comparative ranking methods that avoid absolute scaling issues&lt;br /&gt;
* Objective metrics that better correlate with human perception&lt;br /&gt;
&lt;br /&gt;
Despite these limitations, MOS remains the most widely accepted method for subjective quality assessment and continues to evolve to meet the needs of advancing technology.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://www.itu.int/rec/T-REC-P.800/en ITU-T Recommendation P.800]&lt;br /&gt;
* [https://www.itu.int/rec/T-REC-P.800.1/en ITU-T Recommendation P.800.1]&lt;br /&gt;
* [https://www.itu.int/rec/T-REC-P.808/en ITU-T Recommendation P.808]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=VITS&amp;diff=51</id>
		<title>VITS</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=VITS&amp;diff=51"/>
		<updated>2025-09-22T02:51:59Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: relink&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;VITS&#039;&#039;&#039; (&#039;&#039;&#039;Variational Inference with adversarial learning for end-to-end Text-to-Speech&#039;&#039;&#039;) and &#039;&#039;&#039;VITS2&#039;&#039;&#039; are neural text-to-speech synthesis models that generate speech directly from text input using end-to-end training. VITS was first introduced by researchers at Kakao Enterprise in June 2021, while VITS2 was developed by SK Telecom and published in July 2023 as an improvement over the original model.&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
&lt;br /&gt;
Traditional text-to-speech systems typically employ a two-stage pipeline: first converting text to intermediate representations like mel-spectrograms, then generating audio waveforms from these representations. VITS introduced a single-stage approach that generates natural-sounding audio directly from text using variational inference augmented with normalizing flows and adversarial training.&lt;br /&gt;
&lt;br /&gt;
The models are notable for achieving quality comparable to human speech while maintaining parallel generation capabilities, making them significantly faster than autoregressive alternatives. Human evaluation on the LJ Speech dataset showed that VITS outperformed the best publicly available TTS systems at the time and achieved a [[Mean Opinion Score|mean opinion score (MOS)]] comparable to ground truth recordings.&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
=== VITS (2021) ===&lt;br /&gt;
&lt;br /&gt;
VITS employs a conditional variational autoencoder (VAE) framework combined with several advanced techniques:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Core Components:&#039;&#039;&#039;&lt;br /&gt;
* &#039;&#039;&#039;Posterior Encoder:&#039;&#039;&#039; Processes linear-scale spectrograms during training to learn latent representations&lt;br /&gt;
* &#039;&#039;&#039;Prior Encoder:&#039;&#039;&#039; Contains a text encoder and normalizing flows to model the prior distribution of latent variables&lt;br /&gt;
* &#039;&#039;&#039;Decoder:&#039;&#039;&#039; Based on HiFi-GAN V1 generator, converts latent variables to raw waveforms&lt;br /&gt;
* &#039;&#039;&#039;Discriminator:&#039;&#039;&#039; Multi-period discriminator from HiFi-GAN for adversarial training&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Key Innovations:&#039;&#039;&#039;&lt;br /&gt;
* &#039;&#039;&#039;Monotonic Alignment Search (MAS):&#039;&#039;&#039; Automatically learns alignments between text and speech without external annotations by finding alignments that maximize the likelihood of target speech.&lt;br /&gt;
* &#039;&#039;&#039;Stochastic Duration Predictor:&#039;&#039;&#039; Uses normalizing flows to model the distribution of phoneme durations, enabling synthesis of speech with diverse rhythms from the same text input.&lt;br /&gt;
* &#039;&#039;&#039;Adversarial Training:&#039;&#039;&#039; Improves waveform quality through generator-discriminator competition&lt;br /&gt;
&lt;br /&gt;
The model addresses the one-to-many relationship in speech synthesis, where a single text input can be spoken in multiple ways with different pitches, rhythms, and prosodic patterns.&lt;br /&gt;
&lt;br /&gt;
=== VITS2 (2023) ===&lt;br /&gt;
&lt;br /&gt;
VITS2 introduced several improvements over the original model to address issues including intermittent unnaturalness, computational efficiency, and strong dependence on phoneme conversion.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Major Improvements:&#039;&#039;&#039;&lt;br /&gt;
* &#039;&#039;&#039;Adversarial Duration Predictor:&#039;&#039;&#039; Replaced the flow-based stochastic duration predictor with one trained through adversarial learning, using a time step-wise conditional discriminator to improve efficiency and naturalness.&lt;br /&gt;
* &#039;&#039;&#039;Enhanced Normalizing Flows:&#039;&#039;&#039; Added transformer blocks to normalizing flows to capture long-term dependencies when transforming distributions, addressing limitations of convolution-only approaches.&lt;br /&gt;
* &#039;&#039;&#039;Improved Alignment Search:&#039;&#039;&#039; Modified Monotonic Alignment Search by adding Gaussian noise to calculated probabilities, giving the model additional opportunities to explore alternative alignments during early training.&lt;br /&gt;
* &#039;&#039;&#039;Speaker-Conditioned Text Encoder:&#039;&#039;&#039; For multi-speaker models, conditioned the speaker vector on the third transformer block of the text encoder to better capture speaker-specific pronunciation and intonation characteristics.&lt;br /&gt;
&lt;br /&gt;
== Performance and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
=== VITS Results ===&lt;br /&gt;
&lt;br /&gt;
On the LJ Speech dataset, VITS achieved a MOS of 4.43 (±0.06), compared to 4.46 (±0.06) for ground truth recordings. This outperformed Tacotron 2 + HiFi-GAN at 4.25 (±0.07) and Glow-TTS + HiFi-GAN at 4.32 (±0.07). The model also demonstrated significant improvements in synthesis speed, achieving 67.12× real-time factor compared to 27.48× for Glow-TTS + HiFi-GAN.&lt;br /&gt;
&lt;br /&gt;
=== VITS2 Results ===&lt;br /&gt;
&lt;br /&gt;
VITS2 showed further improvements with a MOS of 4.47 (±0.06) on LJ Speech, representing a 0.09 point increase over VITS. In comparative evaluations, VITS2 achieved a CMOS of 0.201 (±0.105) when compared directly to VITS. The model also improved synthesis speed to 97.25× real-time and reduced training time by approximately 22.7%.&lt;br /&gt;
&lt;br /&gt;
=== Multi-Speaker Capabilities ===&lt;br /&gt;
&lt;br /&gt;
Both models support multi-speaker synthesis. On the VCTK dataset containing 109 speakers, VITS achieved 4.38 (±0.06) MOS compared to 3.82 (±0.07) for Glow-TTS + HiFi-GAN. VITS2 further improved speaker similarity with a MOS of 3.99 (±0.08) compared to VITS&#039;s 3.79 (±0.09) on similarity evaluations.&lt;br /&gt;
&lt;br /&gt;
== End-to-End Capabilities ==&lt;br /&gt;
&lt;br /&gt;
A significant contribution of VITS2 was reducing dependence on phoneme conversion. Using character error rate (CER) evaluation with automatic speech recognition, VITS2 achieved 4.01% CER when using normalized text input compared to 3.92% with phoneme sequences, demonstrating the possibility of fully end-to-end training without explicit phoneme preprocessing.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/jaywalnut310/vits Official VITS repository]&lt;br /&gt;
* [https://github.com/daniilrobnikov/vits2 VITS2 implementation]&lt;br /&gt;
* [https://jaywalnut310.github.io/vits-demo/index.html VITS demo page]&lt;br /&gt;
* [https://vits-2.github.io/demo/ VITS2 demo page]&lt;br /&gt;
* [https://huggingface.co/docs/transformers/model_doc/vits VITS on Hugging Face]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Mean_Opinion_Score&amp;diff=50</id>
		<title>Mean Opinion Score</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Mean_Opinion_Score&amp;diff=50"/>
		<updated>2025-09-22T02:51:14Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add MOS&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Mean Opinion Score&#039;&#039;&#039; (&#039;&#039;&#039;MOS&#039;&#039;&#039;) is a numerical measure used in telecommunications and multimedia engineering to represent the overall quality of a stimulus or system as perceived by human evaluators. It is calculated as the arithmetic mean of individual ratings given by test subjects on a predefined scale, typically ranging from 1 (lowest perceived quality) to 5 (highest perceived quality). MOS is widely used for evaluating voice, video, and audiovisual quality in applications ranging from traditional telephony to modern text-to-speech systems and streaming media.&lt;br /&gt;
&lt;br /&gt;
The methodology originated in the telecommunications industry for assessing telephone call quality and was formally standardized by the International Telecommunication Union (ITU-T) in Recommendation P.800 in 1996. Since then, it has become the gold standard for subjective quality assessment across various domains where human perception of quality is critical.&lt;br /&gt;
&lt;br /&gt;
== History and Development ==&lt;br /&gt;
&lt;br /&gt;
=== Early Origins ===&lt;br /&gt;
&lt;br /&gt;
The concept of Mean Opinion Score emerged in the telecommunications industry during the 1970s as telephone networks became more complex and digital transmission methods were introduced. Initially developed by the ITU, MOS provided a standardized way to assess voice transmission quality over telephone networks by aggregating human judgments of call quality.&amp;lt;ref&amp;gt;https://www.telecomtrainer.com/mos-mean-opinion-score/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The early methodology involved having listeners sit in controlled &amp;quot;quiet rooms&amp;quot; and score telephone call quality as they perceived it. This subjective testing approach had been in use in the telephony industry for decades before formal standardization, reflecting the industry&#039;s recognition that technical measurements alone could not capture the human experience of communication quality.&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/Mean_opinion_score&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Standardization ===&lt;br /&gt;
&lt;br /&gt;
The methodology was formally standardized in ITU-T Recommendation P.800, &amp;quot;Methods for subjective determination of transmission quality,&amp;quot; approved on August 30, 1996.&amp;lt;ref&amp;gt;https://www.itu.int/rec/T-REC-P.800-199608-I/en&amp;lt;/ref&amp;gt; This recommendation established rigorous protocols for conducting subjective quality tests, including specific requirements for test environments and procedures.&lt;br /&gt;
&lt;br /&gt;
The standardization specified that test subjects should be seated in quiet rooms with volumes between 30 and 120 cubic meters, reverberation times less than 500 milliseconds (preferably 200-300 ms), and room noise levels below 30 dBA with no dominant spectral peaks. These environmental controls ensured consistency and reliability in MOS evaluations across different testing facilities and organizations.&lt;br /&gt;
&lt;br /&gt;
=== Evolution and Extensions ===&lt;br /&gt;
&lt;br /&gt;
Following the success of P.800, the ITU-T developed additional recommendations to clarify and extend MOS methodology:&lt;br /&gt;
&lt;br /&gt;
- &#039;&#039;&#039;ITU-T P.800.1&#039;&#039;&#039; (2003, updated 2016): Established terminology for different types of MOS scores, distinguishing between listening quality subjective (MOS-LQS), listening quality objective (MOS-LQO), and listening quality estimated (MOS-LQE) to avoid confusion about the source and nature of scores.&amp;lt;ref&amp;gt;https://studylib.net/doc/8277727/itu-t-rec.-p.800.1--03-2003--mean-opinion-score--mos--ter&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
- &#039;&#039;&#039;ITU-T P.800.2&#039;&#039;&#039;: Prescribed how MOS values should be reported, emphasizing that MOS scores from separate experiments cannot be directly compared unless explicitly designed for comparison and statistically validated.&lt;br /&gt;
&lt;br /&gt;
- &#039;&#039;&#039;ITU-T P.808&#039;&#039;&#039; (2021): Addressed crowdsourcing methods for conducting subjective evaluations, recognizing the need for scalable approaches to MOS testing in the digital age.&amp;lt;ref&amp;gt;https://www.itu.int/rec/dologin_pub.asp?lang=e&amp;amp;id=T-REC-P.808-202106-I!!PDF-E&amp;amp;type=items&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Methodology ==&lt;br /&gt;
&lt;br /&gt;
=== Rating Scales ===&lt;br /&gt;
&lt;br /&gt;
The most commonly used rating scale is the &#039;&#039;&#039;Absolute Category Rating&#039;&#039;&#039; (ACR) scale, which maps subjective quality ratings to numerical values:&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Score !! Quality Level !! Description&lt;br /&gt;
|-&lt;br /&gt;
| 5 || Excellent || Completely natural speech; imperceptible artifacts&lt;br /&gt;
|-&lt;br /&gt;
| 4 || Good || Mostly natural speech; just perceptible but not annoying&lt;br /&gt;
|-&lt;br /&gt;
| 3 || Fair || Equally natural and unnatural; perceptible and slightly annoying&lt;br /&gt;
|-&lt;br /&gt;
| 2 || Poor || Mostly unnatural speech; annoying but not objectionable&lt;br /&gt;
|-&lt;br /&gt;
| 1 || Bad || Completely unnatural speech; very annoying and objectionable&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Alternative scales may use different ranges (e.g., 1-100) or different qualitative descriptors, depending on the specific application and testing requirements.&amp;lt;ref&amp;gt;https://www.twilio.com/docs/glossary/what-is-mean-opinion-score-mos&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Testing Procedures ===&lt;br /&gt;
&lt;br /&gt;
Traditional MOS testing involves several key steps:&lt;br /&gt;
&lt;br /&gt;
# &#039;&#039;&#039;Subject Selection&#039;&#039;&#039;: Recruiting appropriate test participants, typically naive listeners without specialized training in audio quality assessment&lt;br /&gt;
# &#039;&#039;&#039;Environment Control&#039;&#039;&#039;: Conducting tests in acoustically controlled environments meeting ITU-T specifications&lt;br /&gt;
# &#039;&#039;&#039;Stimulus Presentation&#039;&#039;&#039;: Playing audio samples to subjects in randomized order to minimize bias&lt;br /&gt;
# &#039;&#039;&#039;Rating Collection&#039;&#039;&#039;: Having subjects rate each stimulus on the chosen scale&lt;br /&gt;
# &#039;&#039;&#039;Statistical Analysis&#039;&#039;&#039;: Calculating the arithmetic mean and confidence intervals for each stimulus&lt;br /&gt;
&lt;br /&gt;
Modern extensions include comparative methods such as:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Degradation Category Rating&#039;&#039;&#039; (DCR): Subjects compare processed audio to a reference&lt;br /&gt;
* &#039;&#039;&#039;Comparison Category Rating&#039;&#039;&#039; (CCR): Direct comparison between two stimuli&lt;br /&gt;
&lt;br /&gt;
=== Objective Estimation ===&lt;br /&gt;
&lt;br /&gt;
While traditional MOS relies on human evaluation, objective models have been developed to predict MOS scores automatically. Key standardized methods include:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;PESQ&#039;&#039;&#039; (ITU-T P.862): Perceptual Evaluation of Speech Quality, introduced in 200&lt;br /&gt;
* &#039;&#039;&#039;POLQA&#039;&#039;&#039; (ITU-T P.863): Perceptual Objective Listening Quality Assessment, approved in 201&lt;br /&gt;
* &#039;&#039;&#039;PSQM&#039;&#039;&#039; (ITU-T P.861): Perceptual Speech Quality Measure, the first standardized method from 1997&lt;br /&gt;
&lt;br /&gt;
These algorithms analyze acoustic properties of audio signals to estimate human perceptual quality, enabling automated quality monitoring and real-time assessment.&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/Perceptual_Evaluation_of_Speech_Quality&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Applications ==&lt;br /&gt;
&lt;br /&gt;
=== Telecommunications ===&lt;br /&gt;
&lt;br /&gt;
MOS remains fundamental in telecommunications for evaluating voice and video call quality. In Voice over IP (VoIP) systems, MOS scores help assess the impact of network impairments such as packet loss, jitter, and latency on user experience. The G.711 codec, commonly used in VoIP, has a maximum theoretical MOS of 4.4, serving as a benchmark for quality comparisons.&amp;lt;ref&amp;gt;https://obkio.com/blog/measuring-voip-quality-with-mos-score-mean-opinion-score/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Telecommunications companies use MOS for:&lt;br /&gt;
&lt;br /&gt;
* Network planning and optimization&lt;br /&gt;
* Codec evaluation and selection&lt;br /&gt;
* Service level agreement monitoring&lt;br /&gt;
* Competitive benchmarking&lt;br /&gt;
&lt;br /&gt;
=== Speech Synthesis and AI ===&lt;br /&gt;
&lt;br /&gt;
In modern artificial intelligence applications, MOS has become critical for evaluating text-to-speech (TTS) systems. As synthetic speech quality has improved dramatically with neural approaches like WaveNet, Tacotron, and VITS, MOS remains the primary method for assessing how natural and human-like synthesized speech sounds to listeners.&lt;br /&gt;
&lt;br /&gt;
However, recent research has highlighted limitations of MOS for evaluating state-of-the-art speech synthesis systems. Studies have shown that as synthetic speech approaches human quality, MOS becomes less sensitive to remaining differences, leading researchers to explore complementary evaluation methods.&amp;lt;ref&amp;gt;https://www.sciencedirect.com/science/article/abs/pii/S0885230823000967&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Multimedia and Streaming ===&lt;br /&gt;
&lt;br /&gt;
MOS is extensively used in multimedia applications for evaluating:&lt;br /&gt;
&lt;br /&gt;
* Video streaming quality&lt;br /&gt;
* Audio codec performance&lt;br /&gt;
* Compression artifact assessment&lt;br /&gt;
* Real-time communication platforms&lt;br /&gt;
&lt;br /&gt;
Streaming services use MOS to optimize their delivery pipelines, balancing bandwidth efficiency with perceptual quality to ensure user satisfaction across diverse network conditions.&lt;br /&gt;
&lt;br /&gt;
== Modern Challenges and Limitations ==&lt;br /&gt;
&lt;br /&gt;
=== Statistical and Methodological Issues ===&lt;br /&gt;
&lt;br /&gt;
MOS faces several inherent limitations that researchers and practitioners must consider:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Ordinal Scale Problems&#039;&#039;&#039;: MOS ratings are based on ordinal scales where the ranking of items is known but intervals between ratings are not necessarily equal. Mathematically, calculating an arithmetic mean from ordinal data is problematic, and median values would be more appropriate. However, the practice of using arithmetic means is widely accepted and standardized.&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/Mean_opinion_score&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Range-Equalization Bias&#039;&#039;&#039;: Test subjects tend to use the full rating scale during an experiment, making scores relative to the range of quality present in the test rather than absolute measures of quality. This prevents direct comparison of MOS scores from different experiments.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contextual Dependence&#039;&#039;&#039;: MOS values are influenced by the testing context, participant demographics, and the presence of anchor stimuli (very high or low quality samples that influence perception of other stimuli).&lt;br /&gt;
&lt;br /&gt;
=== Scalability and Cost ===&lt;br /&gt;
&lt;br /&gt;
Traditional MOS testing is time-consuming and expensive, requiring recruitment of human evaluators and controlled testing environments. This has led to increased interest in:&lt;br /&gt;
&lt;br /&gt;
* Crowdsourcing platforms for distributed evaluation&lt;br /&gt;
* Objective quality models that predict MOS&lt;br /&gt;
* Automated evaluation metrics that correlate with human perception&lt;br /&gt;
&lt;br /&gt;
=== Limitations in Advanced Applications ===&lt;br /&gt;
&lt;br /&gt;
As technology advances, particularly in AI-generated content, traditional MOS evaluation faces new challenges:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Ceiling Effects&#039;&#039;&#039;: When synthetic speech approaches human quality, MOS becomes less discriminative, with most systems scoring in the 4.0-4.5 range where small differences may not be statistically significant.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Missing Dimensions&#039;&#039;&#039;: MOS provides only an overall quality rating and may miss specific aspects like speaker similarity, emotional expression, or intelligibility of specific linguistic phenomena.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Cultural and Linguistic Bias&#039;&#039;&#039;: MOS scores can vary based on evaluator demographics, language background, and cultural factors, potentially limiting generalizability across diverse user populations.&lt;br /&gt;
&lt;br /&gt;
== Contemporary Developments ==&lt;br /&gt;
&lt;br /&gt;
=== Crowdsourcing and Remote Testing ===&lt;br /&gt;
&lt;br /&gt;
ITU-T P.808 (2021) established guidelines for conducting MOS evaluations using crowdsourcing platforms, recognizing the need for scalable testing methods. This approach enables larger-scale evaluations but introduces new challenges in quality control and participant screening.&amp;lt;ref&amp;gt;https://www.itu.int/rec/dologin_pub.asp?lang=e&amp;amp;id=T-REC-P.808-202106-I!!PDF-E&amp;amp;type=items&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Deep Learning Integration ===&lt;br /&gt;
&lt;br /&gt;
Recent research explores using deep learning models for automatic MOS prediction, potentially enabling real-time quality assessment. Some approaches integrate MOS predictions into other tasks, such as fake audio detection, where predicted MOS scores help identify synthetic speech.&amp;lt;ref&amp;gt;https://arxiv.org/html/2401.13249v2&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Multi-Modal Assessment ===&lt;br /&gt;
&lt;br /&gt;
Modern applications increasingly require evaluation beyond audio quality alone. Research is extending MOS concepts to multi-modal scenarios, including audio-visual quality assessment and the evaluation of text-to-speech avatars that combine voice and visual synthesis.&lt;br /&gt;
&lt;br /&gt;
== Related Standards and Metrics ==&lt;br /&gt;
&lt;br /&gt;
Several ITU-T recommendations work in conjunction with MOS:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;ITU-T G.107&#039;&#039;&#039;: The E-model for objective quality assessment that can be mapped to MOS scales&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.910&#039;&#039;&#039;: Subjective video quality assessment methods&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.863&#039;&#039;&#039;: POLQA objective quality measurement&lt;br /&gt;
* &#039;&#039;&#039;ITU-T P.862&#039;&#039;&#039;: PESQ objective speech quality assessment&lt;br /&gt;
&lt;br /&gt;
Other quality metrics used alongside MOS include technical measurements like signal-to-noise ratio, mean squared error, and mel-cepstral distortion, though these objective measures often correlate poorly with human perception.&lt;br /&gt;
&lt;br /&gt;
== Criticism and Future Directions ==&lt;br /&gt;
&lt;br /&gt;
The speech synthesis research community has increasingly questioned whether MOS alone is sufficient for evaluating modern high-quality systems. Critics argue that the field may have reached &amp;quot;the end of a cul-de-sac by only evaluating the overall quality with MOS&amp;quot; and advocate for developing new evaluation protocols better suited to analyzing advanced speech synthesis technologies.&amp;lt;ref&amp;gt;https://www.research.ed.ac.uk/en/publications/the-limits-of-the-mean-opinion-score-for-speech-synthesis-evaluat&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Proposed alternatives and extensions include:&lt;br /&gt;
&lt;br /&gt;
* Fine-grained evaluation of specific quality dimensions&lt;br /&gt;
* Task-specific intelligibility testing&lt;br /&gt;
* Comparative ranking methods that avoid absolute scaling issues&lt;br /&gt;
* Objective metrics that better correlate with human perception&lt;br /&gt;
&lt;br /&gt;
Despite these limitations, MOS remains the most widely accepted method for subjective quality assessment and continues to evolve to meet the needs of advancing technology.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://www.itu.int/rec/T-REC-P.800/en ITU-T Recommendation P.800]&lt;br /&gt;
* [https://www.itu.int/rec/T-REC-P.800.1/en ITU-T Recommendation P.800.1]&lt;br /&gt;
* [https://www.itu.int/rec/T-REC-P.808/en ITU-T Recommendation P.808]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=VITS&amp;diff=49</id>
		<title>VITS</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=VITS&amp;diff=49"/>
		<updated>2025-09-22T02:39:06Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;&amp;#039;&amp;#039;&amp;#039;VITS&amp;#039;&amp;#039;&amp;#039; (&amp;#039;&amp;#039;&amp;#039;Variational Inference with adversarial learning for end-to-end Text-to-Speech&amp;#039;&amp;#039;&amp;#039;) and &amp;#039;&amp;#039;&amp;#039;VITS2&amp;#039;&amp;#039;&amp;#039; are neural text-to-speech synthesis models that generate speech directly from text input using end-to-end training. VITS was first introduced by researchers at Kakao Enterprise in June 2021, while VITS2 was developed by SK Telecom and published in July 2023 as an improvement over the original model.  == Overview ==  Traditional text-to-speech systems typically...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;VITS&#039;&#039;&#039; (&#039;&#039;&#039;Variational Inference with adversarial learning for end-to-end Text-to-Speech&#039;&#039;&#039;) and &#039;&#039;&#039;VITS2&#039;&#039;&#039; are neural text-to-speech synthesis models that generate speech directly from text input using end-to-end training. VITS was first introduced by researchers at Kakao Enterprise in June 2021, while VITS2 was developed by SK Telecom and published in July 2023 as an improvement over the original model.&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
&lt;br /&gt;
Traditional text-to-speech systems typically employ a two-stage pipeline: first converting text to intermediate representations like mel-spectrograms, then generating audio waveforms from these representations. VITS introduced a single-stage approach that generates natural-sounding audio directly from text using variational inference augmented with normalizing flows and adversarial training.&lt;br /&gt;
&lt;br /&gt;
The models are notable for achieving quality comparable to human speech while maintaining parallel generation capabilities, making them significantly faster than autoregressive alternatives. Human evaluation on the LJ Speech dataset showed that VITS outperformed the best publicly available TTS systems at the time and achieved a [[Mean opinion score|mean opinion score (MOS)]] comparable to ground truth recordings.&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
=== VITS (2021) ===&lt;br /&gt;
&lt;br /&gt;
VITS employs a conditional variational autoencoder (VAE) framework combined with several advanced techniques:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Core Components:&#039;&#039;&#039;&lt;br /&gt;
* &#039;&#039;&#039;Posterior Encoder:&#039;&#039;&#039; Processes linear-scale spectrograms during training to learn latent representations&lt;br /&gt;
* &#039;&#039;&#039;Prior Encoder:&#039;&#039;&#039; Contains a text encoder and normalizing flows to model the prior distribution of latent variables&lt;br /&gt;
* &#039;&#039;&#039;Decoder:&#039;&#039;&#039; Based on HiFi-GAN V1 generator, converts latent variables to raw waveforms&lt;br /&gt;
* &#039;&#039;&#039;Discriminator:&#039;&#039;&#039; Multi-period discriminator from HiFi-GAN for adversarial training&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Key Innovations:&#039;&#039;&#039;&lt;br /&gt;
* &#039;&#039;&#039;Monotonic Alignment Search (MAS):&#039;&#039;&#039; Automatically learns alignments between text and speech without external annotations by finding alignments that maximize the likelihood of target speech.&lt;br /&gt;
* &#039;&#039;&#039;Stochastic Duration Predictor:&#039;&#039;&#039; Uses normalizing flows to model the distribution of phoneme durations, enabling synthesis of speech with diverse rhythms from the same text input.&lt;br /&gt;
* &#039;&#039;&#039;Adversarial Training:&#039;&#039;&#039; Improves waveform quality through generator-discriminator competition&lt;br /&gt;
&lt;br /&gt;
The model addresses the one-to-many relationship in speech synthesis, where a single text input can be spoken in multiple ways with different pitches, rhythms, and prosodic patterns.&lt;br /&gt;
&lt;br /&gt;
=== VITS2 (2023) ===&lt;br /&gt;
&lt;br /&gt;
VITS2 introduced several improvements over the original model to address issues including intermittent unnaturalness, computational efficiency, and strong dependence on phoneme conversion.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Major Improvements:&#039;&#039;&#039;&lt;br /&gt;
* &#039;&#039;&#039;Adversarial Duration Predictor:&#039;&#039;&#039; Replaced the flow-based stochastic duration predictor with one trained through adversarial learning, using a time step-wise conditional discriminator to improve efficiency and naturalness.&lt;br /&gt;
* &#039;&#039;&#039;Enhanced Normalizing Flows:&#039;&#039;&#039; Added transformer blocks to normalizing flows to capture long-term dependencies when transforming distributions, addressing limitations of convolution-only approaches.&lt;br /&gt;
* &#039;&#039;&#039;Improved Alignment Search:&#039;&#039;&#039; Modified Monotonic Alignment Search by adding Gaussian noise to calculated probabilities, giving the model additional opportunities to explore alternative alignments during early training.&lt;br /&gt;
* &#039;&#039;&#039;Speaker-Conditioned Text Encoder:&#039;&#039;&#039; For multi-speaker models, conditioned the speaker vector on the third transformer block of the text encoder to better capture speaker-specific pronunciation and intonation characteristics.&lt;br /&gt;
&lt;br /&gt;
== Performance and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
=== VITS Results ===&lt;br /&gt;
&lt;br /&gt;
On the LJ Speech dataset, VITS achieved a MOS of 4.43 (±0.06), compared to 4.46 (±0.06) for ground truth recordings. This outperformed Tacotron 2 + HiFi-GAN at 4.25 (±0.07) and Glow-TTS + HiFi-GAN at 4.32 (±0.07). The model also demonstrated significant improvements in synthesis speed, achieving 67.12× real-time factor compared to 27.48× for Glow-TTS + HiFi-GAN.&lt;br /&gt;
&lt;br /&gt;
=== VITS2 Results ===&lt;br /&gt;
&lt;br /&gt;
VITS2 showed further improvements with a MOS of 4.47 (±0.06) on LJ Speech, representing a 0.09 point increase over VITS. In comparative evaluations, VITS2 achieved a CMOS of 0.201 (±0.105) when compared directly to VITS. The model also improved synthesis speed to 97.25× real-time and reduced training time by approximately 22.7%.&lt;br /&gt;
&lt;br /&gt;
=== Multi-Speaker Capabilities ===&lt;br /&gt;
&lt;br /&gt;
Both models support multi-speaker synthesis. On the VCTK dataset containing 109 speakers, VITS achieved 4.38 (±0.06) MOS compared to 3.82 (±0.07) for Glow-TTS + HiFi-GAN. VITS2 further improved speaker similarity with a MOS of 3.99 (±0.08) compared to VITS&#039;s 3.79 (±0.09) on similarity evaluations.&lt;br /&gt;
&lt;br /&gt;
== End-to-End Capabilities ==&lt;br /&gt;
&lt;br /&gt;
A significant contribution of VITS2 was reducing dependence on phoneme conversion. Using character error rate (CER) evaluation with automatic speech recognition, VITS2 achieved 4.01% CER when using normalized text input compared to 3.92% with phoneme sequences, demonstrating the possibility of fully end-to-end training without explicit phoneme preprocessing.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/jaywalnut310/vits Official VITS repository]&lt;br /&gt;
* [https://github.com/daniilrobnikov/vits2 VITS2 implementation]&lt;br /&gt;
* [https://jaywalnut310.github.io/vits-demo/index.html VITS demo page]&lt;br /&gt;
* [https://vits-2.github.io/demo/ VITS2 demo page]&lt;br /&gt;
* [https://huggingface.co/docs/transformers/model_doc/vits VITS on Hugging Face]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=48</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=48"/>
		<updated>2025-09-22T02:27:23Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: /* Open Source Models */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
|[[VoxCPM]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English, Chinese&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[IndexTTS2]]&lt;br /&gt;
|Custom (restrictive)&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ✅ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Chatterbox]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[MegaTTS 3]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English, Chinese&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Orpheus TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CSM-1B]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Kokoro-82M]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CosyVoice 2.0]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|Chinese, English, Japanese, Korean&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2024&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
|[[XTTSv2]]&lt;br /&gt;
|CPML (restrictive)&lt;br /&gt;
|English, +16&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2023&lt;br /&gt;
|-&lt;br /&gt;
|[[Tortoise TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|2022&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=IndexTTS2&amp;diff=47</id>
		<title>IndexTTS2</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=IndexTTS2&amp;diff=47"/>
		<updated>2025-09-21T20:46:02Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add IndexTTS 2&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox TTS model|name=IndexTTS2|developer=Bilibili AI Platform Department|release_date=September 2025 (paper)|latest_version=2.0|architecture=Autoregressive Transformer|parameters=Undisclosed|training_data=55,000 hours multilingual|languages=Chinese, English, Japanese|voices=Zero-shot voice cloning|voice_cloning=Yes (emotion-timbre disentangled)|emotion_control=Yes (multimodal input)|streaming=Yes|latency=Not specified|license=Custom/restrictive (commercial license available)|open_source=Yes|code_repository=https://github.com/index-tts/index-tts|model_weights=https://huggingface.co/IndexTeam/IndexTTS-2|demo=https://index-tts.github.io/index-tts2.github.io/|website=https://indextts2.org}}&#039;&#039;&#039;IndexTTS2&#039;&#039;&#039; is an open-source text-to-speech model developed by Bilibili&#039;s AI Platform Department loosely based on [[Tortoise TTS]]. Released in September 2025, it addresses key limitations of traditional TTS models by introducing precise duration control and advanced emotional expression capabilities while maintaining the naturalness advantages of autoregressive generation.&lt;br /&gt;
&lt;br /&gt;
== Development and Background ==&lt;br /&gt;
IndexTTS2 was developed by a team led by Siyi Zhou at Bilibili&#039;s Artificial Intelligence Platform Department in China. The project builds upon the earlier IndexTTS model, incorporating substantial improvements in duration control, emotional modeling, and speech stability. The research was published on arXiv in September 2025, with the paper titled &amp;quot;IndexTTS2: A Breakthrough in Emotionally Expressive and Duration-Controlled Auto-Regressive Zero-Shot Text-to-Speech.&amp;quot;&amp;lt;ref&amp;gt;https://arxiv.org/abs/2506.21619&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The development was motivated by specific limitations in existing autoregressive TTS models, particularly their inability to precisely control speech duration - a critical requirement for applications such as video dubbing that demand strict audio-visual synchronization. Additionally, the team sought to address the limited emotional expressiveness of existing systems, which are often constrained by scarce high-quality emotional training data.&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
IndexTTS2 employs a three-module cascaded architecture:&lt;br /&gt;
&lt;br /&gt;
=== Text-to-Semantic (T2S) Module ===&lt;br /&gt;
The T2S module serves as the core component, utilizing an autoregressive Transformer framework to generate semantic tokens from input text, timbre prompts, style prompts, and optional speech token counts. Key innovations include:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Duration Control Mechanism&#039;&#039;&#039;: A novel duration encoding system where duration information p is computed from target semantic token length T using the formula p = W_num h(T), where W_num represents an embedding table and h(T) returns a one-hot vector&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Emotion-Speaker Disentanglement&#039;&#039;&#039;: Implementation of a Gradient Reversal Layer (GRL) to separate emotional features from speaker-specific characteristics, enabling independent control over timbre and emotion&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Three-Stage Training Strategy&#039;&#039;&#039;: A progressive training approach designed to overcome the scarcity of high-quality emotional data while maintaining model stability&lt;br /&gt;
&lt;br /&gt;
=== Semantic-to-Mel (S2M) Module ===&lt;br /&gt;
The S2M module employs a non-autoregressive architecture based on flow matching to convert semantic tokens into mel-spectrograms. Notable features include:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;GPT Latent Enhancement&#039;&#039;&#039;: Integration of latent features from the T2S module&#039;s final transformer layer to improve speech clarity, particularly during emotionally expressive synthesis&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Speaker Embedding Integration&#039;&#039;&#039;: Concatenation of speaker embeddings with semantic features to ensure timbre consistency&lt;br /&gt;
&lt;br /&gt;
=== Vocoder ===&lt;br /&gt;
The system utilizes [[BigVGANv2]] as its vocoder to convert mel-spectrograms into final audio waveforms, chosen for its superior audio quality and stability compared to previous vocoders.&lt;br /&gt;
&lt;br /&gt;
== Key Features and Capabilities ==&lt;br /&gt;
&lt;br /&gt;
=== Precise Duration Control ===&lt;br /&gt;
IndexTTS2 is claimed to be the first autoregressive zero-shot TTS model to achieve precise duration control. The model supports two generation modes:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specified Duration Mode&#039;&#039;&#039;: Users can explicitly specify the number of generated tokens to control speech duration with millisecond-level precision&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Natural Duration Mode&#039;&#039;&#039;: Free-form generation that faithfully reproduces prosodic features from input prompts without duration constraints&lt;br /&gt;
&lt;br /&gt;
This capability addresses critical requirements for applications like video dubbing, where precise synchronization between audio and visual content is essential.&lt;br /&gt;
&lt;br /&gt;
=== Fine-Grained Emotional Control ===&lt;br /&gt;
The system offers multiple methods for emotional control:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference Audio Emotion&#039;&#039;&#039;: Extraction of emotional characteristics from style prompt audio&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Natural Language Descriptions&#039;&#039;&#039;: Text-based emotion control using a specialized Text-to-Emotion (T2E) module&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Emotion Vector Input&#039;&#039;&#039;: Direct specification of emotional states through numerical vectors&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Cross-Speaker Emotion Transfer&#039;&#039;&#039;: Ability to apply emotional characteristics from one speaker to the voice of another&lt;br /&gt;
&lt;br /&gt;
=== Text-to-Emotion (T2E) Module ===&lt;br /&gt;
A specialized component that enables natural language-based emotion control through:&lt;br /&gt;
&lt;br /&gt;
* Knowledge distillation from DeepSeek-R1 to Qwen3-1.7B&lt;br /&gt;
* Support for seven basic emotions: Anger, Happiness, Fear, Disgust, Sadness, Surprise, and Neutral&lt;br /&gt;
* Generation of emotion probability distributions that are combined with precomputed emotion embeddings&lt;br /&gt;
&lt;br /&gt;
== Training and Dataset ==&lt;br /&gt;
IndexTTS2 was trained on a substantial multilingual corpus:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Total Training Data&#039;&#039;&#039;: 55,000 hours comprising 30,000 hours of Chinese data and 25,000 hours of English data&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Emotional Data&#039;&#039;&#039;: 135 hours of specialized emotional speech from 361 speakers&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Training Infrastructure&#039;&#039;&#039;: 8 NVIDIA A100 80GB GPUs using AdamW optimizer with 2e-4 learning rate&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Training Duration&#039;&#039;&#039;: Three weeks total training time&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Data Sources&#039;&#039;&#039;: Primarily from the Emilia dataset, supplemented with audiobooks and commercial data&lt;br /&gt;
&lt;br /&gt;
The three-stage training methodology includes:&lt;br /&gt;
&lt;br /&gt;
# Foundation training on the full dataset with duration control capabilities&lt;br /&gt;
# Emotional control refinement using curated emotional data with GRL-based disentanglement&lt;br /&gt;
# Robustness improvement through fine-tuning on the complete dataset&lt;br /&gt;
&lt;br /&gt;
== Performance Evaluation ==&lt;br /&gt;
&lt;br /&gt;
=== Objective Metrics ===&lt;br /&gt;
Based on evaluation across multiple datasets (LibriSpeech-test-clean, SeedTTS test-zh/en, AIShell-1), IndexTTS2 demonstrates:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Superior Word Error Rates&#039;&#039;&#039;: Outperforms baseline models including MaskGCT, F5-TTS, CosyVoice2, and SparkTTS across most test sets&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Strong Speaker Similarity&#039;&#039;&#039;: Achieves competitive speaker similarity scores while maintaining improved speech clarity&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Emotional Fidelity&#039;&#039;&#039;: Highest emotion similarity scores among evaluated models&lt;br /&gt;
&lt;br /&gt;
=== Subjective Evaluation ===&lt;br /&gt;
Human evaluation using Mean Opinion Scores (MOS) across multiple dimensions shows:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Quality MOS&#039;&#039;&#039;: Consistent superiority in perceived audio quality&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Similarity MOS&#039;&#039;&#039;: Strong performance in perceived speaker similarity&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Prosody MOS&#039;&#039;&#039;: Enhanced prosodic naturalness compared to baseline models&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Emotion MOS&#039;&#039;&#039;: Significant improvements in emotional expressiveness&lt;br /&gt;
&lt;br /&gt;
=== Duration Control Accuracy ===&lt;br /&gt;
Precision testing reveals minimal token number error rates:&lt;br /&gt;
&lt;br /&gt;
Original duration: &amp;lt;0.02% error rate&lt;br /&gt;
&lt;br /&gt;
Scaled durations (0.875×-1.125×): &amp;lt;0.03% error rate&lt;br /&gt;
&lt;br /&gt;
Larger scaling factors: Maximum 0.067% error rate&lt;br /&gt;
&lt;br /&gt;
== Comparison with Existing Systems ==&lt;br /&gt;
IndexTTS2 distinguishes itself from contemporary TTS systems through:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Versus ElevenLabs&#039;&#039;&#039;: Open-source nature and precise duration control capabilities&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Versus Traditional TTS&#039;&#039;&#039;: Enhanced emotional expressiveness and zero-shot voice cloning&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Versus Other Open-Source Systems&#039;&#039;&#039;: First autoregressive model with precise duration control&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Versus Non-Autoregressive Models&#039;&#039;&#039;: Maintains naturalness advantages while adding duration precision&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
[https://github.com/index-tts/index-tts Official IndexTTS Repository]&lt;br /&gt;
&lt;br /&gt;
[https://index-tts.github.io/index-tts2.github.io/ IndexTTS2 Demo Page]&lt;br /&gt;
&lt;br /&gt;
[https://huggingface.co/IndexTeam IndexTTS2 Models on Hugging Face]&lt;br /&gt;
&lt;br /&gt;
[https://arxiv.org/abs/2506.21619 IndexTTS2 Research Paper]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=46</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=46"/>
		<updated>2025-09-21T20:36:07Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: add some new models&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
|[[IndexTTS2]]&lt;br /&gt;
|Custom (restrictive)&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ✅ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Chatterbox]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[MegaTTS 3]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English, Chinese&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Orpheus TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CSM-1B]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CosyVoice 2.0]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|Chinese, English, Japanese, Korean&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2024&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=ElevenLabs&amp;diff=45</id>
		<title>ElevenLabs</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=ElevenLabs&amp;diff=45"/>
		<updated>2025-09-20T20:48:57Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add ElevenLabs page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox TTS model&lt;br /&gt;
| name = ElevenLabs&lt;br /&gt;
| developer = ElevenLabs Inc.&lt;br /&gt;
| release_date = January 2023 (beta)&lt;br /&gt;
| latest_version = Eleven v3 (alpha)&lt;br /&gt;
| languages = 32+ languages&lt;br /&gt;
| voices = 1000+ voices&lt;br /&gt;
| voice_cloning = Yes (professional &amp;amp; instant)&lt;br /&gt;
| emotion_control = Yes (via audio tags)&lt;br /&gt;
| streaming = Yes&lt;br /&gt;
| latency = ~135ms (Flash models)&lt;br /&gt;
| open_source = No&lt;br /&gt;
| website = https://elevenlabs.io&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;ElevenLabs&#039;&#039;&#039; is a commercial artificial intelligence company specializing in text-to-speech synthesis and voice cloning technology. Founded in 2022 by Piotr Dąbkowski and Mateusz Staniszewski, the company has gained prominence for its AI-generated voices that can replicate human speech patterns, emotions, and intonation across multiple languages.&lt;br /&gt;
&lt;br /&gt;
== History and Founding ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs was co-founded in 2022 by Piotr Dąbkowski, a former Google machine learning engineer, and Mateusz Staniszewski, an ex-Palantir deployment strategist. Both founders, originally from Poland, reportedly drew inspiration from the poor quality of film dubbing they experienced while watching American movies in their home country.&amp;lt;ref&amp;gt;https://venturebeat.com/ai/now-hear-this-voice-cloning-ai-startup-elevenlabs-nabs-19m-from-a16z-and-other-heavy-hitters&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The founders first met as teenagers at Copernicus High School in Warsaw before pursuing separate academic paths—Dąbkowski studying at Oxford and Cambridge, while Staniszewski studied mathematics in London. Their shared vision of making quality content accessible across all languages led to the creation of ElevenLabs as a research-first company.&amp;lt;ref&amp;gt;https://research.contrary.com/company/elevenlabs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The company launched its beta platform in January 2023, quickly gaining traction with over one million users within five months. This rapid adoption demonstrated market demand for high-quality AI voice synthesis technology.&amp;lt;ref&amp;gt;https://research.contrary.com/company/elevenlabs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Funding and Valuation ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs has experienced rapid growth in both user adoption and valuation:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Pre-seed (January 2023)&#039;&#039;&#039;: $2 million led by Credo Ventures and Concept Ventures&lt;br /&gt;
* &#039;&#039;&#039;Series A (June 2023)&#039;&#039;&#039;: $19 million at $100 million valuation, co-led by Andreessen Horowitz, Nat Friedman, and Daniel Gross&lt;br /&gt;
* &#039;&#039;&#039;Series B (January 2024)&#039;&#039;&#039;: $80 million at $1.1 billion valuation, achieving unicorn status&lt;br /&gt;
* &#039;&#039;&#039;Series C (January 2025)&#039;&#039;&#039;: $180 million at $3.3 billion valuation, led by Andreessen Horowitz and ICONIQ Growth&amp;lt;ref&amp;gt;https://en.wikipedia.org/wiki/ElevenLabs&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The company reportedly achieved $200 million in annual recurring revenue (ARR) by August 2025, demonstrating significant commercial traction.&amp;lt;ref&amp;gt;https://sacra.com/c/elevenlabs/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technology and Products ==&lt;br /&gt;
&lt;br /&gt;
=== Core Technology ===&lt;br /&gt;
&lt;br /&gt;
ElevenLabs&#039;s architecture is proprietary and remains undisclosed, with little information about it being publicly available. Some have speculated that early versions of ElevenLabs were based off of Tortoise TTS; however, these rumors remain unverified.&amp;lt;ref&amp;gt;https://github.com/neonbjb/tortoise-tts/discussions/277&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Product Portfolio ===&lt;br /&gt;
&lt;br /&gt;
==== Text-to-Speech Models ====&lt;br /&gt;
&lt;br /&gt;
ElevenLabs offers several model variants optimized for different use cases:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Multilingual v2&#039;&#039;&#039;: High-quality model supporting 29+ languages, optimized for audiobooks and professional content&lt;br /&gt;
* &#039;&#039;&#039;Flash v2.5&#039;&#039;&#039;: Ultra-low latency model (75ms) designed for real-time conversational applications&lt;br /&gt;
* &#039;&#039;&#039;Turbo v2.5&#039;&#039;&#039;: Balanced quality and speed model for general-purpose applications&lt;br /&gt;
* &#039;&#039;&#039;Eleven v3 (alpha)&#039;&#039;&#039;: Latest model featuring advanced emotion control via audio tags&lt;br /&gt;
* &#039;&#039;&#039;Eleven Scribe v1&#039;&#039;&#039;: SoTA automatic speech recognition model&lt;br /&gt;
* &#039;&#039;&#039;Eleven Music v1&#039;&#039;&#039;: Text-to-music model trained on licensed data&amp;lt;ref&amp;gt;https://elevenlabs.io/docs/models&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://elevenlabs.io/music&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==== Voice Cloning ====&lt;br /&gt;
&lt;br /&gt;
The platform provides two voice cloning approaches:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Instant Voice Cloning&#039;&#039;&#039;: Creates voice replicas from short audio samples (1-5 minutes)&lt;br /&gt;
* &#039;&#039;&#039;Professional Voice Cloning&#039;&#039;&#039;: Higher-fidelity, fine-tuning-based cloning requiring longer training samples&lt;br /&gt;
&lt;br /&gt;
==== Additional Features ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;AI Dubbing&#039;&#039;&#039;: Translates and dubs content while preserving original voice characteristics and emotions&lt;br /&gt;
* &#039;&#039;&#039;Voice Design&#039;&#039;&#039;: Tool for creating entirely synthetic voices from text descriptions&lt;br /&gt;
* &#039;&#039;&#039;Speech Classifier&#039;&#039;&#039;: Detection tool to identify AI-generated audio from ElevenLabs&#039; technology&lt;br /&gt;
* &#039;&#039;&#039;Projects&#039;&#039;&#039;: Long-form content creation tool for audiobooks and extended narration&lt;br /&gt;
&lt;br /&gt;
== Business Model and Pricing ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs operates on a freemium subscription model with usage-based pricing:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Free Tier&#039;&#039;&#039;: 10,000 characters per month with basic voices&lt;br /&gt;
* &#039;&#039;&#039;Starter&#039;&#039;&#039;: $5/month with commercial licensing&lt;br /&gt;
* &#039;&#039;&#039;Creator&#039;&#039;&#039;: $11/month with enhanced features&lt;br /&gt;
* &#039;&#039;&#039;Pro&#039;&#039;&#039;: $99/month for professional use&lt;br /&gt;
* &#039;&#039;&#039;Enterprise&#039;&#039;&#039;: Custom pricing with SLAs and dedicated support&lt;br /&gt;
&lt;br /&gt;
The company has evolved its pricing structure multiple times, transitioning from simple character-based billing to more complex model-aware systems and back to unified credit systems as it scaled.&amp;lt;ref&amp;gt;https://flexprice.io/blog/elevenlabs-pricing-breakdown&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance and Benchmarks ==&lt;br /&gt;
&lt;br /&gt;
Independent evaluations have provided mixed results regarding ElevenLabs&#039; performance relative to competitors:&lt;br /&gt;
&lt;br /&gt;
=== Competitive Analysis ===&lt;br /&gt;
&lt;br /&gt;
According to third-party benchmarks:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Voice Quality&#039;&#039;&#039;: ElevenLabs demonstrates superior Mean Opinion Scores (MOS) compared to Google Cloud Text-to-Speech across fiction, non-fiction, and conversational content{https://unrealspeech.com/compare/elevenlabs-vs-google-text-to-speech}&lt;br /&gt;
* &#039;&#039;&#039;Latency&#039;&#039;&#039;: Flash models achieve approximately 135ms Time to First Audio (TTFA), competitive with major cloud providers{https://cartesia.ai/vs/elevenlabs-vs-microsoft-azure-text-to-speech}&lt;br /&gt;
* &#039;&#039;&#039;Accuracy&#039;&#039;&#039;: Word Error Rates vary but generally maintain competitive performance with established providers&lt;br /&gt;
&lt;br /&gt;
However, these evaluations should be interpreted cautiously as they often come from companies with commercial interests in the TTS space, and standardized, independent benchmarking in the industry remains limited.&lt;br /&gt;
&lt;br /&gt;
== Controversies and Ethical Concerns ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs has faced significant criticism regarding the misuse of its technology:&lt;br /&gt;
&lt;br /&gt;
=== Early Misuse Incidents ===&lt;br /&gt;
&lt;br /&gt;
Shortly after the beta launch in January 2023, the platform was exploited by users on 4chan and other forums to create fake audio content. Notable incidents included:&lt;br /&gt;
&lt;br /&gt;
* Creation of celebrity deepfakes, including voices of Emma Watson, Alexandria Ocasio-Cortez, and Ben Shapiro making statements they never made&lt;br /&gt;
* Generation of racist, sexist, and homophobic content using cloned voices&amp;lt;ref&amp;gt;https://www.vice.com/en/article/ai-voice-firm-4chan-celebrity-voices-emma-watson-joe-rogan-elevenlabs/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Political Deepfakes ===&lt;br /&gt;
&lt;br /&gt;
In January 2024, ElevenLabs&#039; technology was used to create a robocall impersonating President Joe Biden, urging New Hampshire voters not to participate in the Democratic primary. The incident prompted investigation by the New Hampshire Attorney General&#039;s office and led to the suspension of the responsible user account.&amp;lt;ref&amp;gt;https://www.bloomberg.com/news/articles/2024-01-26/ai-startup-elevenlabs-bans-account-blamed-for-biden-audio-deepfake&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Legal Challenges ===&lt;br /&gt;
&lt;br /&gt;
The company faces ongoing legal challenges, including:&lt;br /&gt;
&lt;br /&gt;
* A lawsuit from voice actors Mark Boyett and Karissa Vacker, alleging unauthorized use of their voices to create the &amp;quot;Adam&amp;quot; and &amp;quot;Bella&amp;quot; default voices&lt;br /&gt;
* Claims of copyright infringement related to the use of audiobook recordings for training&amp;lt;ref&amp;gt;https://www.thevoicerealm.com/blog/a-look-into-the-elevenlabs-lawsuit/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Safety Measures ===&lt;br /&gt;
&lt;br /&gt;
In response to misuse concerns, ElevenLabs has implemented several safeguards:&lt;br /&gt;
&lt;br /&gt;
* Verification requirements for voice cloning features&lt;br /&gt;
* AI Speech Classifier for detecting ElevenLabs-generated content&lt;br /&gt;
* Partnership with Reality Defender for deepfake detection&lt;br /&gt;
* Mandatory credit card information for certain features&amp;lt;ref&amp;gt;https://www.bloomberg.com/news/articles/2024-07-18/elevenlabs-partners-with-reality-defender-to-combat-deepfake-audio&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Applications and Use Cases ==&lt;br /&gt;
&lt;br /&gt;
ElevenLabs technology is utilized across various industries:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Media and Entertainment&#039;&#039;&#039;: Audiobook production, podcast creation, film dubbing&lt;br /&gt;
* &#039;&#039;&#039;Gaming&#039;&#039;&#039;: Character voice generation for video games&lt;br /&gt;
* &#039;&#039;&#039;Education&#039;&#039;&#039;: Educational content narration and language learning&lt;br /&gt;
* &#039;&#039;&#039;Enterprise&#039;&#039;&#039;: Customer service automation, training materials&lt;br /&gt;
* &#039;&#039;&#039;Accessibility&#039;&#039;&#039;: Tools for visually impaired users&lt;br /&gt;
&lt;br /&gt;
The company reports that 41% of Fortune 500 companies use its platform, with notable customers including The Washington Post, TIME magazine, and HarperCollins Publishers.&amp;lt;ref&amp;gt;https://sacra.com/c/elevenlabs/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://elevenlabs.io/ Official ElevenLabs website]&lt;br /&gt;
* [https://elevenlabs.io/docs/ ElevenLabs Documentation]&lt;br /&gt;
* [https://elevenlabs.io/text-to-speech Text-to-Speech Demo]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=MIT_License&amp;diff=44</id>
		<title>MIT License</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=MIT_License&amp;diff=44"/>
		<updated>2025-09-20T20:10:13Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;&amp;#039;&amp;#039;&amp;#039;MIT License&amp;#039;&amp;#039;&amp;#039; is a permissive free software license originally developed at the Massachusetts Institute of Technology (MIT). It is one of the most popular open-source licenses used in software development and is commonly used for licensing both code and model weights.  == Overview ==  The MIT License is characterized by its simplicity and permissive nature. It allows users to do almost anything with the licensed software, including using, copying, modifying, merg...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;MIT License&#039;&#039;&#039; is a permissive free software license originally developed at the Massachusetts Institute of Technology (MIT). It is one of the most popular open-source licenses used in software development and is commonly used for licensing both code and [[model weights]].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
&lt;br /&gt;
The MIT License is characterized by its simplicity and permissive nature. It allows users to do almost anything with the licensed software, including using, copying, modifying, merging, publishing, distributing, sublicensing, and selling copies, with minimal restrictions.&lt;br /&gt;
&lt;br /&gt;
The license requires only that the original copyright notice and license text be included in all copies or substantial portions of the software. Unlike more restrictive licenses such as the [[GNU General Public License]] (GPL), the MIT License does not require derivative works to be released under the same license terms.&lt;br /&gt;
&lt;br /&gt;
== Key Permissions ==&lt;br /&gt;
&lt;br /&gt;
Under the MIT License, users are granted the following rights:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Commercial use&#039;&#039;&#039;: The software can be used for commercial purposes&lt;br /&gt;
* &#039;&#039;&#039;Modification&#039;&#039;&#039;: The software can be modified and adapted&lt;br /&gt;
* &#039;&#039;&#039;Distribution&#039;&#039;&#039;: The software can be distributed to others&lt;br /&gt;
* &#039;&#039;&#039;Private use&#039;&#039;&#039;: The software can be used privately without disclosure&lt;br /&gt;
* &#039;&#039;&#039;Sublicensing&#039;&#039;&#039;: The software can be incorporated into projects with different licenses&lt;br /&gt;
&lt;br /&gt;
== Requirements ==&lt;br /&gt;
&lt;br /&gt;
The MIT License has minimal requirements:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;License and copyright notice&#039;&#039;&#039;: The original license text and copyright notice must be included with any distribution of the software&lt;br /&gt;
* &#039;&#039;&#039;No trademark use&#039;&#039;&#039;: The license does not grant permission to use the licensor&#039;s trademarks or trade names&lt;br /&gt;
&lt;br /&gt;
== Limitations ==&lt;br /&gt;
&lt;br /&gt;
The MIT License provides no warranty and limits liability:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;No warranty&#039;&#039;&#039;: The software is provided &amp;quot;as is&amp;quot; without any express or implied warranties&lt;br /&gt;
* &#039;&#039;&#039;Limited liability&#039;&#039;&#039;: The license disclaims liability for damages arising from the use of the software&lt;br /&gt;
* &#039;&#039;&#039;No patent protection&#039;&#039;&#039;: Unlike some other licenses, the MIT License does not explicitly grant patent rights&lt;br /&gt;
&lt;br /&gt;
== Comparison with Other Licenses ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! License !! Commercial Use !! Modification !! Distribution !! Copyleft !! Patent Grant&lt;br /&gt;
|-&lt;br /&gt;
| MIT || Yes || Yes || Yes || No || No&lt;br /&gt;
|-&lt;br /&gt;
| [[Apache 2.0]]|| Yes || Yes || Yes || No || Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[GPL v3]]|| Yes || Yes || Yes || Yes || Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[BSD-3-Clause]]|| Yes || Yes || Yes || No || No&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== License Text ==&lt;br /&gt;
&lt;br /&gt;
The complete MIT License text is relatively short:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
MIT License&lt;br /&gt;
&lt;br /&gt;
Copyright (c) [year] [fullname]&lt;br /&gt;
&lt;br /&gt;
Permission is hereby granted, free of charge, to any person obtaining a copy&lt;br /&gt;
of this software and associated documentation files (the &amp;quot;Software&amp;quot;), to deal&lt;br /&gt;
in the Software without restriction, including without limitation the rights&lt;br /&gt;
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell&lt;br /&gt;
copies of the Software, and to permit persons to whom the Software is&lt;br /&gt;
furnished to do so, subject to the following conditions:&lt;br /&gt;
&lt;br /&gt;
The above copyright notice and this permission notice shall be included in all&lt;br /&gt;
copies or substantial portions of the Software.&lt;br /&gt;
&lt;br /&gt;
THE SOFTWARE IS PROVIDED &amp;quot;AS IS&amp;quot;, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR&lt;br /&gt;
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,&lt;br /&gt;
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE&lt;br /&gt;
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER&lt;br /&gt;
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,&lt;br /&gt;
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE&lt;br /&gt;
SOFTWARE.&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* [https://opensource.org/licenses/MIT Open Source Initiative - MIT License]&lt;br /&gt;
* [https://choosealicense.com/licenses/mit/ Choose a License - MIT]&lt;br /&gt;
&lt;br /&gt;
[[Category:Licenses]]&lt;br /&gt;
[[Category:Open Source]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=43</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=43"/>
		<updated>2025-09-20T16:28:18Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add year of release&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ✅ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Chatterbox]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Orpheus TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CSM-1B]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CosyVoice 2.0]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|Chinese, English, Japanese, Korean&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2024&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Chatterbox&amp;diff=42</id>
		<title>Chatterbox</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Chatterbox&amp;diff=42"/>
		<updated>2025-09-20T16:27:59Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add Chatterbox&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox TTS model&lt;br /&gt;
| name = Chatterbox&lt;br /&gt;
| developer = [[Resemble AI]]&lt;br /&gt;
| release_date = May 2025&lt;br /&gt;
| latest_version = Multilingual 2.0&lt;br /&gt;
| architecture = [[CosyVoice 2.0]]-based&lt;br /&gt;
| parameters = 500 million&lt;br /&gt;
| training_data = 500,000 hours cleaned data&lt;br /&gt;
| languages = 23 languages (multilingual version)&lt;br /&gt;
| voices = Zero-shot voice cloning&lt;br /&gt;
| voice_cloning = Yes (5-second reference)&lt;br /&gt;
| emotion_control = Yes (exaggeration parameter)&lt;br /&gt;
| streaming = Yes&lt;br /&gt;
| latency = Sub-200ms&lt;br /&gt;
| license = [[MIT License]]&lt;br /&gt;
| open_source = Yes&lt;br /&gt;
| code_repository = [https://github.com/resemble-ai/chatterbox GitHub]&lt;br /&gt;
| model_weights = [https://huggingface.co/ResembleAI/chatterbox Hugging Face]&lt;br /&gt;
| demo = [https://huggingface.co/spaces/ResembleAI/Chatterbox HF Spaces]&lt;br /&gt;
| website = [https://www.resemble.ai/chatterbox/ resemble.ai/chatterbox]&lt;br /&gt;
}}&lt;br /&gt;
&#039;&#039;&#039;Chatterbox&#039;&#039;&#039; is an open-source [[text-to-speech]] (TTS) model developed by [[Resemble AI]] and released in May 2025. Built on a modified [[Llama]] architecture with 500M parameters, it is marketed as the first open-source TTS model to include controllable emotion exaggeration and has gained attention for claiming to outperform established commercial systems in user preference evaluations. It is built on the [[CosyVoice|CosyVoice 2.0]] architecture.&lt;br /&gt;
&lt;br /&gt;
== Development and Release ==&lt;br /&gt;
&lt;br /&gt;
Chatterbox was developed by a three-person team at Resemble AI, a voice technology company founded by Zohaib Ahmed and Saqib Muhammad.&amp;lt;ref&amp;gt;https://www.digitalocean.com/community/tutorials/resemble-chatterbox-tts-text-to-speech&amp;lt;/ref&amp;gt; The initial English-only version was released in May 2025 under the [[MIT License]], followed by a multilingual version supporting 23 languages in September 2025.&amp;lt;ref&amp;gt;https://www.resemble.ai/introducing-chatterbox-multilingual-open-source-tts-for-23-languages/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The project quickly gained popularity in the open-source community, accumulating over 1 million downloads on [[Hugging Face]] and more than 11,000 stars on [[GitHub]] within weeks of release.&amp;lt;ref name=&amp;quot;multilingual&amp;quot;&amp;gt;https://www.resemble.ai/introducing-chatterbox-multilingual-open-source-tts-for-23-languages/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
Chatterbox utilizes a 500-million parameter model based on a CosyVoice-style modified Llama architecture, significantly smaller than many contemporary TTS systems. The model was trained on approximately 500,000 hours of cleaned audio data and employs what the developers term &amp;quot;alignment-informed inference&amp;quot; for improved stability during generation.&lt;br /&gt;
&lt;br /&gt;
Key technical features include:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Zero-shot voice cloning&#039;&#039;&#039;: Ability to clone voices using as little as 5 seconds of reference audio&lt;br /&gt;
* &#039;&#039;&#039;Emotion exaggeration control&#039;&#039;&#039;: A novel parameter allowing users to adjust emotional intensity from monotone to dramatically expressive&lt;br /&gt;
* &#039;&#039;&#039;Fast inference&#039;&#039;&#039;: Sub-200ms latency for real-time applications&lt;br /&gt;
* &#039;&#039;&#039;Multilingual support&#039;&#039;&#039;: The updated version supports 23 languages including Arabic, Chinese, Hindi, and major European languages&lt;br /&gt;
&lt;br /&gt;
== Performance Claims and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
Resemble AI conducted a comparative evaluation through [[Podonos]], a third-party evaluation service, testing Chatterbox against [[ElevenLabs]], a leading commercial TTS system. In blind A/B testing, 63.75% of evaluators reportedly preferred Chatterbox&#039;s output over ElevenLabs.&amp;lt;ref&amp;gt;https://www.podonos.com/blog/chatterbox&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://www.resemble.ai/chatterbox/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, these results should be interpreted with caution, as the evaluation was limited in scope and conducted by a single third-party service. The testing methodology, sample size, and demographic composition of evaluators have not been independently verified. Additionally, the comparison was limited to a single competitor rather than a comprehensive benchmark against multiple state-of-the-art systems.&lt;br /&gt;
&lt;br /&gt;
== Commercial and Research Impact ==&lt;br /&gt;
&lt;br /&gt;
The release of Chatterbox has been significant for the open-source TTS community, representing one of the first production-grade systems to be freely available under a permissive license. This has enabled developers to integrate high-quality TTS capabilities into applications without licensing costs or vendor dependencies.&lt;br /&gt;
&lt;br /&gt;
The system has found applications in various domains including:&lt;br /&gt;
&lt;br /&gt;
* Audiobook generation and voice narration&lt;br /&gt;
* Game development for non-player character dialogue&lt;br /&gt;
* Educational content creation&lt;br /&gt;
* Accessibility tools for visually impaired users&lt;br /&gt;
* Research and development in speech synthesis&lt;br /&gt;
&lt;br /&gt;
Resemble AI also offers a commercial &amp;quot;Pro&amp;quot; version with enhanced features, service-level agreements, and custom fine-tuning capabilities for enterprise customers requiring guaranteed performance and support. This version is available through their inference partners, such as FAL.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/resemble-ai/chatterbox Official Chatterbox repository]&lt;br /&gt;
* [https://huggingface.co/ResembleAI/chatterbox Model on Hugging Face]&lt;br /&gt;
* [https://huggingface.co/spaces/ResembleAI/Chatterbox Interactive demo]&lt;br /&gt;
* [https://resemble-ai.github.io/chatterbox_demopage/ Demo page with audio samples]&lt;br /&gt;
&lt;br /&gt;
[[Category:Speech synthesis]]&lt;br /&gt;
[[Category:Open-source software]]&lt;br /&gt;
[[Category:Artificial intelligence]]&lt;br /&gt;
[[Category:Voice technology]]&lt;br /&gt;
[[Category:MIT License software]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=41</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=41"/>
		<updated>2025-09-20T16:26:52Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: /* Open Source Models */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ✅ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Chatterbox]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Orpheus TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CSM-1B]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|✅&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|[[CosyVoice 2.0]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|Chinese, English, Japanese, Korean&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2024&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=40</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=40"/>
		<updated>2025-09-20T16:22:49Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: cosyvoice&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Chatterbox]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Orpheus TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[CosyVoice 2.0]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|Chinese, English, Japanese, Korean&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2024&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=VibeVoice&amp;diff=39</id>
		<title>VibeVoice</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=VibeVoice&amp;diff=39"/>
		<updated>2025-09-20T16:18:30Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add VibeVoice infobox&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox TTS model&lt;br /&gt;
| name = VibeVoice&lt;br /&gt;
| developer = [[Microsoft Research]]&lt;br /&gt;
| release_date = August 26, 2025&lt;br /&gt;
| latest_version = 7B&lt;br /&gt;
| architecture = [[Qwen]] 2.5 + Diffusion&lt;br /&gt;
| parameters = 1.5B / 7B&lt;br /&gt;
| training_data = Proprietary dataset&lt;br /&gt;
| languages = English, Chinese&lt;br /&gt;
| voices = 4 speakers maximum&lt;br /&gt;
| voice_cloning = Yes&lt;br /&gt;
| emotion_control = Limited&lt;br /&gt;
| streaming = Yes&lt;br /&gt;
| license = MIT&lt;br /&gt;
| open_source = Limited (code removed)&lt;br /&gt;
| code_repository = [https://github.com/vibevoice-community/VibeVoice Community fork]&lt;br /&gt;
| model_weights = [https://huggingface.co/vibevoice Community backup]&lt;br /&gt;
| website = [https://aka.ms/VibeVoice Microsoft]&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;VibeVoice&#039;&#039;&#039; is an experimental [[text-to-speech]] (TTS) framework developed by [[Microsoft Research]] for generating long-form, multi-speaker conversational audio. It was released in August 2025 and is designed to synthesize long-form speech content such as podcasts and audiobooks with up to 4 speakers and with support for voice cloning.&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Development and Release ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice was developed by a team at Microsoft Research led by Zhiliang Peng, Jianwei Yu, and others, with the technical report published on [[arXiv]] in August 2025.&amp;lt;ref&amp;gt;https://arxiv.org/abs/2508.19205&amp;lt;/ref&amp;gt; The project was initially released as open-source software on GitHub and [[Hugging Face]], with model weights made publicly available under the MIT license.&lt;br /&gt;
&lt;br /&gt;
However, the release was disrupted in September 2025 when Microsoft removed the official repository and model weights from public access. According to Microsoft&#039;s statement, this action was taken after discovering &amp;quot;instances where the tool was used in ways inconsistent with the stated intent&amp;quot; and concerns about responsible AI use.&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt; The repository was later restored without implementation code, while community-maintained forks preserved the original materials. The 1.5B pretrained model remains available on Microsoft&#039;s Hugging Face page, but the 7B model was taken down.&lt;br /&gt;
&lt;br /&gt;
Community forks have preserved backups of the code and model weights, including both the 1.5B and 7B models.&amp;lt;ref&amp;gt;https://github.com/vibevoice-community/VibeVoice&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://huggingface.co/aoi-ot/VibeVoice-Large&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice uses a hybrid architecture combining large language models with diffusion-based audio generation. The system uses two specialized tokenizers operating at an ultra-low 7.5 Hz frame rate:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Acoustic Tokenizer&#039;&#039;&#039;: A [[variational autoencoder]] (VAE) based encoder-decoder that compresses audio signals while preserving fidelity&lt;br /&gt;
* &#039;&#039;&#039;Semantic Tokenizer&#039;&#039;&#039;: A content-focused encoder trained using [[automatic speech recognition]] as a proxy task&lt;br /&gt;
&lt;br /&gt;
The core model utilizes [[Qwen2.5]] as its base large language model (available in 1.5B and 7B parameter variants), integrated with a lightweight diffusion head for generating acoustic features. This design achieves what the researchers claim is an 80-fold improvement in data compression compared to the [[Encodec]] model while maintaining audio quality.&amp;lt;ref&amp;gt;https://arxiv.org/abs/2508.19205&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Capabilities and Limitations ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice can generate speech sequences up to 90 minutes in length with support for up to four distinct speakers. The model demonstrates several emergent capabilities not explicitly trained for, including:&lt;br /&gt;
&lt;br /&gt;
* Cross-lingual speech synthesis&lt;br /&gt;
* Spontaneous singing (though often off-key)&lt;br /&gt;
* Contextual background music generation&lt;br /&gt;
* Voice cloning from short prompts&lt;br /&gt;
&lt;br /&gt;
However, the system has notable limitations:&lt;br /&gt;
&lt;br /&gt;
* Language support restricted to English and Chinese&lt;br /&gt;
* No explicit modeling of overlapping speech&lt;br /&gt;
* Occasional instability, particularly with Chinese text synthesis&lt;br /&gt;
* Uncontrolled generation of background sounds and music&lt;br /&gt;
* Limited commercial viability due to various technical constraints&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
In comparative evaluations against contemporary TTS systems including [[ElevenLabs]], [[Google]]&#039;s Gemini 2.5 Pro TTS, and others, VibeVoice reportedly achieved superior scores in subjective metrics of realism, richness, and user preference. The model also demonstrated competitive [[word error rate]]s when evaluated using speech recognition systems.&amp;lt;ref&amp;gt;https://huggingface.co/microsoft/VibeVoice-1.5B&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, these evaluations were conducted on a limited test set of eight conversational transcripts totaling approximately one hour of audio, raising questions about the generalizability of the results to broader use cases.&lt;br /&gt;
&lt;br /&gt;
== Controversies and Concerns ==&lt;br /&gt;
&lt;br /&gt;
The temporary removal of VibeVoice from public access highlighted ongoing concerns about the potential misuse of high-quality synthetic speech technology. Microsoft explicitly warned about the potential for creating [[deepfake]] audio content for impersonation, fraud, or disinformation purposes.&lt;br /&gt;
&lt;br /&gt;
The model&#039;s ability to generate convincing speech from minimal voice prompts, combined with its long-form generation capabilities, raised particular concerns among AI safety researchers about potential misuse for creating fake audio content at scale.&lt;br /&gt;
&lt;br /&gt;
== Community Response ==&lt;br /&gt;
&lt;br /&gt;
Following Microsoft&#039;s temporary withdrawal of the official release, the open-source community created several preservation efforts:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;[https://github.com/vibevoice-community/VibeVoice vibevoice-community/VibeVoice]&#039;&#039;&#039;: A community-maintained fork preserving the original codebase and model weights&lt;br /&gt;
* &#039;&#039;&#039;[https://github.com/voicepowered-ai/VibeVoice-finetuning VibeVoice-finetuning]&#039;&#039;&#039;: Unofficial tools for fine-tuning the models using [[Low-Rank Adaptation]] (LoRA) techniques&lt;br /&gt;
&lt;br /&gt;
These community efforts have enabled continued research and development despite the official restrictions, though they operate independently of Microsoft&#039;s oversight.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/vibevoice-community/VibeVoice Community-maintained VibeVoice repository]&lt;br /&gt;
* [https://arxiv.org/abs/2508.19205 Original technical report on arXiv]&lt;br /&gt;
&lt;br /&gt;
[[Category:Artificial intelligence]]&lt;br /&gt;
[[Category:Speech synthesis]]&lt;br /&gt;
[[Category:Microsoft Research]]&lt;br /&gt;
[[Category:Open-source software]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Nobold&amp;diff=38</id>
		<title>Template:Nobold</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Nobold&amp;diff=38"/>
		<updated>2025-09-20T16:12:14Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: 1 revision imported: Add infobox template&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;templatestyles src=&amp;quot;Nobold/styles.css&amp;quot;/&amp;gt;&amp;lt;span class=&amp;quot;nobold&amp;quot;&amp;gt;{{{1}}}&amp;lt;/span&amp;gt;&amp;lt;noinclude&amp;gt;&lt;br /&gt;
{{documentation}}&lt;br /&gt;
&amp;lt;!-- PLEASE ADD CATEGORIES AND INTERWIKIS TO THE /doc SUBPAGE, THANKS --&amp;gt;&lt;br /&gt;
&amp;lt;/noinclude&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Main_other&amp;diff=36</id>
		<title>Template:Main other</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Main_other&amp;diff=36"/>
		<updated>2025-09-20T16:12:14Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: 1 revision imported: Add infobox template&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{safesubst:&amp;lt;noinclude/&amp;gt;#switch:&lt;br /&gt;
  &amp;lt;noinclude&amp;gt;&amp;lt;!-- If no or empty &amp;quot;demospace&amp;quot; parameter then detect namespace --&amp;gt;&amp;lt;/noinclude&amp;gt;&lt;br /&gt;
  {{safesubst:&amp;lt;noinclude/&amp;gt;#if:{{{demospace|}}} &lt;br /&gt;
  | {{safesubst:&amp;lt;noinclude/&amp;gt;lc: {{{demospace}}} }}    &amp;lt;noinclude&amp;gt;&amp;lt;!-- Use lower case &amp;quot;demospace&amp;quot; --&amp;gt;&amp;lt;/noinclude&amp;gt;&lt;br /&gt;
  | {{safesubst:&amp;lt;noinclude/&amp;gt;#ifeq:{{safesubst:&amp;lt;noinclude/&amp;gt;NAMESPACE}}|{{safesubst:&amp;lt;noinclude/&amp;gt;ns:0}}&lt;br /&gt;
    | main&lt;br /&gt;
    | other&lt;br /&gt;
    }} &lt;br /&gt;
  }}&lt;br /&gt;
| main     = {{{1|}}}&lt;br /&gt;
| other&lt;br /&gt;
| #default = {{{2|}}}&lt;br /&gt;
}}&amp;lt;noinclude&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{{documentation}}&lt;br /&gt;
&amp;lt;!-- Add categories to the /doc subpage; interwikis go to Wikidata, thank you! --&amp;gt;&lt;br /&gt;
&amp;lt;/noinclude&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Yesno&amp;diff=34</id>
		<title>Template:Yesno</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Yesno&amp;diff=34"/>
		<updated>2025-09-20T16:12:14Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: 1 revision imported: Add infobox template&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{&amp;lt;includeonly&amp;gt;safesubst:&amp;lt;/includeonly&amp;gt;#switch: {{&amp;lt;includeonly&amp;gt;safesubst:&amp;lt;/includeonly&amp;gt;lc: {{{1|¬}}} }}&lt;br /&gt;
 |no&lt;br /&gt;
 |n&lt;br /&gt;
 |f&lt;br /&gt;
 |false&lt;br /&gt;
 |off&lt;br /&gt;
 |0        = {{{no|&amp;lt;!-- null --&amp;gt;}}}&lt;br /&gt;
 |         = {{{blank|{{{no|&amp;lt;!-- null --&amp;gt;}}}}}}&lt;br /&gt;
 |¬        = {{{¬|}}}&lt;br /&gt;
 |yes&lt;br /&gt;
 |y&lt;br /&gt;
 |t&lt;br /&gt;
 |true&lt;br /&gt;
 |on&lt;br /&gt;
 |1        = {{{yes|yes}}}&lt;br /&gt;
 |#default = {{{def|{{{yes|yes}}}}}}&lt;br /&gt;
}}&amp;lt;noinclude&amp;gt;&lt;br /&gt;
{{Documentation}}&lt;br /&gt;
&amp;lt;/noinclude&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Template_link&amp;diff=32</id>
		<title>Template:Template link</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Template_link&amp;diff=32"/>
		<updated>2025-09-20T16:12:14Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: 1 revision imported: Add infobox template&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;span class=&amp;quot;nowrap&amp;quot;&amp;gt;&amp;amp;#123;&amp;amp;#123;&amp;lt;/span&amp;gt;[[Template:{{{1}}}|{{{1}}}]]&amp;lt;span class=&amp;quot;nowrap&amp;quot;&amp;gt;&amp;amp;#125;&amp;amp;#125;&amp;lt;/span&amp;gt;&amp;lt;noinclude&amp;gt;&lt;br /&gt;
{{documentation}}&lt;br /&gt;
&amp;lt;!-- Categories go on the /doc subpage and interwikis go on Wikidata. --&amp;gt;&lt;br /&gt;
&amp;lt;/noinclude&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Tl&amp;diff=30</id>
		<title>Template:Tl</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Tl&amp;diff=30"/>
		<updated>2025-09-20T16:12:14Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: 1 revision imported: Add infobox template&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Template:Template link]]&lt;br /&gt;
&lt;br /&gt;
{{Redirect category shell|&lt;br /&gt;
{{R from move}}&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Infobox&amp;diff=28</id>
		<title>Template:Infobox</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Infobox&amp;diff=28"/>
		<updated>2025-09-20T16:12:14Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: 1 revision imported: Add infobox template&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{#invoke:Infobox|infobox}}&amp;lt;includeonly&amp;gt;{{template other|{{#ifeq:{{PAGENAME}}|Infobox||{{#ifeq:{{str left|{{SUBPAGENAME}}|7}}|Infobox|[[Category:Infobox templates|{{remove first word|{{SUBPAGENAME}}}}]]}}}}|}}&amp;lt;/includeonly&amp;gt;&amp;lt;noinclude&amp;gt;&lt;br /&gt;
{{documentation}}&lt;br /&gt;
&amp;lt;!-- Categories go in the /doc subpage, and interwikis go in Wikidata. --&amp;gt;&lt;br /&gt;
&amp;lt;/noinclude&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=MediaWiki:Common.css&amp;diff=26</id>
		<title>MediaWiki:Common.css</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=MediaWiki:Common.css&amp;diff=26"/>
		<updated>2025-09-20T16:07:17Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;/* CSS placed here will be applied to all skins */&lt;br /&gt;
&lt;br /&gt;
.portal-column-left {&lt;br /&gt;
    float: left;&lt;br /&gt;
    width: 100%;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
.portal-column-right {&lt;br /&gt;
    float: right;&lt;br /&gt;
    width: 100%;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Two column layout on wider screens */&lt;br /&gt;
@media (min-width: 720px) {&lt;br /&gt;
    .portal-column-left {&lt;br /&gt;
        width: 48%;&lt;br /&gt;
        margin-right: 2%;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .portal-column-right {&lt;br /&gt;
        width: 48%;&lt;br /&gt;
        margin-left: 2%;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
/*infobox*/&lt;br /&gt;
&lt;br /&gt;
/* Base infobox styling */&lt;br /&gt;
.infobox {&lt;br /&gt;
    border: 1px solid #a2a9b1;&lt;br /&gt;
    border-spacing: 3px;&lt;br /&gt;
    background-color: #f8f9fa;&lt;br /&gt;
    color: black;&lt;br /&gt;
    margin: 0.5em 0 0.5em 1em;&lt;br /&gt;
    padding: 0.2em;&lt;br /&gt;
    float: right;&lt;br /&gt;
    clear: right;&lt;br /&gt;
    font-size: 88%;&lt;br /&gt;
    line-height: 1.5em;&lt;br /&gt;
    width: 22em;&lt;br /&gt;
    box-sizing: border-box;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
.infobox th {&lt;br /&gt;
    background-color: #eaecf0;&lt;br /&gt;
    text-align: center;&lt;br /&gt;
    font-weight: bold;&lt;br /&gt;
    padding: 4px 8px;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
.infobox td {&lt;br /&gt;
    padding: 4px 8px;&lt;br /&gt;
    word-wrap: break-word;&lt;br /&gt;
    overflow-wrap: break-word;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* TTS-specific styling */&lt;br /&gt;
.tts-infobox {&lt;br /&gt;
    max-width: 100%;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Mobile responsiveness */&lt;br /&gt;
@media screen and (max-width: 640px) {&lt;br /&gt;
    .infobox {&lt;br /&gt;
        float: none !important;&lt;br /&gt;
        margin: 1em 0 !important;&lt;br /&gt;
        width: 100% !important;&lt;br /&gt;
        max-width: none !important;&lt;br /&gt;
        box-sizing: border-box;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .tts-infobox {&lt;br /&gt;
        width: 100% !important;&lt;br /&gt;
        margin-left: 0 !important;&lt;br /&gt;
        margin-right: 0 !important;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .infobox th,&lt;br /&gt;
    .infobox td {&lt;br /&gt;
        font-size: 14px;&lt;br /&gt;
        padding: 6px 8px;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    /* Make links more touch-friendly on mobile */&lt;br /&gt;
    .infobox a {&lt;br /&gt;
        min-height: 44px;&lt;br /&gt;
        display: inline-block;&lt;br /&gt;
        line-height: 1.4;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Tablet responsiveness */&lt;br /&gt;
@media screen and (max-width: 900px) and (min-width: 641px) {&lt;br /&gt;
    .infobox {&lt;br /&gt;
        width: 18em;&lt;br /&gt;
        font-size: 85%;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Very small screens */&lt;br /&gt;
@media screen and (max-width: 480px) {&lt;br /&gt;
    .infobox {&lt;br /&gt;
        font-size: 90% !important;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .infobox th {&lt;br /&gt;
        font-size: 110% !important;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    /* Stack label and value on very small screens */&lt;br /&gt;
    .infobox tr td:first-child {&lt;br /&gt;
        display: block;&lt;br /&gt;
        width: 100%;&lt;br /&gt;
        font-weight: bold;&lt;br /&gt;
        border-bottom: none;&lt;br /&gt;
        padding-bottom: 2px;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .infobox tr td:last-child {&lt;br /&gt;
        display: block;&lt;br /&gt;
        width: 100%;&lt;br /&gt;
        padding-top: 2px;&lt;br /&gt;
        padding-left: 12px;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* High DPI displays */&lt;br /&gt;
@media screen and (-webkit-min-device-pixel-ratio: 2), &lt;br /&gt;
       screen and (min-resolution: 192dpi) {&lt;br /&gt;
    .infobox {&lt;br /&gt;
        border-width: 0.5px;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Print styles */&lt;br /&gt;
@media print {&lt;br /&gt;
    .infobox {&lt;br /&gt;
        float: none !important;&lt;br /&gt;
        width: 100% !important;&lt;br /&gt;
        margin: 1em 0 !important;&lt;br /&gt;
        page-break-inside: avoid;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Dark mode support (optional) */&lt;br /&gt;
@media (prefers-color-scheme: dark) {&lt;br /&gt;
    .infobox {&lt;br /&gt;
        background-color: #2a2a2a;&lt;br /&gt;
        color: #e6e6e6;&lt;br /&gt;
        border-color: #555;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .infobox th {&lt;br /&gt;
        background-color: #404040;&lt;br /&gt;
        color: #e6e6e6;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Infobox_TTS_model&amp;diff=25</id>
		<title>Template:Infobox TTS model</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Infobox_TTS_model&amp;diff=25"/>
		<updated>2025-09-20T16:06:57Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;includeonly&amp;gt;&lt;br /&gt;
{| class=&amp;quot;infobox tts-infobox&amp;quot; style=&amp;quot;width: 22em; font-size: 90%; text-align: left; border: 1px solid #aaa; background-color: #f9f9f9; color: black; margin: 0.5em 0 0.5em 1em; padding: 0.2em; float: right; clear: right;&amp;quot;&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; font-size: 125%; font-weight: bold; background-color: #ccccff; color: black;&amp;quot; | {{{name|{{PAGENAME}}}}}&lt;br /&gt;
{{#if:{{{image|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; padding: 0.5em;&amp;quot; {{!}} {{{image}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{caption|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; font-style: italic; padding: 0.2em;&amp;quot; {{!}} {{{caption}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Model Information&lt;br /&gt;
{{#if:{{{developer|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; width: 30%; vertical-align: top;&amp;quot; {{!}} Developer:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{developer}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{release_date|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Release date:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{release_date}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{latest_version|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Latest version:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{latest_version}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{architecture|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Architecture:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{architecture}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{parameters|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Parameters:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{parameters}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{training_data|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Training data:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{training_data}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Capabilities&lt;br /&gt;
{{#if:{{{languages|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Languages:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{languages}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{voices|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Voices:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{voices}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{voice_cloning|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Voice cloning:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{voice_cloning}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{emotion_control|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Emotion control:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{emotion_control}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{streaming|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Streaming:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{streaming}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{latency|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Latency:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{latency}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Availability&lt;br /&gt;
{{#if:{{{license|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} License:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{license}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{open_source|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Open source:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{open_source}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{code_repository|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Repository:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{code_repository}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{model_weights|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Model weights:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{model_weights}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{demo|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Demo:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{demo}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{website|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; vertical-align: top;&amp;quot; {{!}} Website:&lt;br /&gt;
{{!}} style=&amp;quot;word-wrap: break-word;&amp;quot; {{!}} {{{website}}}&lt;br /&gt;
}}&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/includeonly&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Orpheus_TTS&amp;diff=24</id>
		<title>Orpheus TTS</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Orpheus_TTS&amp;diff=24"/>
		<updated>2025-09-20T16:06:28Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: add infobox&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Infobox TTS model&lt;br /&gt;
| name = Orpheus TTS&lt;br /&gt;
| developer = [[Canopy Labs]]&lt;br /&gt;
| release_date = March 18, 2025&lt;br /&gt;
| latest_version = 3B 0.1&lt;br /&gt;
| architecture = [[LLM-Based]]&lt;br /&gt;
| parameters = 3 billion&lt;br /&gt;
| training_data = 100k+ hours (English)&lt;br /&gt;
| languages = English (multilingual in preview)&lt;br /&gt;
| voices = 8 distinct voices&lt;br /&gt;
| voice_cloning = Yes (zero-shot)&lt;br /&gt;
| emotion_control = Yes (tag-based)&lt;br /&gt;
| streaming = Yes&lt;br /&gt;
| latency = ~200ms&lt;br /&gt;
| license = [[Apache License 2.0]]&lt;br /&gt;
| open_source = Yes&lt;br /&gt;
| code_repository = [https://github.com/canopyai/Orpheus-TTS GitHub]&lt;br /&gt;
| model_weights = [https://huggingface.co/canopylabs/orpheus-3b-0.1-ft Hugging Face]&lt;br /&gt;
| demo = [https://huggingface.co/spaces/MohamedRashad/Orpheus-TTS HF Spaces]&lt;br /&gt;
| website = [https://canopylabs.ai/releases/towards_human_sounding_tts Canopy Labs]&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Orpheus TTS&#039;&#039;&#039; is an open-source [[text-to-speech]] (TTS) system developed by Canopy Labs and released in March 2025. Built on the [[Llama (language model)|Llama-3.2-3B]] architecture, it uses a novel approach of using large-language-models with audio tokens instead of traditional TTS-specific architectures.&lt;br /&gt;
&lt;br /&gt;
== Development and Release ==&lt;br /&gt;
&lt;br /&gt;
Orpheus TTS was developed by Canopy Labs, an artificial intelligence startup founded with the stated mission of creating &amp;quot;digital humans that are indistinguishable from real humans.&amp;quot;&amp;lt;ref&amp;gt;https://canopylabs.ai/&amp;lt;/ref&amp;gt; The code and model weights were publicly released on March 18, 2025, under the [[Apache License]] 2.0, making both the model weights and training code freely available.&amp;lt;ref&amp;gt;https://github.com/canopyai/Orpheus-TTS&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
Orpheus TTS differs from conventional text-to-speech systems by using a modified Meta&#039;s Llama-3.2-3B language model as its foundation. It takes in a text prompt and generates audio tokens using the 24 kHz variant of the [[SNAC]] audio tokenizer. The system was trained on a dataset comprising over 100,000 hours of English speech data combined with billions of tokens of textual QA pairs, a hybrid approach designed to maintain linguistic understanding while adding speech synthesis capabilities.&lt;br /&gt;
&lt;br /&gt;
The model supports streaming out-of-the-box and can achieve ~200 milliseconds of streaming latency.&amp;lt;ref&amp;gt;https://www.baseten.co/blog/canopy-labs-selects-baseten-as-preferred-inference-provider-for-orpheus-tts-model/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance Claims and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
Canopy Labs claims that Orpheus TTS delivers &amp;quot;natural intonation, emotion, and rhythm that is superior to SOTA closed source models,&amp;quot; positioning it as competitive with established commercial systems such as [[ElevenLabs]] and other proprietary text-to-speech services.&amp;lt;ref&amp;gt;https://huggingface.co/canopylabs/orpheus-3b-0.1-ft&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, these performance assertions are based primarily on internal evaluations and subjective assessments rather than standardized benchmarks or peer-reviewed studies. The lack of comprehensive comparative analysis with established TTS systems has led to some skepticism within the research community about the extent of its claimed superiority.&lt;br /&gt;
&lt;br /&gt;
== Model Variants ==&lt;br /&gt;
&lt;br /&gt;
The model is available in three variants:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Pretrained model&#039;&#039;&#039;: The base model trained on the full dataset, suitable for research and custom fine-tuning&lt;br /&gt;
* &#039;&#039;&#039;Fine-tuned production model&#039;&#039;&#039;: An optimized version designed for everyday TTS applications, fine-tuned on several voices&lt;br /&gt;
* &#039;&#039;&#039;Multilingual model&#039;&#039;&#039;: A family of fine-tuned models with support for other languages&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/canopyai/Orpheus-TTS Official Orpheus TTS repository]&lt;br /&gt;
* [https://huggingface.co/canopylabs/orpheus-3b-0.1-ft Model on Hugging Face]&lt;br /&gt;
* [https://www.baseten.co/library/orpheus-tts/ Orpheus TTS on Baseten]&lt;br /&gt;
&lt;br /&gt;
[[Category:Speech synthesis]]&lt;br /&gt;
[[Category:Open-source software]]&lt;br /&gt;
[[Category:Artificial intelligence]]&lt;br /&gt;
[[Category:Large language models]]&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Orpheus_TTS&amp;diff=23</id>
		<title>Orpheus TTS</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Orpheus_TTS&amp;diff=23"/>
		<updated>2025-09-20T16:04:20Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Orpheus TTS&#039;&#039;&#039; is an open-source [[text-to-speech]] (TTS) system developed by Canopy Labs and released in March 2025. Built on the [[Llama (language model)|Llama-3.2-3B]] architecture, it uses a novel approach of using large-language-models with audio tokens instead of traditional TTS-specific architectures.&lt;br /&gt;
&lt;br /&gt;
== Development and Release ==&lt;br /&gt;
&lt;br /&gt;
Orpheus TTS was developed by Canopy Labs, an artificial intelligence startup founded with the stated mission of creating &amp;quot;digital humans that are indistinguishable from real humans.&amp;quot;&amp;lt;ref&amp;gt;https://canopylabs.ai/&amp;lt;/ref&amp;gt; The code and model weights were publicly released on March 18, 2025, under the [[Apache License]] 2.0, making both the model weights and training code freely available.&amp;lt;ref&amp;gt;https://github.com/canopyai/Orpheus-TTS&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
Orpheus TTS differs from conventional text-to-speech systems by using a modified Meta&#039;s Llama-3.2-3B language model as its foundation. It takes in a text prompt and generates audio tokens using the 24 kHz variant of the [[SNAC]] audio tokenizer. The system was trained on a dataset comprising over 100,000 hours of English speech data combined with billions of tokens of textual QA pairs, a hybrid approach designed to maintain linguistic understanding while adding speech synthesis capabilities.&lt;br /&gt;
&lt;br /&gt;
The model supports streaming out-of-the-box and can achieve ~200 milliseconds of streaming latency.&amp;lt;ref&amp;gt;https://www.baseten.co/blog/canopy-labs-selects-baseten-as-preferred-inference-provider-for-orpheus-tts-model/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance Claims and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
Canopy Labs claims that Orpheus TTS delivers &amp;quot;natural intonation, emotion, and rhythm that is superior to SOTA closed source models,&amp;quot; positioning it as competitive with established commercial systems such as [[ElevenLabs]] and other proprietary text-to-speech services.&amp;lt;ref&amp;gt;https://huggingface.co/canopylabs/orpheus-3b-0.1-ft&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, these performance assertions are based primarily on internal evaluations and subjective assessments rather than standardized benchmarks or peer-reviewed studies. The lack of comprehensive comparative analysis with established TTS systems has led to some skepticism within the research community about the extent of its claimed superiority.&lt;br /&gt;
&lt;br /&gt;
== Model Variants ==&lt;br /&gt;
&lt;br /&gt;
The model is available in three variants:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Pretrained model&#039;&#039;&#039;: The base model trained on the full dataset, suitable for research and custom fine-tuning&lt;br /&gt;
* &#039;&#039;&#039;Fine-tuned production model&#039;&#039;&#039;: An optimized version designed for everyday TTS applications, fine-tuned on several voices&lt;br /&gt;
* &#039;&#039;&#039;Multilingual model&#039;&#039;&#039;: A family of fine-tuned models with support for other languages&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/canopyai/Orpheus-TTS Official Orpheus TTS repository]&lt;br /&gt;
* [https://huggingface.co/canopylabs/orpheus-3b-0.1-ft Model on Hugging Face]&lt;br /&gt;
* [https://www.baseten.co/library/orpheus-tts/ Orpheus TTS on Baseten]&lt;br /&gt;
&lt;br /&gt;
[[Category:Speech synthesis]]&lt;br /&gt;
[[Category:Open-source software]]&lt;br /&gt;
[[Category:Artificial intelligence]]&lt;br /&gt;
[[Category:Large language models]]&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Orpheus_TTS&amp;diff=22</id>
		<title>Orpheus TTS</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Orpheus_TTS&amp;diff=22"/>
		<updated>2025-09-20T16:04:01Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add Orpheus TTS&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Orpheus TTS&#039;&#039;&#039; is an open-source [[text-to-speech]] (TTS) system developed by Canopy Labs and released in March 2025. Built on the [[Llama (language model)|Llama-3.2-3B]] architecture, it uses a novel approach of using large-language-models with audio tokens instead of traditional TTS-specific architectures.&lt;br /&gt;
&lt;br /&gt;
== Development and Release ==&lt;br /&gt;
&lt;br /&gt;
Orpheus TTS was developed by Canopy Labs, an artificial intelligence startup founded with the stated mission of creating &amp;quot;digital humans that are indistinguishable from real humans.&amp;quot;&amp;lt;ref&amp;gt;https://canopylabs.ai/&amp;lt;/ref&amp;gt; The code and model weights were publicly released on March 18, 2025, under the [[Apache License]] 2.0, making both the model weights and training code freely available.&amp;lt;ref&amp;gt;https://github.com/canopyai/Orpheus-TTS&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
Orpheus TTS differs from conventional text-to-speech systems by using a modified Meta&#039;s Llama-3.2-3B language model as its foundation. It takes in a text prompt and generates audio tokens using the 24 kHz variant of the [[SNAC]] audio tokenizer. The system was trained on a dataset comprising over 100,000 hours of English speech data combined with billions of tokens of textual QA pairs, a hybrid approach designed to maintain linguistic understanding while adding speech synthesis capabilities.&lt;br /&gt;
&lt;br /&gt;
The model supports streaming out-of-the-box and can achieve ~200 milliseconds of streaming latency.&amp;lt;ref&amp;gt;https://www.baseten.co/blog/canopy-labs-selects-baseten-as-preferred-inference-provider-for-orpheus-tts-model/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance Claims and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
Canopy Labs claims that Orpheus TTS delivers &amp;quot;natural intonation, emotion, and rhythm that is superior to SOTA closed source models,&amp;quot; positioning it as competitive with established commercial systems such as [[ElevenLabs]] and other proprietary text-to-speech services.&amp;lt;ref&amp;gt;https://huggingface.co/canopylabs/orpheus-3b-0.1-ft&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, these performance assertions are based primarily on internal evaluations and subjective assessments rather than standardized benchmarks or peer-reviewed studies. The lack of comprehensive comparative analysis with established TTS systems has led to some skepticism within the research community about the extent of its claimed superiority.&lt;br /&gt;
&lt;br /&gt;
== Model Variants ==&lt;br /&gt;
&lt;br /&gt;
The model is available in three variants:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Pretrained model&#039;&#039;&#039;: The base model trained on the full dataset, suitable for research and custom fine-tuning&lt;br /&gt;
* &#039;&#039;&#039;Fine-tuned production model&#039;&#039;&#039;: An optimized version designed for everyday TTS applications, fine-tuned on several voices&lt;br /&gt;
* &#039;&#039;&#039;Multilingual model&#039;&#039;&#039;: A family of fine-tuned models with support for other languages&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
{{reflist}}&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/canopyai/Orpheus-TTS Official Orpheus TTS repository]&lt;br /&gt;
* [https://huggingface.co/canopylabs/orpheus-3b-0.1-ft Model on Hugging Face]&lt;br /&gt;
* [https://www.baseten.co/library/orpheus-tts/ Orpheus TTS on Baseten]&lt;br /&gt;
&lt;br /&gt;
[[Category:Speech synthesis]]&lt;br /&gt;
[[Category:Open-source software]]&lt;br /&gt;
[[Category:Artificial intelligence]]&lt;br /&gt;
[[Category:Large language models]]&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Infobox_TTS_model&amp;diff=20</id>
		<title>Template:Infobox TTS model</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Infobox_TTS_model&amp;diff=20"/>
		<updated>2025-09-20T16:02:15Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Protected &amp;quot;Template:Infobox TTS model&amp;quot; ([Edit=Allow only administrators] (indefinite) [Move=Allow only administrators] (indefinite))&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;includeonly&amp;gt;&lt;br /&gt;
{| class=&amp;quot;infobox&amp;quot; style=&amp;quot;width: 22em; font-size: 90%; text-align: left; border: 1px solid #aaa; background-color: #f9f9f9; color: black; margin: 0.5em 0 0.5em 1em; padding: 0.2em; float: right; clear: right;&amp;quot;&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; font-size: 125%; font-weight: bold; background-color: #ccccff; color: black;&amp;quot; | {{{name|{{PAGENAME}}}}}&lt;br /&gt;
{{#if:{{{image|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; padding: 0.5em;&amp;quot; {{!}} {{{image}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{caption|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; font-style: italic; padding: 0.2em;&amp;quot; {{!}} {{{caption}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Model Information&lt;br /&gt;
{{#if:{{{developer|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; width: 30%;&amp;quot; {{!}} Developer:&lt;br /&gt;
{{!}} {{{developer}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{release_date|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Release date:&lt;br /&gt;
{{!}} {{{release_date}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{latest_version|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Latest version:&lt;br /&gt;
{{!}} {{{latest_version}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{architecture|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Architecture:&lt;br /&gt;
{{!}} {{{architecture}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{parameters|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Parameters:&lt;br /&gt;
{{!}} {{{parameters}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{training_data|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Training data:&lt;br /&gt;
{{!}} {{{training_data}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Capabilities&lt;br /&gt;
{{#if:{{{languages|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Languages:&lt;br /&gt;
{{!}} {{{languages}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{voices|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Voices:&lt;br /&gt;
{{!}} {{{voices}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{voice_cloning|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Voice cloning:&lt;br /&gt;
{{!}} {{{voice_cloning}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{emotion_control|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Emotion control:&lt;br /&gt;
{{!}} {{{emotion_control}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{streaming|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Streaming:&lt;br /&gt;
{{!}} {{{streaming}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{latency|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Latency:&lt;br /&gt;
{{!}} {{{latency}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Availability&lt;br /&gt;
{{#if:{{{license|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} License:&lt;br /&gt;
{{!}} {{{license}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{open_source|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Open source:&lt;br /&gt;
{{!}} {{{open_source}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{code_repository|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Repository:&lt;br /&gt;
{{!}} {{{code_repository}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{model_weights|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Model weights:&lt;br /&gt;
{{!}} {{{model_weights}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{demo|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Demo:&lt;br /&gt;
{{!}} {{{demo}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{website|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Website:&lt;br /&gt;
{{!}} {{{website}}}&lt;br /&gt;
}}&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/includeonly&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Infobox_TTS_model&amp;diff=19</id>
		<title>Template:Infobox TTS model</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Infobox_TTS_model&amp;diff=19"/>
		<updated>2025-09-20T16:02:09Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;&amp;lt;includeonly&amp;gt; {| class=&amp;quot;infobox&amp;quot; style=&amp;quot;width: 22em; font-size: 90%; text-align: left; border: 1px solid #aaa; background-color: #f9f9f9; color: black; margin: 0.5em 0 0.5em 1em; padding: 0.2em; float: right; clear: right;&amp;quot; ! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; font-size: 125%; font-weight: bold; background-color: #ccccff; color: black;&amp;quot; | {{{name|{{PAGENAME}}}}} {{#if:{{{image|}}}| {{!}}-  {{!}} colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; padding: 0.5em;&amp;quot; {{!}} {{{image}}...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;includeonly&amp;gt;&lt;br /&gt;
{| class=&amp;quot;infobox&amp;quot; style=&amp;quot;width: 22em; font-size: 90%; text-align: left; border: 1px solid #aaa; background-color: #f9f9f9; color: black; margin: 0.5em 0 0.5em 1em; padding: 0.2em; float: right; clear: right;&amp;quot;&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; font-size: 125%; font-weight: bold; background-color: #ccccff; color: black;&amp;quot; | {{{name|{{PAGENAME}}}}}&lt;br /&gt;
{{#if:{{{image|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; padding: 0.5em;&amp;quot; {{!}} {{{image}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{caption|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; font-style: italic; padding: 0.2em;&amp;quot; {{!}} {{{caption}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Model Information&lt;br /&gt;
{{#if:{{{developer|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold; width: 30%;&amp;quot; {{!}} Developer:&lt;br /&gt;
{{!}} {{{developer}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{release_date|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Release date:&lt;br /&gt;
{{!}} {{{release_date}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{latest_version|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Latest version:&lt;br /&gt;
{{!}} {{{latest_version}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{architecture|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Architecture:&lt;br /&gt;
{{!}} {{{architecture}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{parameters|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Parameters:&lt;br /&gt;
{{!}} {{{parameters}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{training_data|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Training data:&lt;br /&gt;
{{!}} {{{training_data}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Capabilities&lt;br /&gt;
{{#if:{{{languages|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Languages:&lt;br /&gt;
{{!}} {{{languages}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{voices|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Voices:&lt;br /&gt;
{{!}} {{{voices}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{voice_cloning|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Voice cloning:&lt;br /&gt;
{{!}} {{{voice_cloning}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{emotion_control|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Emotion control:&lt;br /&gt;
{{!}} {{{emotion_control}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{streaming|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Streaming:&lt;br /&gt;
{{!}} {{{streaming}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{latency|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Latency:&lt;br /&gt;
{{!}} {{{latency}}}&lt;br /&gt;
}}&lt;br /&gt;
|-&lt;br /&gt;
! colspan=&amp;quot;2&amp;quot; style=&amp;quot;text-align: center; background-color: #ddddff; font-weight: bold;&amp;quot; | Availability&lt;br /&gt;
{{#if:{{{license|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} License:&lt;br /&gt;
{{!}} {{{license}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{open_source|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Open source:&lt;br /&gt;
{{!}} {{{open_source}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{code_repository|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Repository:&lt;br /&gt;
{{!}} {{{code_repository}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{model_weights|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Model weights:&lt;br /&gt;
{{!}} {{{model_weights}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{demo|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Demo:&lt;br /&gt;
{{!}} {{{demo}}}&lt;br /&gt;
}}&lt;br /&gt;
{{#if:{{{website|}}}|&lt;br /&gt;
{{!}}- &lt;br /&gt;
{{!}} style=&amp;quot;font-weight: bold;&amp;quot; {{!}} Website:&lt;br /&gt;
{{!}} {{{website}}}&lt;br /&gt;
}}&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;/includeonly&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=18</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=18"/>
		<updated>2025-09-20T15:59:04Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: add chatterbox&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Chatterbox]]&lt;br /&gt;
|MIT&lt;br /&gt;
|English&lt;br /&gt;
|✅&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Orpheus TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=17</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=17"/>
		<updated>2025-09-20T15:49:25Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: /* Open Source Models */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|-&lt;br /&gt;
|[[Orpheus TTS]]&lt;br /&gt;
|Apache-2.0&lt;br /&gt;
|English&lt;br /&gt;
|❌&lt;br /&gt;
|❌&lt;br /&gt;
|✅&lt;br /&gt;
|2025&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=16</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=16"/>
		<updated>2025-09-20T01:49:22Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: /* Open Source Models */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| [[F5-TTS]] || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[MaskGCT]] || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[StyleTTS 2]] || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| [[VibeVoice]] || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=VibeVoice&amp;diff=15</id>
		<title>VibeVoice</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=VibeVoice&amp;diff=15"/>
		<updated>2025-09-20T01:48:44Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;&amp;#039;&amp;#039;&amp;#039;VibeVoice&amp;#039;&amp;#039;&amp;#039; is an experimental text-to-speech (TTS) framework developed by Microsoft Research for generating long-form, multi-speaker conversational audio. It was released in August 2025 and is designed to synthesize long-form speech content such as podcasts and audiobooks with up to 4 speakers and with support for voice cloning.&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt;  == Development and Release ==  VibeVoice was developed by a team at Microsoft Res...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;VibeVoice&#039;&#039;&#039; is an experimental [[text-to-speech]] (TTS) framework developed by [[Microsoft Research]] for generating long-form, multi-speaker conversational audio. It was released in August 2025 and is designed to synthesize long-form speech content such as podcasts and audiobooks with up to 4 speakers and with support for voice cloning.&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Development and Release ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice was developed by a team at Microsoft Research led by Zhiliang Peng, Jianwei Yu, and others, with the technical report published on [[arXiv]] in August 2025.&amp;lt;ref&amp;gt;https://arxiv.org/abs/2508.19205&amp;lt;/ref&amp;gt; The project was initially released as open-source software on GitHub and [[Hugging Face]], with model weights made publicly available under the MIT license.&lt;br /&gt;
&lt;br /&gt;
However, the release was disrupted in September 2025 when Microsoft removed the official repository and model weights from public access. According to Microsoft&#039;s statement, this action was taken after discovering &amp;quot;instances where the tool was used in ways inconsistent with the stated intent&amp;quot; and concerns about responsible AI use.&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt; The repository was later restored without implementation code, while community-maintained forks preserved the original materials. The 1.5B pretrained model remains available on Microsoft&#039;s Hugging Face page, but the 7B model was taken down.&lt;br /&gt;
&lt;br /&gt;
Community forks have preserved backups of the code and model weights, including both the 1.5B and 7B models.&amp;lt;ref&amp;gt;https://github.com/vibevoice-community/VibeVoice&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;https://huggingface.co/aoi-ot/VibeVoice-Large&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Technical Architecture ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice uses a hybrid architecture combining large language models with diffusion-based audio generation. The system uses two specialized tokenizers operating at an ultra-low 7.5 Hz frame rate:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Acoustic Tokenizer&#039;&#039;&#039;: A [[variational autoencoder]] (VAE) based encoder-decoder that compresses audio signals while preserving fidelity&lt;br /&gt;
* &#039;&#039;&#039;Semantic Tokenizer&#039;&#039;&#039;: A content-focused encoder trained using [[automatic speech recognition]] as a proxy task&lt;br /&gt;
&lt;br /&gt;
The core model utilizes [[Qwen2.5]] as its base large language model (available in 1.5B and 7B parameter variants), integrated with a lightweight diffusion head for generating acoustic features. This design achieves what the researchers claim is an 80-fold improvement in data compression compared to the [[Encodec]] model while maintaining audio quality.&amp;lt;ref&amp;gt;https://arxiv.org/abs/2508.19205&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Capabilities and Limitations ==&lt;br /&gt;
&lt;br /&gt;
VibeVoice can generate speech sequences up to 90 minutes in length with support for up to four distinct speakers. The model demonstrates several emergent capabilities not explicitly trained for, including:&lt;br /&gt;
&lt;br /&gt;
* Cross-lingual speech synthesis&lt;br /&gt;
* Spontaneous singing (though often off-key)&lt;br /&gt;
* Contextual background music generation&lt;br /&gt;
* Voice cloning from short prompts&lt;br /&gt;
&lt;br /&gt;
However, the system has notable limitations:&lt;br /&gt;
&lt;br /&gt;
* Language support restricted to English and Chinese&lt;br /&gt;
* No explicit modeling of overlapping speech&lt;br /&gt;
* Occasional instability, particularly with Chinese text synthesis&lt;br /&gt;
* Uncontrolled generation of background sounds and music&lt;br /&gt;
* Limited commercial viability due to various technical constraints&amp;lt;ref&amp;gt;https://github.com/microsoft/VibeVoice&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Performance and Evaluation ==&lt;br /&gt;
&lt;br /&gt;
In comparative evaluations against contemporary TTS systems including [[ElevenLabs]], [[Google]]&#039;s Gemini 2.5 Pro TTS, and others, VibeVoice reportedly achieved superior scores in subjective metrics of realism, richness, and user preference. The model also demonstrated competitive [[word error rate]]s when evaluated using speech recognition systems.&amp;lt;ref&amp;gt;https://huggingface.co/microsoft/VibeVoice-1.5B&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
However, these evaluations were conducted on a limited test set of eight conversational transcripts totaling approximately one hour of audio, raising questions about the generalizability of the results to broader use cases.&lt;br /&gt;
&lt;br /&gt;
== Controversies and Concerns ==&lt;br /&gt;
&lt;br /&gt;
The temporary removal of VibeVoice from public access highlighted ongoing concerns about the potential misuse of high-quality synthetic speech technology. Microsoft explicitly warned about the potential for creating [[deepfake]] audio content for impersonation, fraud, or disinformation purposes.&lt;br /&gt;
&lt;br /&gt;
The model&#039;s ability to generate convincing speech from minimal voice prompts, combined with its long-form generation capabilities, raised particular concerns among AI safety researchers about potential misuse for creating fake audio content at scale.&lt;br /&gt;
&lt;br /&gt;
== Community Response ==&lt;br /&gt;
&lt;br /&gt;
Following Microsoft&#039;s temporary withdrawal of the official release, the open-source community created several preservation efforts:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;[https://github.com/vibevoice-community/VibeVoice vibevoice-community/VibeVoice]&#039;&#039;&#039;: A community-maintained fork preserving the original codebase and model weights&lt;br /&gt;
* &#039;&#039;&#039;[https://github.com/voicepowered-ai/VibeVoice-finetuning VibeVoice-finetuning]&#039;&#039;&#039;: Unofficial tools for fine-tuning the models using [[Low-Rank Adaptation]] (LoRA) techniques&lt;br /&gt;
&lt;br /&gt;
These community efforts have enabled continued research and development despite the official restrictions, though they operate independently of Microsoft&#039;s oversight.&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/vibevoice-community/VibeVoice Community-maintained VibeVoice repository]&lt;br /&gt;
* [https://arxiv.org/abs/2508.19205 Original technical report on arXiv]&lt;br /&gt;
&lt;br /&gt;
[[Category:Artificial intelligence]]&lt;br /&gt;
[[Category:Speech synthesis]]&lt;br /&gt;
[[Category:Microsoft Research]]&lt;br /&gt;
[[Category:Open-source software]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=MediaWiki:Common.css&amp;diff=14</id>
		<title>MediaWiki:Common.css</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=MediaWiki:Common.css&amp;diff=14"/>
		<updated>2025-09-20T01:33:24Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Protected &amp;quot;MediaWiki:Common.css&amp;quot; ([Edit=Allow only administrators] (indefinite) [Move=Allow only administrators] (indefinite))&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;/* CSS placed here will be applied to all skins */&lt;br /&gt;
&lt;br /&gt;
.portal-column-left {&lt;br /&gt;
    float: left;&lt;br /&gt;
    width: 100%;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
.portal-column-right {&lt;br /&gt;
    float: right;&lt;br /&gt;
    width: 100%;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Two column layout on wider screens */&lt;br /&gt;
@media (min-width: 720px) {&lt;br /&gt;
    .portal-column-left {&lt;br /&gt;
        width: 48%;&lt;br /&gt;
        margin-right: 2%;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .portal-column-right {&lt;br /&gt;
        width: 48%;&lt;br /&gt;
        margin-left: 2%;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=MediaWiki:Common.css&amp;diff=13</id>
		<title>MediaWiki:Common.css</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=MediaWiki:Common.css&amp;diff=13"/>
		<updated>2025-09-20T01:33:18Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;/* CSS placed here will be applied to all skins */  .portal-column-left {     float: left;     width: 100%; }  .portal-column-right {     float: right;     width: 100%; }  /* Two column layout on wider screens */ @media (min-width: 720px) {     .portal-column-left {         width: 48%;         margin-right: 2%;     }          .portal-column-right {         width: 48%;         margin-left: 2%;     } }&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;/* CSS placed here will be applied to all skins */&lt;br /&gt;
&lt;br /&gt;
.portal-column-left {&lt;br /&gt;
    float: left;&lt;br /&gt;
    width: 100%;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
.portal-column-right {&lt;br /&gt;
    float: right;&lt;br /&gt;
    width: 100%;&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
/* Two column layout on wider screens */&lt;br /&gt;
@media (min-width: 720px) {&lt;br /&gt;
    .portal-column-left {&lt;br /&gt;
        width: 48%;&lt;br /&gt;
        margin-right: 2%;&lt;br /&gt;
    }&lt;br /&gt;
    &lt;br /&gt;
    .portal-column-right {&lt;br /&gt;
        width: 48%;&lt;br /&gt;
        margin-left: 2%;&lt;br /&gt;
    }&lt;br /&gt;
}&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Box_footer&amp;diff=12</id>
		<title>Template:Box footer</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Box_footer&amp;diff=12"/>
		<updated>2025-09-20T01:32:53Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Protected &amp;quot;Template:Box footer&amp;quot; ([Edit=Allow only administrators] (indefinite) [Move=Allow only administrators] (indefinite))&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;|}&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Box_footer&amp;diff=11</id>
		<title>Template:Box footer</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Box_footer&amp;diff=11"/>
		<updated>2025-09-20T01:32:47Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;|}&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;|}&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Box_header&amp;diff=10</id>
		<title>Template:Box header</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Box_header&amp;diff=10"/>
		<updated>2025-09-20T01:32:32Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Protected &amp;quot;Template:Box header&amp;quot; ([Edit=Allow only administrators] (indefinite) [Move=Allow only administrators] (indefinite))&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width:100%; background:{{{3|#f8f9fa}}}; border:1px solid #a2a9b1; margin-bottom:5px;&amp;quot;&lt;br /&gt;
! style=&amp;quot;background:{{{2|#eaecf0}}}; padding:5px; text-align:left; font-weight:bold;&amp;quot; | {{{1}}}&lt;br /&gt;
|-&lt;br /&gt;
| style=&amp;quot;padding:8px;&amp;quot; |&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Template:Box_header&amp;diff=9</id>
		<title>Template:Box header</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Template:Box_header&amp;diff=9"/>
		<updated>2025-09-20T01:32:24Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Created page with &amp;quot;{| style=&amp;quot;width:100%; background:{{{3|#f8f9fa}}}; border:1px solid #a2a9b1; margin-bottom:5px;&amp;quot; ! style=&amp;quot;background:{{{2|#eaecf0}}}; padding:5px; text-align:left; font-weight:bold;&amp;quot; | {{{1}}} |- | style=&amp;quot;padding:8px;&amp;quot; |&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width:100%; background:{{{3|#f8f9fa}}}; border:1px solid #a2a9b1; margin-bottom:5px;&amp;quot;&lt;br /&gt;
! style=&amp;quot;background:{{{2|#eaecf0}}}; padding:5px; text-align:left; font-weight:bold;&amp;quot; | {{{1}}}&lt;br /&gt;
|-&lt;br /&gt;
| style=&amp;quot;padding:8px;&amp;quot; |&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=8</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=8"/>
		<updated>2025-09-19T03:53:20Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| F5-TTS || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| MaskGCT || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| StyleTTS 2 || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| VibeVoice || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Emilia_Dataset&amp;diff=7</id>
		<title>Emilia Dataset</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Emilia_Dataset&amp;diff=7"/>
		<updated>2025-09-19T03:43:07Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Remove related links&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The &#039;&#039;&#039;Emilia Dataset&#039;&#039;&#039; is a large-scale, multilingual, and diverse speech generation dataset derived from in-the-wild speech data. Emilia starts with over 101k hours of speech across six languages, including a wide range of speaking styles for more natural and spontaneous speech generation.&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
&lt;br /&gt;
The Emilia dataset is constructed from a large collection of publicly-available audio on the Internet, such as podcasts, debates, and audiobooks. The dataset was created using &#039;&#039;&#039;Emilia-Pipe&#039;&#039;&#039;, an open-source preprocessing pipeline used to process, transcribe, and filter the dataset.&lt;br /&gt;
&lt;br /&gt;
== Dataset Statistics ==&lt;br /&gt;
&lt;br /&gt;
=== Original Emilia Dataset ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Language !! Code !! Duration (Hours) &lt;br /&gt;
|-&lt;br /&gt;
| English || EN || 46.8k &lt;br /&gt;
|-&lt;br /&gt;
| Chinese || ZH || 49.9k &lt;br /&gt;
|-&lt;br /&gt;
| German || DE || 1.6k &lt;br /&gt;
|-&lt;br /&gt;
| French || FR || 1.4k &lt;br /&gt;
|-&lt;br /&gt;
| Japanese || JA || 1.7k &lt;br /&gt;
|-&lt;br /&gt;
| Korean || KO || 0.2k &lt;br /&gt;
|-&lt;br /&gt;
| &#039;&#039;&#039;Total&#039;&#039;&#039; || - || &#039;&#039;&#039;101.7k&#039;&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Emilia-Large Dataset ===&lt;br /&gt;
The dataset has been expanded to Emilia-Large, a dataset with over 216k hours of speech, making it one of the largest openly-available speech datasets. Emilia-Large combines the original 101k-hour Emilia dataset (licensed under CC BY-NC 4.0) with the new Emilia-YODAS dataset (licensed under CC BY 4.0).&lt;br /&gt;
&lt;br /&gt;
The Emilia-YODAS dataset is based on the YODAS2 dataset, sourced from publicly-available YouTube videos licensed under the Creative Commons license.&lt;br /&gt;
&lt;br /&gt;
== Technical Specifications ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Sampling Rate&#039;&#039;&#039;: 24 kHz&lt;br /&gt;
* &#039;&#039;&#039;Audio Format&#039;&#039;&#039;: WAV files, mono channel&lt;br /&gt;
* &#039;&#039;&#039;Sample Width&#039;&#039;&#039;: 16-bit&lt;br /&gt;
* &#039;&#039;&#039;Audio Quality&#039;&#039;&#039;: DNSMOS P.835 OVRL score of 2.50&lt;br /&gt;
* &#039;&#039;&#039;Languages Supported&#039;&#039;&#039;: 6 (English, Chinese, German, French, Japanese, Korean)&lt;br /&gt;
&lt;br /&gt;
== Emilia-Pipe Processing Pipeline ==&lt;br /&gt;
&lt;br /&gt;
Emilia-Pipe consists of six steps: Standardization, Source Separation, Speaker Diarization, Segmentation by VAD, ASR, and Filtering.&lt;br /&gt;
&lt;br /&gt;
=== Processing Steps ===&lt;br /&gt;
&lt;br /&gt;
==== 1. Standardization ====&lt;br /&gt;
Audio files are converted to WAV format, resampled to 24 kHz and set to mono-channel. Normalization is performed such that the amplitude levels range between -1 to 1, optimizing for a standard decibel level to minimize distortion.&lt;br /&gt;
&lt;br /&gt;
==== 2. Source Separation ====&lt;br /&gt;
This step involves the extraction of clean vocal tracks from audio that may contain background noise or music. The authors employ the Ultimate Vocal Remover model, which has been pretrained, to precisely isolate vocal elements.&lt;br /&gt;
&lt;br /&gt;
==== 3. Speaker Diarization ====&lt;br /&gt;
Speaker diarization techniques partition long-form speech data into multiple utterances based on the speaker using the PyAnnote speaker diarization 3.1 pipeline.&lt;br /&gt;
&lt;br /&gt;
==== 4. Segmentation (VAD) ====&lt;br /&gt;
Voice Activity Detection is used to further segment the audio into smaller, manageable chunks suitable for training.&lt;br /&gt;
&lt;br /&gt;
==== 5. Automated Speech Recognition (ASR) ====&lt;br /&gt;
ASR techniques transcribe the segmented speech data. The medium version of the Whisper model is employed, with batched inference for parallel processing.&lt;br /&gt;
&lt;br /&gt;
==== 6. Filtering ====&lt;br /&gt;
Segments not matching predetermined quality standards (e.g., DNSMOS score) or confidence indicators regarding language differentiation are eliminated, yielding a refined dataset.&lt;br /&gt;
&lt;br /&gt;
== Licensing and Access ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Emilia Dataset&#039;&#039;&#039;: CC BY-NC 4.0 (Non-commercial use only)&lt;br /&gt;
* &#039;&#039;&#039;Emilia-YODAS Dataset&#039;&#039;&#039;: CC BY 4.0&lt;br /&gt;
* &#039;&#039;&#039;Emilia-Pipe Pipeline&#039;&#039;&#039;: Open-source&lt;br /&gt;
&lt;br /&gt;
Users are permitted to use Emilia dataset only for non-commercial purposes under the CC BY-NC-4.0 license. Emilia does not own the copyright to the audio files; the copyright remains with the original owners of the videos or audio.&lt;br /&gt;
&lt;br /&gt;
== Usage ==&lt;br /&gt;
&lt;br /&gt;
=== Loading the Dataset ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
from datasets import load_dataset&lt;br /&gt;
dataset = load_dataset(&amp;quot;amphion/Emilia-Dataset&amp;quot;)&lt;br /&gt;
print(dataset)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading Specific Languages ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
from datasets import load_dataset&lt;br /&gt;
path = &amp;quot;Emilia/DE/*.tar&amp;quot;&lt;br /&gt;
dataset = load_dataset(&amp;quot;amphion/Emilia-Dataset&amp;quot;, &lt;br /&gt;
                      data_files={&amp;quot;de&amp;quot;: path}, &lt;br /&gt;
                      split=&amp;quot;de&amp;quot;, &lt;br /&gt;
                      streaming=True)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://huggingface.co/datasets/amphion/Emilia-Dataset Hugging Face Dataset Page]&lt;br /&gt;
* [https://emilia-dataset.github.io/Emilia-Demo-Page/ Demo Page]&lt;br /&gt;
* [https://github.com/open-mmlab/Amphion/tree/main/preprocessors/Emilia Emilia-Pipe Source Code]&lt;br /&gt;
* [https://arxiv.org/abs/2407.05361 Original Research Paper]&lt;br /&gt;
* [https://arxiv.org/abs/2501.15907 Extended Research Paper]&lt;br /&gt;
&lt;br /&gt;
[[Category:Datasets]]&lt;br /&gt;
[[Category:Speech Datasets]]&lt;br /&gt;
[[Category:Open Source]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Emilia_Dataset&amp;diff=6</id>
		<title>Emilia Dataset</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Emilia_Dataset&amp;diff=6"/>
		<updated>2025-09-19T03:42:43Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add Emilia dataset&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The &#039;&#039;&#039;Emilia Dataset&#039;&#039;&#039; is a large-scale, multilingual, and diverse speech generation dataset derived from in-the-wild speech data. Emilia starts with over 101k hours of speech across six languages, including a wide range of speaking styles for more natural and spontaneous speech generation.&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
&lt;br /&gt;
The Emilia dataset is constructed from a large collection of publicly-available audio on the Internet, such as podcasts, debates, and audiobooks. The dataset was created using &#039;&#039;&#039;Emilia-Pipe&#039;&#039;&#039;, an open-source preprocessing pipeline used to process, transcribe, and filter the dataset.&lt;br /&gt;
&lt;br /&gt;
== Dataset Statistics ==&lt;br /&gt;
&lt;br /&gt;
=== Original Emilia Dataset ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Language !! Code !! Duration (Hours) &lt;br /&gt;
|-&lt;br /&gt;
| English || EN || 46.8k &lt;br /&gt;
|-&lt;br /&gt;
| Chinese || ZH || 49.9k &lt;br /&gt;
|-&lt;br /&gt;
| German || DE || 1.6k &lt;br /&gt;
|-&lt;br /&gt;
| French || FR || 1.4k &lt;br /&gt;
|-&lt;br /&gt;
| Japanese || JA || 1.7k &lt;br /&gt;
|-&lt;br /&gt;
| Korean || KO || 0.2k &lt;br /&gt;
|-&lt;br /&gt;
| &#039;&#039;&#039;Total&#039;&#039;&#039; || - || &#039;&#039;&#039;101.7k&#039;&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Emilia-Large Dataset ===&lt;br /&gt;
The dataset has been expanded to Emilia-Large, a dataset with over 216k hours of speech, making it one of the largest openly-available speech datasets. Emilia-Large combines the original 101k-hour Emilia dataset (licensed under CC BY-NC 4.0) with the new Emilia-YODAS dataset (licensed under CC BY 4.0).&lt;br /&gt;
&lt;br /&gt;
The Emilia-YODAS dataset is based on the YODAS2 dataset, sourced from publicly-available YouTube videos licensed under the Creative Commons license.&lt;br /&gt;
&lt;br /&gt;
== Technical Specifications ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Sampling Rate&#039;&#039;&#039;: 24 kHz&lt;br /&gt;
* &#039;&#039;&#039;Audio Format&#039;&#039;&#039;: WAV files, mono channel&lt;br /&gt;
* &#039;&#039;&#039;Sample Width&#039;&#039;&#039;: 16-bit&lt;br /&gt;
* &#039;&#039;&#039;Audio Quality&#039;&#039;&#039;: DNSMOS P.835 OVRL score of 2.50&lt;br /&gt;
* &#039;&#039;&#039;Languages Supported&#039;&#039;&#039;: 6 (English, Chinese, German, French, Japanese, Korean)&lt;br /&gt;
&lt;br /&gt;
== Emilia-Pipe Processing Pipeline ==&lt;br /&gt;
&lt;br /&gt;
Emilia-Pipe consists of six steps: Standardization, Source Separation, Speaker Diarization, Segmentation by VAD, ASR, and Filtering.&lt;br /&gt;
&lt;br /&gt;
=== Processing Steps ===&lt;br /&gt;
&lt;br /&gt;
==== 1. Standardization ====&lt;br /&gt;
Audio files are converted to WAV format, resampled to 24 kHz and set to mono-channel. Normalization is performed such that the amplitude levels range between -1 to 1, optimizing for a standard decibel level to minimize distortion.&lt;br /&gt;
&lt;br /&gt;
==== 2. Source Separation ====&lt;br /&gt;
This step involves the extraction of clean vocal tracks from audio that may contain background noise or music. The authors employ the Ultimate Vocal Remover model, which has been pretrained, to precisely isolate vocal elements.&lt;br /&gt;
&lt;br /&gt;
==== 3. Speaker Diarization ====&lt;br /&gt;
Speaker diarization techniques partition long-form speech data into multiple utterances based on the speaker using the PyAnnote speaker diarization 3.1 pipeline.&lt;br /&gt;
&lt;br /&gt;
==== 4. Segmentation (VAD) ====&lt;br /&gt;
Voice Activity Detection is used to further segment the audio into smaller, manageable chunks suitable for training.&lt;br /&gt;
&lt;br /&gt;
==== 5. Automated Speech Recognition (ASR) ====&lt;br /&gt;
ASR techniques transcribe the segmented speech data. The medium version of the Whisper model is employed, with batched inference for parallel processing.&lt;br /&gt;
&lt;br /&gt;
==== 6. Filtering ====&lt;br /&gt;
Segments not matching predetermined quality standards (e.g., DNSMOS score) or confidence indicators regarding language differentiation are eliminated, yielding a refined dataset.&lt;br /&gt;
&lt;br /&gt;
== Licensing and Access ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Emilia Dataset&#039;&#039;&#039;: CC BY-NC 4.0 (Non-commercial use only)&lt;br /&gt;
* &#039;&#039;&#039;Emilia-YODAS Dataset&#039;&#039;&#039;: CC BY 4.0&lt;br /&gt;
* &#039;&#039;&#039;Emilia-Pipe Pipeline&#039;&#039;&#039;: Open-source&lt;br /&gt;
&lt;br /&gt;
Users are permitted to use Emilia dataset only for non-commercial purposes under the CC BY-NC-4.0 license. Emilia does not own the copyright to the audio files; the copyright remains with the original owners of the videos or audio.&lt;br /&gt;
&lt;br /&gt;
== Usage ==&lt;br /&gt;
&lt;br /&gt;
=== Loading the Dataset ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
from datasets import load_dataset&lt;br /&gt;
dataset = load_dataset(&amp;quot;amphion/Emilia-Dataset&amp;quot;)&lt;br /&gt;
print(dataset)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading Specific Languages ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
from datasets import load_dataset&lt;br /&gt;
path = &amp;quot;Emilia/DE/*.tar&amp;quot;&lt;br /&gt;
dataset = load_dataset(&amp;quot;amphion/Emilia-Dataset&amp;quot;, &lt;br /&gt;
                      data_files={&amp;quot;de&amp;quot;: path}, &lt;br /&gt;
                      split=&amp;quot;de&amp;quot;, &lt;br /&gt;
                      streaming=True)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== External Links ==&lt;br /&gt;
&lt;br /&gt;
* [https://huggingface.co/datasets/amphion/Emilia-Dataset Hugging Face Dataset Page]&lt;br /&gt;
* [https://emilia-dataset.github.io/Emilia-Demo-Page/ Demo Page]&lt;br /&gt;
* [https://github.com/open-mmlab/Amphion/tree/main/preprocessors/Emilia Emilia-Pipe Source Code]&lt;br /&gt;
* [https://arxiv.org/abs/2407.05361 Original Research Paper]&lt;br /&gt;
* [https://arxiv.org/abs/2501.15907 Extended Research Paper]&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
* [[F5-TTS]]&lt;br /&gt;
* [[Text-to-Speech]]&lt;br /&gt;
* [[Speech Generation]]&lt;br /&gt;
* [[Multilingual Datasets]]&lt;br /&gt;
* [[Voice Cloning]]&lt;br /&gt;
&lt;br /&gt;
[[Category:Datasets]]&lt;br /&gt;
[[Category:Speech Datasets]]&lt;br /&gt;
[[Category:Open Source]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=5</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=5"/>
		<updated>2025-09-19T03:28:52Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Update styling&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Welcome to TTS Wiki =&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
Note: This Wiki is still a work-in-progress. Contributions are welcome!&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| F5-TTS || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| MaskGCT || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| StyleTTS 2 || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| VibeVoice || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=4</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=4"/>
		<updated>2025-09-19T03:26:42Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{DISPLAYTITLE:TTS Wiki}}&lt;br /&gt;
&lt;br /&gt;
= Welcome to TTS Wiki =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| F5-TTS || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| MaskGCT || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| StyleTTS 2 || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| VibeVoice || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=3</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=3"/>
		<updated>2025-09-19T03:25:24Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Add home page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{DISPLAYTITLE:TTS Wiki}}&lt;br /&gt;
__NOTOC__&lt;br /&gt;
&lt;br /&gt;
= Welcome to TTS Wiki =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;TTS Wiki&#039;&#039;&#039; is a collaborative knowledge base dedicated to documenting and comparing latest &#039;&#039;&#039;Text-to-Speech (TTS)&#039;&#039;&#039; models and technologies. Our mission is to provide comprehensive, up-to-date information about the rapidly evolving landscape of speech synthesis.&lt;br /&gt;
&lt;br /&gt;
== Getting Started ==&lt;br /&gt;
&lt;br /&gt;
=== For Developers ===&lt;br /&gt;
* [[Installation Guides]] - Setup instructions for various TTS models&lt;br /&gt;
* [[Installation Guides|Finetuning Guides]] - Walkthroughs for fine-tuning TTS models&lt;br /&gt;
* [[Licensing Overview]] - Commercial usage rights and restrictions&lt;br /&gt;
&lt;br /&gt;
== Model Categories ==&lt;br /&gt;
&lt;br /&gt;
=== Open Source Models ===&lt;br /&gt;
{| class=&amp;quot;wikitable sortable&amp;quot; style=&amp;quot;width: 100%;&amp;quot;&lt;br /&gt;
! Model !! License !! Languages !! Voice Cloning !! Conversational &lt;br /&gt;
!Fine-Tuning!! Date Released&lt;br /&gt;
|-&lt;br /&gt;
| F5-TTS || MIT, CC-BY-NC || English, Chinese || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| MaskGCT || MIT, CC-BY-NC || English, Chinese, Korean, Japanese, French, German || ✅ || ❌ &lt;br /&gt;
|❌|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| StyleTTS 2 || MIT || English || ✅ || ❌ &lt;br /&gt;
|✅|| 2024&lt;br /&gt;
|-&lt;br /&gt;
| VibeVoice || MIT || English, Chinese || ❌ || ✅ &lt;br /&gt;
|✅|| 2025&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Commercial Services ===&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
== Contributing ==&lt;br /&gt;
&lt;br /&gt;
Contributions are welcome! Here&#039;s how you can help:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Add new models&#039;&#039;&#039; - Document recently released TTS systems&lt;br /&gt;
* &#039;&#039;&#039;Update comparisons&#039;&#039;&#039; - Share performance benchmarks and quality tests&lt;br /&gt;
* &#039;&#039;&#039;Write tutorials and guides&#039;&#039;&#039; - Help others learn to use different TTS tools&lt;br /&gt;
* &#039;&#039;&#039;Upload samples&#039;&#039;&#039; - Provide audio examples for model comparisons (please do not upload copyrighted content!)&lt;br /&gt;
* &#039;&#039;&#039;Fix information&#039;&#039;&#039; - Correct outdated or inaccurate details&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Disclaimer:&#039;&#039;&#039; This wiki is maintained by the community and information may not always be current. Always verify details with official sources before making production decisions. Voice cloning should only be used with proper consent and for ethical purposes.&lt;br /&gt;
&lt;br /&gt;
[[Category:Main]]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
	<entry>
		<id>https://tts.wiki/index.php?title=Main_Page&amp;diff=2</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://tts.wiki/index.php?title=Main_Page&amp;diff=2"/>
		<updated>2025-09-19T03:10:56Z</updated>

		<summary type="html">&lt;p&gt;Ttswikiadmin: Protected &amp;quot;Main Page&amp;quot;: High traffic page ([Edit=Allow only administrators] (indefinite) [Move=Allow only administrators] (indefinite))&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;strong&amp;gt;MediaWiki has been installed.&amp;lt;/strong&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Consult the [https://www.mediawiki.org/wiki/Special:MyLanguage/Help:Contents User&#039;s Guide] for information on using the wiki software.&lt;br /&gt;
&lt;br /&gt;
== Getting started ==&lt;br /&gt;
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:Configuration_settings Configuration settings list]&lt;br /&gt;
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:FAQ MediaWiki FAQ]&lt;br /&gt;
* [https://lists.wikimedia.org/postorius/lists/mediawiki-announce.lists.wikimedia.org/ MediaWiki release mailing list]&lt;br /&gt;
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Localisation#Translation_resources Localise MediaWiki for your language]&lt;br /&gt;
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:Combating_spam Learn how to combat spam on your wiki]&lt;/div&gt;</summary>
		<author><name>Ttswikiadmin</name></author>
	</entry>
</feed>