OuteTTS version 0.3 introduces multiple model variants tailored for diverse use cases.
This release significantly enhances the naturalness and coherence of speech synthesis by adding punctuation support, improving the flow and clarity of generated speech.
The following punctuation marks are supported: `'.', '!', '?', ',', '"', '„', '¡', '¿', '…', '...', '。', '!', '?', ',', '؟'`. These are converted into special tokens, for instance, `.` is transformed into `<|period|>`.
Additionally, the models were trained on refined and extended datasets, offering broader linguistic coverage. With this version, two new languages, **German (de)** and **French (fr)**, are supported, bringing the total to six languages: **English (en)**, **Japanese (jp)**, **Korean (ko)**, **Chinese (zh)**, **French (fr)**, and **German (de)**.
OuteTTS is a solution designed to extend any existing large language model (LLM) with text-to-speech (TTS) and speech-to-speech capabilities. By preserving the original architecture, it ensures high compatibility with a broad range of libraries and tools, making it easy to integrate speech functionalities without compromising on flexibility.
Experimental voice control features are also included, though they are in very early stage of development. Due to limited data, these features may produce inconsistent results and might sometimes be ignored by the model.
Special thanks to **Hugging Face** 🤗 for providing the GPU grant that made training this model possible!
## Available Variants
### OuteTTS-0.3-500M
- **Base**: Qwen2.5-0.5B (Apache-2.0)
- **TTS Model License**: CC-BY-SA-4.0
- **Training**: 10,000 hours of speech audio (~4 billion tokens)
- **Supported Languages**: en, jp, ko (small dataset), zh, fr, de
### OuteTTS-0.3-1B
- **Base**: OLMo-1B (Apache-2.0)
- **TTS Model License**: CC-BY-NC-SA-4.0 _(Incorporates the Emilia dataset, for improved quality)_
- **Training**: 20,000 hours of speech audio (~8 billion tokens)
text="Speech synthesis is the artificial production of human speech.",
temperature=0.4,
repetition_penalty=1.1,
max_length=4096,
speaker=speaker,
)
output = interface.generate(config=gen_cfg)
# Save the generated speech to a file
output.save("output.wav")
```
### Additional Usage Examples
> [!IMPORTANT]
> For additional usage examples and recommendations, visit the: [GitHub repository](https://github.com/edwko/OuteTTS?tab=readme-ov-file#usage).
### Generation Performance
> [!IMPORTANT]
> The model performs best with 30-second generation batches. This window is reduced based on the length of your speaker samples. For example, if the speaker reference sample is 10 seconds, the effective window becomes approximately 20 seconds. I am currently working on adding batched generation capabilities to the library, along with further improvements that are not yet implemented.
---
## Dataset Attribution
The OuteTTS-0.3-1B training data incorporates various publicly available speech datasets. Below is a summary of the key data sources: