Update README.md
Browse files
README.md
CHANGED
|
@@ -1,4 +1,35 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
pipeline_tag: feature-extraction
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
pipeline_tag: feature-extraction
|
| 4 |
+
---
|
| 5 |
+
|
| 6 |
+
# Soprano: Instant, Ultra‑Realistic Text‑to‑Speech
|
| 7 |
+
|
| 8 |
+
<div align="center">
|
| 9 |
+
|
| 10 |
+
<img width="640" height="320" alt="soprano-github" src="https://github.com/user-attachments/assets/4d612eac-23b8-44e6-8c59-d7ac14ebafd1" />
|
| 11 |
+
|
| 12 |
+
[](https://github.com/ekwek1/soprano)
|
| 13 |
+
[](https://huggingface.co/spaces/ekwek/Soprano-TTS)
|
| 14 |
+
</div>
|
| 15 |
+
|
| 16 |
+
### 📰 News
|
| 17 |
+
**2026.01.13 - [Soprano-Factory](https://github.com/ekwek1/soprano-factory) released! You can now train/fine-tune your own Soprano models.**
|
| 18 |
+
2025.12.22 - Soprano-80M released! [Code](https://github.com/ekwek1/soprano) | [Demo](https://huggingface.co/spaces/ekwek/Soprano-TTS)
|
| 19 |
+
|
| 20 |
+
---
|
| 21 |
+
|
| 22 |
+
This repository contains **Soprano-Encoder**, which converts raw audio into audio tokens that the LLM backbone can recognize.
|
| 23 |
+
|
| 24 |
+
## Overview
|
| 25 |
+
|
| 26 |
+
**Soprano** is an ultra‑lightweight, on-device text‑to‑speech (TTS) model designed for expressive, high‑fidelity speech synthesis at unprecedented speed. Soprano was designed with the following features:
|
| 27 |
+
- Up to **2000x** real-time generation on GPU and **20x** real-time on CPU
|
| 28 |
+
- **Lossless streaming** with **<15 ms** latency on GPU, **<250 ms** on CPU
|
| 29 |
+
- **<1 GB** memory usage with a compact 80M parameter architecture
|
| 30 |
+
- **Infinite generation length** with automatic text splitting
|
| 31 |
+
- Highly expressive, crystal clear audio generation at **32kHz**
|
| 32 |
+
- Widespread support for CUDA, CPU, and MPS devices on Windows, Linux, and Mac
|
| 33 |
+
- Supports WebUI, CLI, and OpenAI-compatible endpoint for easy and production-ready inference
|
| 34 |
+
|
| 35 |
---
|