ekwek commited on
Commit
2a5aeee
·
verified ·
1 Parent(s): 30ae4dd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md CHANGED
@@ -1,4 +1,35 @@
1
  ---
2
  license: apache-2.0
3
  pipeline_tag: feature-extraction
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  ---
 
1
  ---
2
  license: apache-2.0
3
  pipeline_tag: feature-extraction
4
+ ---
5
+
6
+ # Soprano: Instant, Ultra‑Realistic Text‑to‑Speech
7
+
8
+ <div align="center">
9
+
10
+ <img width="640" height="320" alt="soprano-github" src="https://github.com/user-attachments/assets/4d612eac-23b8-44e6-8c59-d7ac14ebafd1" />
11
+
12
+ [![Alt Text](https://img.shields.io/badge/Github-Repo-black?logo=github)](https://github.com/ekwek1/soprano)
13
+ [![Alt Text](https://img.shields.io/badge/HuggingFace-Demo-yellow?logo=huggingface)](https://huggingface.co/spaces/ekwek/Soprano-TTS)
14
+ </div>
15
+
16
+ ### 📰 News
17
+ **2026.01.13 - [Soprano-Factory](https://github.com/ekwek1/soprano-factory) released! You can now train/fine-tune your own Soprano models.**
18
+ 2025.12.22 - Soprano-80M released! [Code](https://github.com/ekwek1/soprano) | [Demo](https://huggingface.co/spaces/ekwek/Soprano-TTS)
19
+
20
+ ---
21
+
22
+ This repository contains **Soprano-Encoder**, which converts raw audio into audio tokens that the LLM backbone can recognize.
23
+
24
+ ## Overview
25
+
26
+ **Soprano** is an ultra‑lightweight, on-device text‑to‑speech (TTS) model designed for expressive, high‑fidelity speech synthesis at unprecedented speed. Soprano was designed with the following features:
27
+ - Up to **2000x** real-time generation on GPU and **20x** real-time on CPU
28
+ - **Lossless streaming** with **<15 ms** latency on GPU, **<250 ms** on CPU
29
+ - **<1 GB** memory usage with a compact 80M parameter architecture
30
+ - **Infinite generation length** with automatic text splitting
31
+ - Highly expressive, crystal clear audio generation at **32kHz**
32
+ - Widespread support for CUDA, CPU, and MPS devices on Windows, Linux, and Mac
33
+ - Supports WebUI, CLI, and OpenAI-compatible endpoint for easy and production-ready inference
34
+
35
  ---