Documentation Index
Fetch the complete documentation index at: https://mintlify.com/KittenML/KittenTTS/llms.txt
Use this file to discover all available pages before exploring further.
kitten-tts-nano is the smallest model in the KittenTTS family, designed for real-time inference on constrained hardware. It is available in two variants: standard fp32 and quantized int8.
Variants
| Variant | Model ID | Size | Format |
|---|---|---|---|
| nano (fp32) | KittenML/kitten-tts-nano-0.8-fp32 | 56 MB | Full precision |
| nano (int8) | KittenML/kitten-tts-nano-0.8-int8 | 25 MB | int8 quantized |
Both variants share the same 15M parameter architecture. The int8 variant uses quantization to reduce file size at the cost of minor precision loss.
When to use this model
- Edge devices with limited CPU, memory, or storage
- Real-time text-to-speech on constrained hardware
- Embedded systems and IoT devices
- Mobile applications where binary size matters
- High-throughput batch workloads where inference speed is critical
Usage
nano (fp32)
nano (int8)
Models are downloaded from Hugging Face on first use and cached locally. Pass
cache_dir to control where they are stored.Related models
- kitten-tts-mini — 80M params, highest quality
- kitten-tts-micro — 40M params, balanced quality and speed