Configuration Parsing Warning:Invalid JSON for config file config.json

Flare-TTS 28M

Welcome to Flare-TTS 28M, an open-source text-to-speech model with 28 million parameters trained on LJSpeech. This model was originally created by LH-Tech AI.

Quality and results

This model is okayish quality but it still sounds a bit robotish but you can clearly understand what the model tries to say. See this model as a proof-of-concept or a first-beta. Example:

Training process

We trained this model for ~300 epochs on a single A6000 GPU for ~24 hours. The full training code can be found in this repo as start.sh and train.py. Just run start.sh to train this model yourself.

Architecture

This model was trained using CoquiTTS. For the architecture we chose GlowTTS.

Training dataset

We trained on the full LJSpeech dataset. Thanks to keithito for this :-)

How to use

As soon as you have the model checkpoint (model.pth) and config.json on your device, you can generate a sample using:

tts --text "Hello world, this is my first trained TTS model." \
    --model_path model.pth \
    --config_path config.json \
    --out_path output_1.wav

Final thoughts

We don't think it's perfect - it's more like a proof of concept. So please do not use this model for production use cases but more for experiments. We are happy to share more of this soon - stay tuned for Flare-TTS v2 :D

Credit

Credit goes to LH-Tech AI. :-)

Notes:

Due to LH-Tech AI training this model with unoffical CompactAI trainig configurations, it will not be present in any of CompactAI's inference engines!

Downloads last month
42
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train CompactAI-O/Flare-TTS-28M

Space using CompactAI-O/Flare-TTS-28M 1

Collection including CompactAI-O/Flare-TTS-28M