SkillAgentSearch skills...

Rasa

Expressive TTS Dataset for Assamese, Bengali, and Tamil.

Install / Use

/learn @AI4Bharat/Rasa
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

Rasa: Building Expressive Speech Synthesis Systems for Indian Languages in Low-resource Settings

🎉 Accepted at INTERSPEECH 2024

We release Rasa, the first multilingual expressive TTS dataset for any Indian language, which contains 10 hours of neutral speech and 1-3 hours of expressive speech for each of the 6 Ekman emotions covering 3 languages: Assamese, Bengali, & Tamil. Our ablation studies reveal that just 1 hour of neutral and 30 minutes of expressive data can yield a Fair system as indicated by MUSHRA scores. Increasing neutral data to 10 hours, with minimal expressive data, significantly enhances expressiveness. This offers a practical recipe for resource-constrained languages, prioritizing easily obtainable neutral data alongside smaller amounts of expressive data. We show the importance of syllabically balanced data and pooling emotions to enhance expressiveness. We also highlight challenges in generating specific emotions, e.g., fear and surprise.

TL;DR: We open-source Expressive Text-To-Speech dataset and models for 3 Indian languages: Assamese, Bengali, and, Tamil..

Authors: Praveen S V*, Ashwin Sankar*, Giri Raju, Mitesh M. Khapra

Downloads:

Model checkpoints can be downloaded at https://github.com/AI4Bharat/Rasa/releases

Data can be downloaded at https://huggingface.co/datasets/ai4bharat/Rasa

Setup:

Environment Setup

git clone https://github.com/ai4bharat/rasa.git
cd rasa
conda env create -f environment.yml

cd Trainer
pip3 install -e .[all]
cd ..

cd TTS
pip3 install -e .[all]
cd ..

Data Setup

The data should be formatted similar to LJSpeech.

Training:

  1. Set the configuration with main.py, vocoder.py, configs and run.sh. Make sure to update the CUDA_VISIBLE_DEVICES in all these files.
  2. Train by executing sh configs/train.sh

Inference

    output_dir="/path/to/where/you/want/the/output/saved"
    base_path="/path/to/base/directory/of/fastpitch"
    vocoder_path="/path/to/base/directory/of/hifigan"

    mkdir -p ${output_dir}

    python3 -m TTS.bin.synthesize --text "/path/to/metadata_test.csv" \
        --model_path ${base_path}/best_model.pth \
        --config_path ${base_path}/config.json \
        --vocoder_path ${vocoder_path}/best_model.pth \
        --vocoder_config_path ${vocoder_path}/config.json \
        --out_path ${output_dir} \
        --use_cuda t \
        --use_emotion t \

License

CC-BY-4.0

Citation

If you use this dataset, please cite:

@inproceedings{ai4bharat2024rasa,
  author={Praveen Srinivasa Varadhan and Ashwin Sankar and Giri Raju and Mitesh M. Khapra},
  title={{Rasa: Building Expressive Speech Synthesis Systems for Indian Languages in Low-resource Settings}},
  year=2024,
  booktitle={Proc. INTERSPEECH 2024},
}

Code Reference:

  1. https://github.com/coqui-ai/TTS
  2. https://github.com/coqui-ai/Trainer
View on GitHub
GitHub Stars15
CategoryDevelopment
Updated2mo ago
Forks3

Languages

Python

Security Score

75/100

Audited on Jan 8, 2026

No findings