Resource name | link |
---|---|
Dataset | Lahaja |
Model - M1 | Coming soon |
Model - M2 | Coming soon |
Model - M3 | Coming soon |
Model - M4 | Coming soon |
Hindi, one of the most spoken language of India, exhibits a diverse array of accents due to its usage among individuals from diverse linguistic origins. To enable a robust evaluation of Hindi ASR systems on multiple accents, we create a benchmark, LAHAJA, which contains read and extempore speech on a diverse set of topics and use cases, with a total of 12.5 hours of Hindi audio, sourced from 132 speakers spanning 83 districts of India. We evaluate existing open-source and commercial models on LAHAJA and find their performance to be poor. We then train models using different datasets and find that our model trained on multilingual data with good speaker diversity outperforms existing models by a significant margin. We also present a fine grained analysis which shows that the performance declines for speakers from North-East and South India, especially with content heavy in named entities and specialized terminology.
Resource name | link |
---|---|
Dataset | Lahaja |
Model - M1 | Coming soon |
Model - M2 | Coming soon |
Model - M3 | Coming soon |
Model - M4 | Coming soon |
The lahaja contains two thing:
audio
folder which contains all the speech datameta-data-lahaja.csv
mapping file, that contains transcripts along with the additional speaker details (given below)The csv contains the following data with respect to the audio:
verbatim
— Verbatim version of the transcript.normalized
— Normalized version of the transcript.duration
— Duration of audio recorded (seconds)scenario
— Type of speech data (read/extempore)fname
— File namenative_language
— Native language of the participantgender
— Gender (Male/Female)age_group
— Age group of the participantnative_state
— Speaker’s native statenative_district
— Speaker’s native districtsp_id
— Unique speaker idtext
— Cleaned version of normalized transcript (after braces removal).lang
— langauge id (hi)job_category
— speaker’s job categoryoccupation_domain
— speaker’s domain of occupation (Education and Research, Healthcare [Medical & Pharma], Government, Technology and Services, Information and Media, Financial Services [Banking and Insurance], Transportation and Logistics, Entertainment, Social service, Manufacturing & Retail)occupation
— Speaker’s occupationjob_type
— speakers’s job category (Part Time, Full Time, Other)age-group
— speaker’s age group (18-30, 30-45, 45-60 & 60+ )qual
— speaker’s highest education qualificationLahaja folder tree
Lahaja
├── audio
│ ├── <filename>.wav
│ .
│ .
│ .
└── meta-data-lahaja.csv
Install NeMo
Training command
python ${RUNNER_PATH}/speech_to_text_hybrid_rnnt_ctc_bpe.py \
--config-path=${CONFIG_PATH} \
--config-name=${CONFIG_NAME} \
trainer.devices=-1 \
trainer.accelerator="gpu" \
trainer.strategy="ddp_find_unused_parameters_true" \
exp_manager.name=<exp_name> \
exp_manager.wandb_logger_kwargs.name=<exp_name> \
exp_manager.explicit_log_dir=$LOG_DIR
Inference
OMP_NUM_THREADS=64 python ${RUNNER_PATH}/transcribe_speech.py \
model_path=$MODEL_PATH \
dataset_manifest=$MANIFEST_PATH.json \
output_filename=$SAVE_FILE_NAME \
langid=$LANGID \
batch_size=64 \
compute_timestamps=False \
compute_langs=False \
cuda=$GPU_ID \
amp=True \
append_pred=False
If you benefit from this dataset, kindly cite as follows:
@misc{
to be updated
} -->