PyTorch
ELM
hubert

AfriHuBERT: A self-supervised speech representation model for African languages

Model description

This is a compact multilingual self-supervised speech encoder based on mHuBERT-147. We performed continued pretraining through multilingual adaptive finetuning (MAFT) on over 10,000 hours of African language data aggregated from various sources. This model can be considered the fourth iteration of mHuBERT-147, specifically trained on African languages. According to the paper, this is the AfriHuBERT-n model. You can click here for the AfriHuBERT-o model.

Pretraining data

  • Dataset: AfriHuBERT was trained on data from 11 major sources, including BibleTTS, Kallaama, MMS Ulab v2, NaijaVoices, and NCHLT. All sources and their licenses are shown in the table below. Please refer to the paper for more information. Logo

Language Coverage

AfriHuBERT covers 1,230 languages in total including 1,226 indigenous African languages

BibTeX entry and citation info.

@misc{alabi2024afrihubertselfsupervisedspeechrepresentation,
      title={AfriHuBERT: A self-supervised speech representation model for African languages}, 
      author={Jesujoba O. Alabi and Xuechen Liu and Dietrich Klakow and Junichi Yamagishi},
      year={2024},
      eprint={2409.20201},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2409.20201}, 
}
Downloads last month
134
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ajesujoba/AfriHuBERT

Finetuned
(8)
this model

Datasets used to train ajesujoba/AfriHuBERT