Iqra’Eval is a shared task aimed at advancing automatic assessment of Qur’anic recitation pronunciation by leveraging computational methods to detect and diagnose pronunciation errors. The focus on Qur’anic recitation provides a standardized and well-defined context for evaluating Modern Standard Arabic (MSA) pronunciation, where precise articulation is not only valued but essential for correctness according to established Tajweed rules.
Participants will develop systems capable of:
The Iqra’Eval shared task focuses on automatic mispronunciation detection and diagnosis in Qur’anic recitation. Given:
the goal is to identify any pronunciation errors, localize them within the phoneme sequence, and classify the type of error based on Tajweed rules.
Each participant’s system must predict the sequence of phonemes that the reciter actually produced. A standardized phonemizer (Nawar Halabi’s phonetizer) will be used to generate the “gold” phoneme sequence from the reference transcript for comparison.
Key subtasks:
Suppose the reference verse (fully vowelized) is:
إِنَّ اللَّهَ عَلَىٰ كُلِّ شَيْءٍ قَدِيرٌ
(inna l-lāha ʿalā kulli shay’in qadīrun)
The gold phoneme sequence (using the standard phonemizer) might be:
inna l l aa h a ʕ a l a k u l l i ʃ a y ’ i n q a d i r u n
If a reciter mispronounces “قَدِيرٌ” (qadīrun) as “كَدِيرٌ” (kadīrun), that corresponds to a substitution at the very start of that word: phoneme /q/ → /k/.
A well-trained system should:
Figure: Example of a phoneme-level comparison between reference vs. predicted for an Arabic Qur’anic recitation.
Systems will be scored on their ability to detect and correctly classify phoneme-level errors:
A final Composite Error Score (CES) will be computed by combining:
(Detailed evaluation weights and scripts will be made available on June 5, 2025.)
Participants are required to submit a CSV file named submission.csv
containing the predicted phoneme sequences for each audio sample. The file must have exactly two columns:
Below is a minimal example illustrating the required format:
ID,Labels 0000_0001, i n n a m a a y a k h a l l a h a m i n ʕ i b a a d i h u l ʕ u l a m 0000_0002, m a a n a n s a k h u m i n i ʕ a a y a t i n 0000_0003, y u k h i k u m u n n u ʔ a u ʔ a m a n a t a n m m i n h u …
The first column (ID) should match exactly the audio filenames (without extension). The second column (Labels) is the predicted phoneme string.
Important:
submission.csv
.All data are hosted on Hugging Face. Two main splits are provided:
df = load_dataset("mostafaashahin/IqraEval_Training_Data", split="train")
df = load_dataset("mostafaashahin/IqraEval_Training_Data", split="dev")
A sample submission file (sample_submission.csv
) is also provided in the repository.
Column Definitions:
sentence
: Original sentence text (may be partially diacritized or non-diacritized).q_index
: If from the Quran, the verse index (0–6265, including Basmalah); otherwise -1
.start_word_index
, end_word_index
: Word positions within the verse (or -1
if non-Quranic).tashkeel_sentence
: Fully diacritized sentence (auto-generated via a diacritization tool).phoneme
: Phoneme sequence corresponding to the diacritized sentence (Nawar Halabi phonetizer).
Data Splits:
• Training (train): 79 hours total
• Development (dev): 3.4 hours total
We also provide a high-quality TTS corpus for auxiliary experiments (e.g., data augmentation, synthetic pronunciation error simulation). This TTS set can be loaded via:
df_tts = load_dataset("IqraEval/Iqra_TTS", split="train")
Researchers who wish to experiment with “synthetic mispronunciations” can use the TTS waveform + forced-alignment pipeline to generate various kinds of pronunciation errors in a controlled manner.
For detailed instructions on data access, phonetizer installation, and baseline usage, please refer to the GitHub README.
Further details on evaluation criteria (exact scoring weights), submission templates, and any clarifications will be posted on the shared task website when test data are released (June 5, 2025). Stay tuned!