hypaai commited on
Commit
1241eff
·
verified ·
1 Parent(s): 93a5744

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -72,7 +72,7 @@ In addition to the text-to-text translations, the dataset includes parallel spee
72
  - **Text-to-Text Translations:** English sentences paired with their translations in Igbo, Yoruba, and Hausa.
73
  - **Speech-to-Speech Recordings:** Audio recordings of native speakers reading both the English texts and the corresponding translated texts.
74
 
75
- This dual modality (text and audio) supports various downstream tasks such as machine translation, automatic speech recognition (ASR), text-to-speech (TTS), and cross-lingual transfer learning.
76
 
77
  ---
78
 
@@ -85,7 +85,7 @@ This dual modality (text and audio) supports various downstream tasks such as ma
85
  - Igbo
86
  - Yoruba
87
  - Hausa
88
- - **Format:** Typically stored in CSV/TSV or JSON files where each record contains:
89
  - The English sentence.
90
  - The corresponding translations for each target language.
91
  - **Splits:** The dataset is divided according to each low-resource language to mirror Fleurs partitioning but particularly for African Languages.
@@ -101,7 +101,7 @@ This dual modality (text and audio) supports various downstream tasks such as ma
101
  - **Format:** Audio files (e.g., WAV) with accompanying metadata files (e.g., CSV/JSON) that include:
102
  - Unique identifier linking to text entries.
103
  - Language code.
104
- - Duration and other audio properties.
105
  - **Parallelism:** Each audio file is aligned with the corresponding text in both the source (English) and target languages.
106
 
107
  ---
@@ -195,6 +195,7 @@ The Hypa_Fleurs dataset can be used for various research and development tasks,
195
  - **Speech Recognition (ASR):** Developing systems that can transcribe speech in under-resourced languages.
196
  - **Text-to-Speech (TTS):** Creating natural-sounding TTS systems using paired audio-text data.
197
  - **Cross-lingual Learning:** Supporting transfer learning and multilingual model training.
 
198
 
199
  ---
200
 
 
72
  - **Text-to-Text Translations:** English sentences paired with their translations in Igbo, Yoruba, and Hausa.
73
  - **Speech-to-Speech Recordings:** Audio recordings of native speakers reading both the English texts and the corresponding translated texts.
74
 
75
+ This dual modality (text and audio) supports various downstream tasks such as machine translation, automatic speech recognition (ASR), text-to-speech (TTS), language identification (LI), and cross-lingual transfer learning.
76
 
77
  ---
78
 
 
85
  - Igbo
86
  - Yoruba
87
  - Hausa
88
+ - **Format:** Typically stored in CSV or JSON files where each record contains:
89
  - The English sentence.
90
  - The corresponding translations for each target language.
91
  - **Splits:** The dataset is divided according to each low-resource language to mirror Fleurs partitioning but particularly for African Languages.
 
101
  - **Format:** Audio files (e.g., WAV) with accompanying metadata files (e.g., CSV/JSON) that include:
102
  - Unique identifier linking to text entries.
103
  - Language code.
104
+ - Duration, sample rate, and other audio properties.
105
  - **Parallelism:** Each audio file is aligned with the corresponding text in both the source (English) and target languages.
106
 
107
  ---
 
195
  - **Speech Recognition (ASR):** Developing systems that can transcribe speech in under-resourced languages.
196
  - **Text-to-Speech (TTS):** Creating natural-sounding TTS systems using paired audio-text data.
197
  - **Cross-lingual Learning:** Supporting transfer learning and multilingual model training.
198
+ - **Language Identification (LI):** Identifying spoken or written languages (speech or text).
199
 
200
  ---
201