nithinraok commited on
Commit
2326ee7
1 Parent(s): f5d70c1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -241,7 +241,7 @@ The tokenizers for these models were built using the text transcripts of the tra
241
 
242
  The model was trained on 64K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
243
 
244
- The training dataset consists of private subset with 40K hours of English speech plus 25K hours from the following public datasets:
245
 
246
  - Librispeech 960 hours of English speech
247
  - Fisher Corpus
 
241
 
242
  The model was trained on 64K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
243
 
244
+ The training dataset consists of private subset with 40K hours of English speech plus 24K hours from the following public datasets:
245
 
246
  - Librispeech 960 hours of English speech
247
  - Fisher Corpus