You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I am trying to train the titanet-large model with a huge dataset (10M+ samples) and it is taking more than 20 hours (still running) to generate the manifest.json file for trainig data. i run: python NeMo/scripts/speaker_tasks/filelist_to_manifest.py --filelist path/to/train.txt --id -2 --out path/to/manifest.json --split
Is there any possibility to make it faster? or to skip this step and be able to train the model with the files list .txt?
The text was updated successfully, but these errors were encountered:
Hi,
I am trying to train the titanet-large model with a huge dataset (10M+ samples) and it is taking more than 20 hours (still running) to generate the manifest.json file for trainig data. i run:
python NeMo/scripts/speaker_tasks/filelist_to_manifest.py --filelist path/to/train.txt --id -2 --out path/to/manifest.json --split
Is there any possibility to make it faster? or to skip this step and be able to train the model with the files list .txt?
The text was updated successfully, but these errors were encountered: