-
Notifications
You must be signed in to change notification settings - Fork 59
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cant start training #3
Comments
What kind of cluster environments are you using? @yubouf, I strongly recommend to add more documents about |
i am using conda environment. and using local machine so have changed to run.pl in cmd.sh |
Oh, I see. |
after exporting CUDA_VISIBLE_DEVICES=1 here is the log
|
|
looks like training started but stopped
log:
|
The losses of the two epochs look good. What does |
here it is
|
where are the hyper parameter of model? maybe reducing the batch size would help |
See |
Thanks for the help. I really appreciate your quick reply. @yubouf After reducing the batch size training completed with 29% DER.
Also there is diarization_data with mix audio what's that for? I think I am missing something. Can you spread some light on what should be dataset format and structure for speaker diarization. |
Both. The latest network configuration is based on 'End-to-End Neural Speaker Diarization with Self-attention'.
"mini_librispeech" model is prepared just for the code integration tests, not related to the papers. I'm afraid the current code is not intended for the inference-only purpose. EEND/egs/mini_librispeech/v1/run.sh Lines 106 to 117 in 9a0f211
train_clean_2 and dev_clean_2 are not actual training and test data for our model. |
In
in
same goes with utt2spk
please elaborate where i am wrong and whats actually in those files. Thanks! |
Explanation of Kaldi's data directory: To know how we generate the simulated training data, see |
I already have audio recordings so no need to simulate but do I need to get the transcript? |
No. You don’t have to prepare the text file. |
Thanks for the links.
Is tbeg(2.82) and tdur(4.27) randomly generated here as I couldn't find difference after hearing the mix audio file. Same goes with found in Lastly in spk2utt and utt2spk : which require Cheers! |
Yes, the training data is the simulated two-speaker mixture of "mini_librespeech" utterances with randomly chosen silence intervals. Suppose you already have your two-speaker mixtures for training data:
Then, you can generate |
hi, i got all the files and started training but nothing is happening. There is nothing inside |
train log
|
The log indicates that
I have no idea about those lines. |
before
can you suggest how to debug further |
When you interrupt the program by Ctrl+C, you will find the stack trace and possible cause of the stop. I'm afraid that it's hard to find the problem because it might be related to the data preparation of your data. If you could open the data for me, I could run that for debugging. |
i am getting this results. can you help me with inference
|
Copied my earlier comment.
"mini_librispeech" model is prepared just for the code integration tests, not related to the papers. I'm afraid the current code is not intended for the inference-only purpose. EEND/egs/mini_librispeech/v1/run.sh Lines 106 to 117 in 9a0f211
|
|
ok, can you suggest how much hours of data is needed to build a good speaker diarization system? Also, can we do this without timestamps? As you know getting audio with accurate timestamps is a difficult task. |
We didn't use manual timestamps for simulated mixtures and two-channel recordings. |
@yubouf Could you please reveal the GPU you used, so I can roughly estimate the training time in my case? |
GeForce GTX 1080 Ti. |
Thank you very much |
Is there any way to train multi-speaker recordings with callhome recipe? I get this error when no of speakers are more than two File "/home/sysadmin/EEND/eend/feature.py", line 282, in get_labeledSTFT |
The model should have a fixed number of speakers as in config |
@008karan Hi,I have the same problem as you, could you share your solution? Thank you! |
train.py -c conf/train.yaml data data exp/diarize/model/data.data.trainStarted at Fri Dec 20 12:27:28 IST 2019python version: 3.7.5 (default, Oct 25 2019, 15:51:11) [GCC 7.3.0] |
Have you solved the problem? |
I was testing the setup on mini librispeech data .This is log when I started training
can you guys suggest whats going wrong?
The text was updated successfully, but these errors were encountered: