I'm trying to retrain your InterSubnet. I have some questions about training datasets.
I saw you used interspeech 2020 datasets in your code but the subset of the interspeech 2021 datasets in your paper. Which is correct ?
If you use the subset of interspeech 2021 datasets, what kind of dataset did you use? Fullband or Wideband? Only use clean read_speech or do you use emotional speech and non-English speech?
Thanks
great job!
But I found that when training the cIRM will using drop_band=2, but the validating will not using drop_band? why?and how to fix this problem?
First of all, thank you for your work. I would like to ask if it is possible to provide a pre-trained model to test the optimal performance of this algorithm.