Open iver56 opened 3 years ago
I am curious - how do you plan on using micaugment? Is it the idea that we can submit a trained micaugment model as part of the transform?
I haven't thought it through, but yeah, we should ideally have a pretrained model that is ready to be used. The model can be uploaded as a binary in a github release, and can be downloaded and stored locally on demand (the first time the transform gets used). This approach is inspired by the way Keras did pretrained imagenet models.
Would https://github.com/akashrajkn/micaugment be suitable?
I think it is suitable - however, I still have to update the repo with a pretrained model.
It would be awesome if you could make that happen 🤩 But I guess the pretrained model would depend on a specific sample rate, right? Ideally, torch-audiomentations should be compatible with a wide range of sample rates 🤔 Maybe it could do some resampling to match the sample rate used in the model
À la https://arxiv.org/abs/2010.09658