-
-
Hello, I am processing and training the AISHELL-4 dataset using the command:
python diaper/train.py -c DiaPer/models/10attractors/SC_LibriSpeech_2spk_adapted1-10_finetuneAISHELL4mix/train.yaml,
…
-
Hi,
First, thank you very much for your work on MelHuBERT and for making your checkpoints publicly available. I have been trying to use MelHuBERT models, and I noticed that the mean and standard de…
-
There should be a functionality where instead of having to download entire dataset and train on it, we could download just partial data and use only that for training. And if not, then the documentati…
-
Hi,
I've just ended a training of a conformer using the sentencepiece featurizer on LibriSpeech over 50 epochs.
Here are the results if you want to update your readme:
```
dataset_config:
t…
-
# Task Name
Audio Spatial Distance Prediction
## Task Objective
Audio Spatial Distance Prediction is a task that aims to predict spatial distance from the source of the sound based on the giv…
-
# ❓ Questions & Help
I'm learning how to use LibriSpeech Dataset to train the squeezeformer network.
After 20 epochs training, both the evaluation WER (0.6368) and CER (0.4251) are still very hi…
-
**Binary**: [1.0.1 Linux](https://github.com/MontrealCorpusTools/Montreal-Forced-Aligner/releases/tag/v1.0.1).
**Command**: I was following the [librispeech example](https://montreal-forced-aligner…
-
I am attempting to prepare the Multilingual LibriSpeech (MLS) dataset using the [lhotse.recipes.mls](https://github.com/lhotse-speech/lhotse/blob/master/lhotse/recipes/mls.py):
```bash
lhotse pr…
-
Hi, thank you for the amazing work. May I ask two questions about Table 1?
1. Could you please provide more detailed descriptions about the ```baseline acoustic codec``` in your paper? Does it
- ex…