-
# ❓ Questions & Help
HI, I can not save model and vocab when using spm.SentencePieceTrainer.Train.
## Details
This is my config:
python3 ./openspeech_cli/hydra_train.py dataset=librispeech…
-
Extract the English subset in the Multilingual Librispeech dataset (https://huggingface.co/datasets/facebook/multilingual_librispeech). The resulting data must look like this:
```
{
'audio': the…
-
Extract the French subset in the Multilingual Librispeech dataset (https://huggingface.co/datasets/facebook/multilingual_librispeech). The resulting data must look like this:
```
{
'audio': the …
-
The goal is to discretize the speech data from the French Librispeech dataset you previously worked on (@abheesht17 & Adithiya): https://huggingface.co/datasets/abheesht/librispeech_fr. Remember that…
-
```
from datasets import load_dataset
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
import soundfile as sf
import torch
from jiwer import wer
librispeech_eval = load_dataset("li…
-
https://www.openslr.org/12
-
Hello and thanks for your work and sharing the pretrained models
1. I noticed that all the models shared have `2spk` in their names, does that mean that they only support two speakers?
2. If I want …
-
### System Info
DGX V100 and DGX A100
### Who can help?
@ncomly-nvidia to add more folks.
### Information
- [ ] The official example scripts
- [ ] My own modified scripts
### Tasks
- [ ] An of…
-
Hi, thank you for the amazing work. May I ask two questions about Table 1?
1. Could you please provide more detailed descriptions about the ```baseline acoustic codec``` in your paper? Does it
- ex…
-
# Task Name
Audio Spatial Distance Prediction
## Task Objective
Audio Spatial Distance Prediction is a task that aims to predict spatial distance from the source of the sound based on the giv…