-
in the train stage,it still fail , File "/opt/whisper/whisper-at/src/whisper_at_train/models.py", line 172, in forward
audio_rep = audio_rep.reshape(Bself.n_layer, audio_rep.shape[2], audio_rep.shap…
-
can support real time asr?
-
### System Info
(h2ogpt) jon@pseudotensor:~/h2ogpt$ pip freeze | grep langchain
langchain==0.0.235
langchainplus-sdk==0.0.20
Python 3.10
(h2ogpt) jon@pseudotensor:~/h2ogpt$ lsb_release -a
No…
-
Hello @kan-bayashi I can see gan_tts task now work with joint TTS training but wondering if you have example for specify pretrained models, switching vocoder impl and expected training time difference…
-
## ReazonSpeech v1.1モデルのリリース
### ゴール
* 2023年春頃を目標に`reazonspeech-espnet-v1.1` をリリースする。
* 音声認識の精度・ロバストネスを向上させ、さらに長時間音声の文字起こしに対応する。
### v1からの改善点
1. ノイズに対するロバストネスを向上させる #4
2. 複数の発話が含まれる音声に…
-
Hello,
I am using VOSK for my development and it works wonderfully on my local machine. I need to start building my pipeline, and I was looking at using GCS to host my data, workflow and app.
I…
-
Hi @csukuangfj I tried sherpa with mgb2 streaming transducer from here https://github.com/k2-fsa/icefall/tree/master/egs/mgb2/ASR/pruned_transducer_stateless5. I used https://github.com/k2-fsa/icefal…
-
### System Info
Copy-and-paste the text below in your GitHub issue and FILL OUT the two last points.
- `transformers` version: 4.21.1
- Platform: Linux-4.15.0-177-generic-x86_64-with-glibc2.27
…
-
I want to retrain the ASR and F0 models with my own data set. Can I provide the code, or are there any precautions when building the training code?
-
Hi,
I've been testing the NeMo transcription engine for some time now and I came across a question. In the case of voice with background music, how would NeMo behave?
Would I need to train with au…