-
[I hit send too soon on this; I'm updating the comment.]
I think the time might have come to create an 's5b' version of the WSJ setup.
WSJ is the oldest setup and the local scripts are not up to t…
-
I noticed that the default Mandarin acoustic model uses [phone groups](https://github.com/MontrealCorpusTools/mfa-models/blob/main/config/acoustic/phone_groups/mandarin_mfa.yaml) to combine tones. Doe…
-
being unable to reopen #169 (because it was closed by a contrib), so i'll reference it here. someone with enough perms may delete this one and re-open #169
Thank you!
-
Unfortunately, I wasn't able to find the transcriptions corresponding to the simulated meeting data. If we want to investigate E2E approaches during the workshop, I think it will be necessary to have …
TCord updated
4 years ago
-
Hi Nickolay,
Here is a scenario, suppose an audio is playing in a user's Andorid device and meanwhile an announcement occurs outside, now microphone is capturing multiple inputs. How can we recognis…
-
Currently we are using a Google speech recognition API to translate speech into text. Idk how private that API is, so I would like to switch it maybe IBM, or something that works offline
-
Seems that the new oneAPI Intel installer places things in different paths than expected by FindBLAS.cmake - after setting everything up (Intel oneAPI 2021.2) I could not make the cmake process work …
-
compile the latest code( e4eb4f6725d836b7915230f54adedfb605379254 )
failed with error:
```
Running sparse-matrix-test ... 0s... SUCCESS sparse-matrix-test
Running numpy-array-test ...c++ -std=…
-
Dear Owner.
We are using the vosk api and websockets package to implement a multi-access service for our self-generated model.
In particular, we are developing an engine that evaluates how well a …
-
From the README (https://github.com/DigitalPhonetics/VoicePAT?tab=readme-ov-file#anonymization) it says that to anonymize my own data I should modify the following fields in the config file:
```
dat…