-
When recording in real time, for example, I specified the output voice to be English, and the audio played was a mixture of Japanese and English, sometimes translating Japanese into English and someti…
Ye83 updated
6 months ago
-
A typical example where stress is important in Japanese is 橋 (bridge, read as はし, hashi) vs 箸 (chopsticks, read also as はし, hashi). The former (橋) has stress in the second syllable, where the latter (…
-
I have downloaded the Faster Whisper ASR models in the tools/asr/models/ folder but still fail to load them .. only encounter with the resource_tracker warning and not going further. I have tried v2…
-
### Describe the proposed feature and its relevance to inferencing
Document how to take a random PyTorch model and make it run in Godot Engine IREE.gd
### Context or use case
1. https://githu…
-
**Is your feature request related to a problem? Please describe.**
The current logic of misspelling identification relies on `vocab.txt` from the transformer model. BERT tokenisers break not such com…
R1j1t updated
8 months ago
-
I only see examples pertaining to transcription, but can't faster-whisper also be used for translation and include parameters such as beam size, specifying the language translating from, etc., all the…
-
TL;DR:
* Encountering frequent NaN values mainly for the Loss, during training with [a large JPN dataset ](https://huggingface.co/datasets/oshizo/japanese-wikipedia-paragraphs)(10.5 million rows).
…
-
whisper supports outputting multiple file formats, currently `txt,vtt,srt,tsv,json,all`, does the project have the ability to output multiple formats? Here is the whisper command line help, you can sp…
-
Here are a few points regarding the Readium CSS integration which should be double-checked or fixed in the mobile toolkits.
* [`ReadiumCSS-default.css` needs to be injected](https://github.com/read…
-
Is there support for non-latin languages like Chinese, Japanese or Thai?