-
I am writing to ask for your help with a problem I am having with the tokenizer. I have been trying to solve it for a while now, but I have been unsuccessful.
However, I am having trouble with : Trac…
-
-
During the process of fine-tuning LLama3 using LLama.cpp on my Mac, I encountered this error. I'm a beginner and don't know what caused this issue. I hope an expert can help me.
The model used is: …
-
Hi. I have a phoneme-based Zipformer model.
Before this [PR](https://github.com/k2-fsa/sherpa-onnx/pull/828), I was able to apply hotwords encoding for phoneme sequences, e.g. `ɪ z/dʒ ʌ s t/b ɛ s t…
-
Background: Using the same gguf model with the same parameters and inputs, using -- top k=1 (greedy strategy);
llamafile-0.8.6 llamacpp-b2249
When generating the first token, the distribution of lo…
-
#### Problem description
A gensim model was trained under Python 2.7 with a **chinese** dataset.
However, now we are using Python3.6, and we got some broken strings in .vocab.keys() as title.
…
-
Xlsx 类型的文件要下载什么插件呀
-
Loading data...
Vocab size: 4762
491it [00:00, 98460.66it/s]
40it [00:00, ?it/s]
42it [00:00, 42113.50it/s]
Traceback (most recent call last):
File "C:\Users\dell\Desktop\Chinese-Text-Classifi…
-
Note that the issue tracker is NOT the place for general support.
I deployed a model, but encountered the problem of garbled Chinese, what is the reason?
such as:
this is my model.json
```js…
-
I want to use the Megatron framework for Chinese NLP pre-training tasks. Currently, I have Chinese corpus resources and a vocab.txt file. However, for most frameworks, it seems that vocab.json and mer…