-
-
I found in the pre-trained datasets, there are some docs has large amount chars, which cause a long time to encode them. For example, a doc has 15955671 chars, will cost 6.6 hours to encode it.
Ho…
-
**Describe the bug**
Error converting mistral to onnx
**Expected behavior**
```
!pip install virtualenv
!virtualenv myenv
!source /content/myenv/bin/activate
!git clone https://github.com/n…
-
Hi, I am currently testing with `TinyLlama/TinyLlama-1.1B-Chat-v0.3` model on NVIDIA Tesla T4 and the Docker image version is 0.1.0b1. Unfortunately, there is an error when doing inference, and here i…
-
Hi,
Posting here even though this is not related to the code itself.
**Context:**
I have tried to used `Chat-v0.3` directly using the checpoints [code]() I just added `eos_token_id=tokeniz…
-
Dear Authors,
Thanks so much for your amazing project.
Would it be possible for you plan to release the following:
1. the optimizer states
2. the scheduler
3. a checkpoint just before cooling do…
-
非常有意思的工作,但是huggingface 最近总是连接超时,是否可以放一个国内可以下载的链接呢。
-
See https://huggingface.co/PY007/TinyLlama-1.1B-step-50K-105b/blob/main/tokenizer_config.json#L22
-
I adapted TimDettmers filtered Openassistant dataset in order for it to take the Llama 2 prompt format (e.g. with INST), see [here](https://huggingface.co/datasets/Trelis/openassistant-llama-style/).
…
-
Hi, I see the mention of running this model on llama.cpp in README. Did you get a manage to get it to run and quantize with good output? I'm trying to evaluate if this model can be used for speculativ…