-
Hi! I tried evaluating 1bitLLM/bitnet_b1_58-3B from hugging face. i am getting the error ValueError: Tokenizer class BitnetTokenizer does not exist or is not currently imported.
Kindly help!
-
Hi there,
I was trying to import the weights from `gemma_2b` to transformers `GemmaForCausalLM` but it seems to be missing the `lm_head.weight` key in the state dict. Code below:
```python
fro…
-
It will be wonderfull if there were Helper functions for external retrieval like:
- lm.get_URL(path) for using an URL
- lm_get_MD(path) for using Markdown document
- lm_get_PDF(path) for using P…
-
one very slow kernel in train_gpt2.
`BEAM=2 DEBUG=2 python3 test/external/external_test_lm_head.py` gives 136ms. changing vocab_size to 50304, the kernel became 20ms. fixing this in optimization ca…
-
### System information
Type | Version/Name
--- | ---
Distribution Name | Windows 10 22H2 19045.4474
Architecture | x86_64
OpenZFS Version | zfs-2.2.99-149-gf9ec771f74
### Describe the pr…
-
(pytorch2.0.1zh) zh@inspur:~/TrOCR$ python -m transformers.onnx --model ./cust-data/weights/ --feature=vision2seq-lm --framework=pt --preprocessor=tokenizer --atol 1e-4 hand-write-onnx
usage: Hugging…
-
An empty json object is created when using the following code (from [lm_analysis_plots.R](https://github.com/HARPgroup/HARParchive/blob/master/HARP-2024-2025/functions/lm_analysis_plots.R) )
```
dat…
-
lm-studio is an open source tool for running LLMs locally. After having done this by hand (as the plugin is doing), and trying other tools to do this, it's clear (to me) that lm-studio is going to be…
-
I am trying to eval the finetuned model 70B with torch run and getting error
Here is my config file
```
model:
_component_: torchtune.models.llama3.lora_llama3_70b
lora_attn_modules: ['q_pr…
-
I repeatedly confirmed that the mlx and mlx_lm modules have been installed, but tinychat reported an error when adjusting the model during the conversation. The specific error is as follows:
resp = …