-
I am using
`shap==0.38.1`
`torch==1.7.1`
I have a MLP that takes a text input (preprocessed/encoded as LongTensor) and the model trains fine for the classification problem at hand. I use the Long…
-
### Prerequisites
- [X] I have read the [documentation](https://hf.co/docs/autotrain).
- [X] I have checked other issues for similar problems.
### Backend
Colab
### Interface Used
CLI
### CLI Co…
-
# 在arm64 windows 电脑上没有识别出结果
脚本:
```bash
@echo off
setlocal
:: 指定模型路径
set MODEL_PATH=C:\\Users\\Admin123\\.cache\\modelscope\\hub\\lovemefan\\SenseVoiceGGUF\\gguf-fp16-sense-voice-small.bin
::…
-
### What happened?
I encountered an issue while loading a custom model in llama.cpp after converting it from PyTorch to GGUF format. Although the model was able to run inference successfully in PyTor…
-
I tried to train matmulfreellm model following "https://github.com/ridgerchu/matmulfreellm/issues/9#issuecomment-2193970930". But I kept faced " cuda device-side assert triggered". Could you give me m…
-
All requests end with 'finish_reason': 'length' when the max_tokens=-1 parameter is set.
What could be the problem?
**Model**:
https://huggingface.co/IlyaGusev/saiga_mistral_7b_gguf/resolve/main/…
-
The model configuration is not the same as described in the paper. There is a softmax layer missing at the end of the model. The paper concatenates the attention * vision features for all the glimpses…
-
### What happened?
For llama cpp I had downloaded the q4_k_m quantized [model](https://huggingface.co/jxtngx/Meta-Llama-3.2-1B-Q4_K_M-GGUF/tree/main) and used [llama-bench](https://github.com…
-
https://github.com/codertimo/BERT-pytorch/blob/d10dc4f9d5a6f2ca74380f62039526eb7277c671/bert_pytorch/model/attention/multi_head.py#L15
Looks that **self.d_k = d_model // h ---> embed size 768 divi…
-
I am trying to integrate SAM.The model is getting loaded but when I try to annotate. I am getting this error here. I will really appreciate any help on this
(mask_decoder): MaskDecoder(
(transf…