-
Is this out of scope? I hope not, would be nice to have a one-stop shop for interpretability tooling.
### Proposal
It should be easy to get the most bare-bones interpretability research off the…
-
### Feature request
Currently, if fp16 is used with grounding dino via https://huggingface.co/docs/transformers/main/en/model_doc/grounding-dino, there is an error of the following:
```
...
Fi…
-
-
# Description
When attempting to set up llama cpp python for GPU support using CUDA toolkit, following the documented steps, the initialization of the llama-cpp model fails with an access violation…
-
I'm trying to use the DH benchmark from this year's OAEI.
I get the error below. Do you have any idea what is going wrong? I also included the `config.json` and `configMatcher.json`. To test I only …
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ ] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
## Environment
- Platform: Debian Linux
- GPU: A100
- Torch: '2.1.2+cu121'
- Transfomers: '4.37.2'
## Issue
I'm seeing random and sudden loss spikes during training, if there is a simpler wa…
-
## 🐛 Bug: Opening a bug. But really, it is a question. In the NACRF implementation, (specifically the Fast Structured Decoding for Sequence Models paper implementation), I do not see the Multi-head po…
-
I'd like to get a summary for the following BERT model
```
model = torch.hub.load('huggingface/pytorch-transformers', 'model', 'bert-base-uncased')
```
What shape or shapes should I use? how to …
-
Hi guys,
I am following the Megatron-LM example to pre-train a BERT model but I'm getting this error:
```
[rank0]: Traceback (most recent call last):
[rank0]: File "/root/Megatron-LM/pretrai…