-
Are you interested in learning more about GPU programming and developing cool optimizations? Do you want to help build next generation and state-of-the-art machine learning models and layers? Do you w…
-
I have gone through the example: opensearch-py-ml/examples/demo_deploy_cliptextmodel.html
Model is correctly registered in opensearch cluster but the final command of the example:
ml_client.depl…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
### System Info
the offical docker env in docker/Dockerfile.multi of commit "c629546"
### Who can help?
@byshiue @ncomly-nvidia
I try to convert deepseek-v2-lite
```
python convert_checkpoint.py …
-
**Describe the bug**
I try to use LLaVA example and faced to key mismatch error. I am on latest commit in main branch. (094d66b)
[rank0]: RuntimeError: Error(s) in loading state_dict for LLaVAMode…
-
Opening this to add support for all models following #34282
Lets bring support for flex attention to more models! 🤗
- [x] Gemma2
It would be great to add the support for more architectures s…
-
### System Info
transformers==4.45.2
when preparing the cross_attention_mask in ```_prepare_cross_attention_mask``` function we get the``` cross_attn_mask``` to the shape of [batch,text_tokens,i…
-
![image](https://github.com/user-attachments/assets/3bc230bc-5029-4657-b107-0f1a1b54be15)
Error:
`Phi3Transformer does not support an attention implementation through torch.nn.functional.scaled_do…
-
## 🐛 Bug
I get the following assertion error from Thunder JIT:
```py
File ~/dev/lightning-thunder/thunder/core/jit_ext.py:1731, in thunder_general_jit(fn, args, kwargs, record_history, sharp_edges,…
-
output = self.model(sequences, attention_mask=attention_mask, position_ids=position_ids)
File "/root/miniconda3/envs/OpenRLHF/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553…