-
TensorRT-LLM version:v0.13.0
GPU:A100
Convert script:
```
python3 convert_checkpoint.py \
--model_dir /model/Qwen2.5-14B-Instruct \
--output_dir /model…
-
**Describe the bug**
ttnn.tril gives low PCC when using sharded strategies and when input tenor (of rank 3 or 4) has second to innermost dimension not divisible by 32. Problem has been observed on Wor…
-
# Brief describe
add shape mismatch for this input variation
```
"Tensor self = ?", "Tensor other = 0.0"
```
Run [`pytest tests/pattern/test_add_shape_mismatch.py`](https://github.com/tenstorrent/pyt…
-
# Bug Report
### Describe the bug
As described in issue https://github.com/onnx/onnx/issues/5880, due to shape infer implementation, for large models, the dim info of input or output is someti…
-
Traceback (most recent call last):
File "/home/lpl/muavic/demo/run_demo.py", line 220, in
AV_RESOURCES = load_av_models(args.av_models_path)
File "/home/lpl/muavic/demo/demo_utils.py", lin…
-
Hi VideoLLaMA Team,
I am facing issues while loading all the base models in 4-bit precision. The following lines try to load the ```mm_projector_weights``` which are stored in 16-bit precision into…
-
### 🐛 Describe the bug
Here, I came across this issue with MAISI network from MONAI.
To reproduce, you would need to pull branches from :
https://github.com/Project-MONAI/MONAI/pull/8153 and…
-
### Checklist
- [ ] The issue exists after disabling all extensions
- [ ] The issue exists on a clean installation of webui
- [ ] The issue is caused by an extension, but I believe it is caused by a …
kkget updated
3 weeks ago
-
Hi, thanks for your excellent work! I'm trying to run `bash eval_calvin.sh`.
When running to FeedbackPolicy/models/policy.py, there is an issue where the shape of the vision_x input to vision_encoder…
-
### Describe the issue
I am trying to load XGBoost onnx models using onnxruntime on Windows machine.
The model size is 52 mb and the RAM it is consuming on loading is 1378.9 MB. The time to load …