-
## Description
## Environment
**TensorRT Version**: 8.5
**NVIDIA GPU**: 2060
**NVIDIA Driver Version**: 515
**CUDA Version**: 11.6
**CUDNN Version**: 8.8
**Operating System**: ubantu …
-
Hello.
Steps to install:
```
git clone https://github.com/bmaltais/kohya_ss.git
cd kohya_ss
git checkout v21.5.11
./setup.sh
```
Output when I do `accelerate config`:
```
$ . venv/bi…
-
## Description
According to the tutorial, in order to quantify more layers and get int8-in-int8-out, I need to follow the conv with a Q, but if the relu is followed by a Pad, what should I do?
F…
-
Hello
error when in Processing Images with TensorRT. Please see error below.
306lab:~/jetson-inference/python/training/imagenet$ imagenet-console.py --model=cat_dog/resnet18.onnx --input_blob=in…
-
@chensuyue I have tried the case in `examples/onnxrt/image_recognition/mobilenet_v2/quantization/ptg_static`, but can not reproduce the results shown in https://intel.github.io/neural-compressor/lates…
-
My goal is to convert a Torch model trained via mmdetection to ONNX. I first followed the [guide](https://mmdeploy.readthedocs.io/en/latest/build/linux.html) to build toolchains, dependencies, and MMD…
-
@mmaaz60 Can you please provide the code for converting this model to int8? It is successfully converting to ONNX and I am able to infer using ONNXRuntime. However, is there any way to decrease the in…
-
### Search before asking
- [X] I have searched the YOLOv5 [issues](https://github.com/ultralytics/yolov5/issues) and found no similar bug report.
### YOLOv5 Component
_No response_
### Bug
When …
-
Meta just released there LLaMA model family
https://github.com/facebookresearch/llama
Can we got support for that?
They calim that the 13B model is better than GPT 3 175B model
----------------…
-
### bug描述 Describe the Bug
### 错误信息
错误引入 PR:https://github.com/PaddlePaddle/Paddle/pull/50915
case 地址:https://github.com/PaddlePaddle/PaddleTest/tree/develop/inference/python_api_test/test_nlp_…