-
(Apologies if this is already supported, the docs are unclear/confusing)
DJL-serving claims support for TensorRT model on https://github.com/deepjavalibrary/djl-serving
DJL FAQ doesn't mention s…
-
多输入为有不同的shape,2/3/4维都有,请问这个QAT范式支持吗?如果支持的话,在保存模型前的一次前向推理应该怎么写呢?如官方给出的example:
predict = net.forward(F.placeholder([1, 3, 224, 224], F.NC4HW4))
那么请问我这个多输入的应该怎么写吗?我尝试按照net的输入顺序一次填入,执行失败了(没有报错),结果出现很…
-
### Description
Hello, I am trying to inference detection model using 2 tpu's using model pipelining. I have google coral dev board and google coral usb accelerator connected to coral devboard. I am …
-
### Before Asking
- [X] I have read the [README](https://github.com/meituan/YOLOv6/blob/main/README.md) carefully. 我已经仔细阅读了README上的操作指引。
- [X] I want to train my custom dataset, and I have read the …
-
Keyerrors which occur in ptq process mainly due to that the head or tail of a block (sub-graph) is not a `torch.nn.Module`.
Will fix it next week.
-
Hi
I was successd in PTQ but failed with fast-finetune
the code is like below:
```
quant_mode = 'calib'
quantizer = torch_quantizer(quant_mode=quant_mode,
…
-
Hi,
I am trying to export the torch-native PTQ model using the timm repo. I am working on exporting the deit_tiny_patch_16 model.
The model conversion is however failing with runtime error:
Ru…
-
### 请提出你的问题
**运行环境**
win10,GPU
paddlenlp 2.5.2
paddlepaddle-gpu 2.4.2
**执行命令**
`python finetune.py --device gpu --logging_steps 1 --save_steps 100 --eval_steps 100 --seed 42 --mod…
-
-
环境:
Ubuntu18.04,cuda10.1,
Torch==1.8.1+cu101, Torchvision==0.9.1+cu101,
onnx==1.11.0,onnx-simplifier==0.3.10,onnxoptimizer==0.2.7,onnxruntime==1.11.1
安装:
使用的MQBench是最新的main分支代码。安装过程是按照文档的说明进行安装…