Open ApolloRay opened 7 months ago
python demo_txt2img_xl.py "a photo of an astronaut riding a horse on mars" --version xl-turbo --onnx-dir ./dreamshaper_model/dreamshaper_onnx/ --engine-dir engine-sdxl-turbo --height 1024 --width 1024 --int8
Description: run with this code in A10 (23G), it will show OOM.
if i use width = 512 and height = 512 , it can run. But the inference time unet-int8 ~ 300ms , unet - fp16 ~ 250ms.
@rajeevsrao @azhurkevich Is it expected? Thanks!
same problem as #3724
@rajeevsrao @azhurkevich Is it expected? Thanks!
Can anyone help?
@ApolloRay maybe you can follow this blog post
@ApolloRay maybe you can follow this blog post
from utils import load_calib_prompts I can't find any info about utils.
Description
Environment
TensorRT Version:8.6
NVIDIA GPU:A10
NVIDIA Driver Version:525.147.05
CUDA Version:12.0
CUDNN Version:8.9
Operating System:
Python Version (if applicable):
Tensorflow Version (if applicable):
PyTorch Version (if applicable):
Baremetal or Container (if so, version):
Relevant Files
Model link: dreamshaper(turbo version)
Steps To Reproduce
Commands or scripts:
Have you tried the latest release?:
Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (
polygraphy run <model.onnx> --onnxrt
):