-
### System Info
I am currently mostly working with the ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 Docker image on MacOS. Currently, I am only trying to find out which reranker models with …
-
When migrating from automatic1111 to diffusers, I'm experiencing a significant degradation in image quality despite using the same parameters. The images generated with diffusers are of noticeably low…
-
Hello, I have no idea what I'm doing. I'm trying to run on an ARM64 system, and I get the following errors:
I ran the loadgen install and everything worked fine, running the vision benchmarks, fai…
-
### Describe the issue
Hey
We are planning to add GPU inference (using Mirosoft.ML.OnnxRuntime.Gpu 1.17.0) as an option in our C# software.
However, when switching from the CPU ONNX runtime to th…
-
Hi, Thanks for such a robust work!
We have supported ArenaHard dataset in Opencompass now, OpenCompass is an evaluation platform that can partition tasks and support different model inference backend…
-
### Describe the issue
Im using A1111 and an extension to mask the background. When i try to run the generation to get the mask, i run into some issues. Since there is no install instructions anywher…
-
Hi! Very impressive project!
My main goal is to export the model to intermediate format and test accelerability on many platforms. I am trying to accelerate the assembled convolution module for be…
-
Hello,
I am seeking advice on the best practices for tracking all inputs and predictions made by a model when using Triton Inference Server. Specifically, I would like to track every interaction th…
-
🦥 Unsloth: Will patch your computer to enable 2x faster free finetuning.
==((====))== Unsloth: Fast Llama patching release 2024.6
\\ /| GPU: NVIDIA A100 80GB PCIe MIG 7g.80gb. Max memory: 7…
-
### Describe the issue
I designed and trained a 6D pose estimation algorithm model using pytorch. After that I use torch.onnx.export to convert the pth format parameter file into an onnx inference fi…
W-QY updated
2 months ago