-
### Describe the issue
I tried to use CPUExecutionProvider and CUDAExecutionProvider to inference the same single conv node, and turns out the result does not match after 4 decimals. I'm wondering …
-
Hi, guys
I have rented two a800 and chosen the mirror of opensora 1.1 on the cloud platform.
But when I try to run the command below:
**python scripts/inference-long.py configs/opensora-v1-1/…
-
Hello,
First of all, thanks for the Groq platform! I use it as llm backend to create agent. But it often comes up `RateLimit` error. I handled like this:
```
def inference(self, model: str,…
-
### System Info
- `transformers` version: 4.41.1
- Platform: Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.26
- Python version: 3.10.14
- Huggingface_hub version: 0.23.2
- Safetensors …
-
This is coolest model that I saw. Thank you for this prefect work. For using this to different platforms, is there any supports or information to make inference?
-
### :question: Question
I am sorry for back to back question. But this is very important for me.
I previously used retinaNet for detection and i used 2d data but now i have shifted to 3d data.
I am…
-
/kind feature
## **Why you need this feature:**
At the moment, Model serving (via KServe) and Feature Serving (via Feast—the Feature Store) are separate components without any guidance on how to b…
-
### 🐛 Describe the bug
TorchServe version is 0.10.0.
It's my code:
```
def get_inference_stub(address: str, port: Union[str, int]= 7070):
channel = grpc.insecure_channel(address + ':' + str(p…
-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and fou…
-
### Your current environment
```text
The output of `python collect_env.py`
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
…