-
### System Info
Using Windows 11
### Information
- [X] The official example scripts
- [X] My own modified scripts
### 🐛 Describe the bug
When running:
```
docker run -it -p 5000:5000 -v C:/Us…
-
## Bug Description
I'm trying to serve torch-tensorrt optimized model to Nvidia Triton server based on the provided tutorial
https://pytorch.org/TensorRT/tutorials/serving_torch_tensorrt_with_t…
-
In `reverie/backend_server/persona/prompt_template/run_gpt_prompt.py`, multiple requests to OpenAI are made with a hardcoded model `gpt-35-turbo-0125`, which is currently not a valid/supported model o…
-
Hello, I hope you are doing well.
I intend to run resnet50 in the server scenario (datacenter) using the script in the docs:
```
cm run script --tags=run-mlperf,inference,_r4.1-dev \
--mode…
-
Envoy supports sending the full request body to the external authorization server via the with_request_body filter configuration. Do you think that it is possible to expose such feature on the Securit…
-
Here is a trace from my Intel Arc A770 via Docker:
```
$ ollama run deepseek-coder-v2
>>> write fizzbuzz
"""""""""""""""""""""""""""""""
```
And here is an trace from Arch linux running on …
-
### Your current environment
The output of `python collect_env.py`
```text
Collecting environment information...
PyTorch version: 2.4.0+cu121
Is debug build: False
CUDA used to build PyTorch…
-
### What is the issue?
No issues with any model that fits into a single 3090 but seems to run out of memory when trying to distribute to the second 3090.
```
INFO [wmain] starting c++ runner | ti…
-
### What's the use case?
I have a job with a dynamic graph, using DynamicOut.
The ops are configured with Pydantic configs, allowing us to parametrize the ops in the launchpad.
In each Op, we…
-
### Self Checks
- [X] This is only for bug report, if you would like to ask a question, please head to [Discussions](https://github.com/langgenius/dify/discussions/categories/general).
- [X] I have s…