-
i sent your mail for the access datasets and AI inference application before 2 weeks ago.
When can you give me the answer?
-
### The Feature
While not documented, `seed` is "supported" now. https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/inference#parameter-list
Similar to #2962. Related to https:/…
-
## Overview
Currently, users can specify the `num_inference_steps` parameter in the T2I (Text-to-Image) pipeline (see the implementation [here](https://github.com/livepeer/ai-worker/blob/0a26654ccc…
-
TBD
-
Since jetson supports triton inference server, I am considering applying it.
So, I have a few questions.
1. In an environment where multiple AI models are run in Jetson, is there any advantage to …
-
**Problem**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
Jan only supports openai compatible api. So using https://github.com/zhu327/gemini-openai-pro…
-
We are currently using a _pulling update_ mechanism to get the health check.
https://github.com/containers/podman-desktop-extension-ai-lab/blob/529bc5bef181032081fb5a616c0de7afabd27c4e/packages/bac…
-
[generative-models/blob/fbdc58cab9f4ee2be7a5e1f2e2787ecd9311942f/README.md#installation](https://github.com/Stability-AI/generative-models/blob/fbdc58cab9f4ee2be7a5e1f2e2787ecd9311942f/README.md#insta…
-
Hi @hopef !
I found a huge gap of inference latency on orin of BEVFusion's lidar-scn backbone, between using **the engine returned by spconv::load_engine_from_onnx(30ms)** and using **the original …
-
We're exploring various optimizations available in the [Diffusers library](https://huggingface.co/docs/diffusers/main/en/optimization/opt_overview) to enhance VRAM usage and inference speed. @titan-no…