-
### Your current environment
```text
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.4 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubunt…
-
fast_inference api推理会把参考声音融入进去是怎么回事,有没有好心的大佬知道这个问题
-
### Description
The kibana team has requested that we add pagination and sorting options to the`GET _inference/_all` API to efficiently handle these operations in the backend. Currently, they have ad…
-
### Description
Investigate caching to improve performance especially when generating text embeddings.
-
Many packages, such as Dictionaries.jl or StaticArrays.jl, seemingly need access to Julia's type inference functionality to be able to offer their users convenient construction of their custom data st…
-
### Description
Improve validation and exception handling within the inference API.
Here are a few areas to get started
- When a text embedding service is created, during the creation process w…
-
## ❓ Questions and Help
When I made multiprocess inference of huggingface transformers frame, I used xmp.spawn(perform_inference, args=(args,), nprocs=4), and I wanted to run 4 scripts once. However,…
-
A test failed on a tracked branch
```
Error: Expected status code 200, got 500 with body '{"error":{"root_cause":[{"type":"status_exception","reason":"Model importing failed due to an Exception [java…
-
## Bug Description
### To Reproduce
Steps to reproduce the behavior:
enable pipeline params:
pipeline_enabled: true
pipeline_params:
_reduce_whitespace: true
_run_ml_inference: …
-
### Discussed in https://github.com/beartype/beartype/discussions/402
Originally posted by **Moosems** June 30, 2024
I want to have beartype print out the proper type annotation for a variable…