-
***Under Construction***
The Answer Engine, released in version 0.13, provides a Q&A interface for Tabby's users to interact with the LLM, optionally within the context of a connected repository. T…
-
**Description**
I'm trying to build with my pending core PR with:
```shell
./build.py \
-v \
--no-container-build \
--repo-tag=core:pull/108/head
```
It fails with:
```
Cloning into…
-
### Issues Policy acknowledgement
- [X] I have read and agree to submit bug reports in accordance with the [issues policy](https://www.github.com/mlflow/mlflow/blob/master/ISSUE_POLICY.md)
### W…
-
In plumbing through the Security Assistant Knowledge Base API integration tests (https://github.com/elastic/kibana/pull/192665), I ended up having to add support for a `modelId` override to our API's …
spong updated
1 month ago
-
I followed the [document](https://docs.mlcommons.org/inference/benchmarks/image_classification/resnet50) to inference ResNet50, using MLCommons-Python -> edge -> Tensorflow -> CUDA -> Native
The comm…
-
VLLM 0.6.2 had just released few hours ago, it said no support multi image inference with Qwen2-VL.
I've try it, but it require the newest transformer and automatic install it.
When I start it u…
-
I'm preforming the instruction in Startup example, and
`pip3 install -r ./open_source/deps/requirements_torch_gpu_cuda12.txt` completed successfully, but when I installed maga_transformer-0.2.0+cud…
-
Hi,
when running the tutorial `OnnxRuntimeServerSSDModel.ipynb` I have this response from the server
```python
response = requests.post(inference_url, headers=request_headers, data=request_messa…
-
### System Info
Docker image: nvcr.io/nvidia/tritonserver:24.07-trtllm-python-py3
Device: 8x H100
trt-llm backend: v0.11.0
### Who can help?
@byshiue @schetlur-nv
### Information
- [ ] The off…
-
**Describe the bug**
Getting zombie process exception as already reported for the [sagemaker-inference-toolkit](https://github.com/aws/sagemaker-inference-toolkit/pull/133)
**To reproduce**
Using…