-
-
-
Objective: TriagerX is a novel AI-enabled software analytics tool that we developed via the IBM CAS project (with Dr. Uddin). TriagerX aims to assign an issue to components/teams and developers and to…
-
### Your current environment
The output of `python collect_env.py`
```text
Collecting environment information...
PyTorch version: 2.5.1+cpu
Is debug build: False
CUDA used to build PyTorch…
-
It is recommended to use Instance profile instead of Accesskey and Secret Key to configure Amazon Bedrock as served entity in model serving endpoint.
### Use-cases
### Attempted Solutions
…
-
I am currently using the LLM Serving Sim paper to see how this simulator works, and I am using it to figure out the results of the simulator according to the settings.
However, the following error …
-
Hi,
Could you please help provide guide on integrating deepspeed approach of using multi-GPU Intel Flex 140 to run model inference using FastAPI and uvicorn setting ?
model id: 'meta-llama/Llama-2-7…
-
Environment:
Platform: 6548N+4ARC770
Docker Image: intelanalytics/ipex-llm-serving-xpu:2.1.0
servicing script:
![image](https://github.com/user-attachments/assets/3949f088-d83f-4844-9ab3-0f0c986…
-
How can we create model endpoints using Modelmesh on the cluster?
This should involve instructions on deploying modelmesh using Operator Hub and adding the model to a s3 bucket.
Create the predi…
-
Explore [text-generation-webui](https://github.com/oobabooga/text-generation-webui/) to serve language models on ODH.
How flexible is it for different models?
Would we need a custom server.py f…