-
## Steps to reproduce
**Important note:** this happens in a very specific client environment, simply reproducing this deployment in a LXD cloud will not be sufficient to get the error. I am p…
-
### ClearML serving design document v2.0
**Goal: Create a simple interface to serve multiple models with scalable serving engines on top of Kubernetes**
Design Diagram (edit [here](https://excalid…
-
## Description
When I package a kedro pipeline for serving with modelify my kedro hooks no longer work.
## Context
I was using hooks for initialization of some structures and had to make work…
-
The GCP integration test for this accelerator is disabled because the model serving features used in notebook 4 are not yet available on GCP.
-
/kind feature
**Describe the solution you'd like**
[A clear and concise description of what you want to happen.]
Currently, when an InferenceService is deployed without CPU or Memory resource l…
-
OS type
Ubuntu
Description
When running the example Translation using Docker Compose, one of the images takes additional time to pull a model from the Huggingface upon startup. During this period…
-
hello I got the .har by google chrome and placed it in the corresponding folder. But when I try to work with the chatgpt.com provider, it doesn't work. log:
Read .har file: ./har_and_cookies/chatgpt…
-
/kind bug
cc: @gilcardoai
**What steps did you take and what happened:**
Upgraded from v0.10 to v0.11 to prepare for a cluster upgrade.
Spun up a new InferenceService to get the associated v…
-
-
Infrastructure for serving the trained model with RESTful API.
1.) API gateway for APP to upload a image
2.) pass the image to get prediction out of the trained model
3.) Infrastructure for servi…