Closed ylwu-amzn closed 3 months ago
Does this mean models deployed inside OpenSearch ml-common?
Does this mean models deployed inside OpenSearch ml-common?
Yes, it is going to be released with 2.15. Related PR: https://github.com/opensearch-project/ml-commons/pull/2508
Great! Now is it possible to deploy a small language model such as microsoft/Phi-3-mini-4k-instruct-onnx on OpenSearch and use this feature to do question answering inferencing within OpenSearch? If yes, then would you please share the steps?
@asfoorial , no, this feature is to support using current supported local models in ML inference processor. SLM is not a supported local model type. So you can't deploy it inside OpenSearch cluster. You can deploy it somewhere else and use connector to connect. One update in 2.15 is we support private IP setting https://github.com/opensearch-project/ml-commons/pull/2534
We have released ML inference processor in 2.14 which only supports remote model. We are going to enhance it to support local model too.