rh-aiservices-bu / llm-on-openshift

Resources, demos, recipes,... to work with LLMs on OpenShift with OpenShift AI or Open Data Hub.
Apache License 2.0
90 stars 86 forks source link

LLM on OpenShift

In this repo you will find resources, demos, recipes... to work with LLMs on OpenShift with OpenShift AI or Open Data Hub.

Content

Inference Servers

The following Inference Servers for LLMs can be deployed standalone on OpenShift:

Serving Runtimes deployment

The following Runtimes can be imported in the Single-Model Serving stack of Open Data Hub or OpenShift AI.

Vector Databases

The following Databases can be used as a Vector Store for Retrieval Augmented Generation (RAG) applications:

Inference and application examples

LLM clients