bentoml / BentoML

The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!
https://bentoml.com
Apache License 2.0
7.17k stars 792 forks source link

docs: Add explanations for model loading acceleration #5066

Closed Sherlock113 closed 2 weeks ago

Sherlock113 commented 2 weeks ago

What does this PR address?

Fixes #(issue)

Before submitting: