sayakpaul / ml-deployment-k8s-fastapi

This project shows how to serve an ONNX-optimized image classification model as a web service with FastAPI, Docker, and Kubernetes.
https://medium.com/google-developer-experts/load-testing-tensorflow-serving-and-fastapi-on-gke-411bc14d96b2
Apache License 2.0
198 stars 36 forks source link

Set up another API similarly with the pure TensorFlow model for comparison #28

Closed sayakpaul closed 2 years ago

deep-diver commented 2 years ago

Pure TF model with FastAPI? Or TF serving?

sayakpaul commented 2 years ago

First option. I think we should cover the second in a standalone project.

sayakpaul commented 2 years ago

TFServing is a better alternative: https://github.com/deep-diver/ml-deployment-k8s-tfserving