Toolkit for allowing inference and serving with PyTorch on SageMaker. Dockerfiles used for building SageMaker Pytorch Containers are at https://github.com/aws/deep-learning-containers.
Description of changes:
Since the EI-enabled client framework is CPU only, we will load PyTorch models to CPU by default when the user is using Elastic Inference. The model still runs on the server in CUDA
By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.
Issue #, if available:
Description of changes: Since the EI-enabled client framework is CPU only, we will load PyTorch models to CPU by default when the user is using Elastic Inference. The model still runs on the server in CUDA
By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.