aws / sagemaker-pytorch-inference-toolkit

Toolkit for allowing inference and serving with PyTorch on SageMaker. Dockerfiles used for building SageMaker Pytorch Containers are at https://github.com/aws/deep-learning-containers.
Apache License 2.0
131 stars 70 forks source link

Add support for SAGEMAKER_MAX_PAYLOAD_IN_MB #143

Closed namannandan closed 1 year ago

namannandan commented 1 year ago

Description of changes: Sagemaker batch transform sets SAGEMAKER_MAX_PAYLOAD_IN_MB env var which is currently not handled and passed to the model server. Adding support for the same based on the following PR in base toolkit: https://github.com/aws/sagemaker-inference-toolkit/pull/121

Testing:

By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.

sagemaker-bot commented 1 year ago

AWS CodeBuild CI Report

Powered by github-codebuild-logs, available on the AWS Serverless Application Repository

lxning commented 1 year ago

this PR breaks the PT container env convention (see notes). It will be confusing for cx. In summary, there is no acton needed in PT container.

@namannandan pls close this PR w/o merge