Closed ctandrewtran closed 1 year ago
Is it possible to deploy to sagemaker by passing the task + model id and provide custom inference code?
Indirectly, the "parameter" for TASK and MODEL_ID are passed in as environment variables so when you use an inference.py
you can access them through os.environ.get()
and then the model would be loaded from the hub and not from s3.
Is it possible to deploy to sagemaker by passing the task + model id and provide custom inference code?
Indirectly, the "parameter" for TASK and MODEL_ID are passed in as environment variables so when you use an
inference.py
you can access them throughos.environ.get()
and then the model would be loaded from the hub and not from s3.
Thank you for the response!
To double check my understanding, I would do the following then:
AutoTokenizer(model_environ_var)
etcIs there any difference between the above and simply doing AutoTokenizer(flan-xxl)
?
Your assumption is correct.
Thank you!
Hello Phil-
Big fan of your work. Been following this tutorial, and have a quick question.
I've noticed there are two ways to deploy with certain use cases
Is it possible to deploy to sagemaker by passing the task + model id and provide custom inference code?