Open AIByteSmith opened 1 year ago
Wondering if it is possible to add streaming support for models hosted on Sagemaker. This example makes it seem possible, but the implementation looks awkward. https://github.com/aws/amazon-sagemaker-examples/blob/main/inference/generativeai/llm-workshop/lab6-stream-with-pagination/stream_pagination_lmi.ipynb
Hello Streaming is not yet support on SageMaker, easyllm is not using any layer in between. I hope they ll add support soon then i ll add it as well.
Wondering if it is possible to add streaming support for models hosted on Sagemaker. This example makes it seem possible, but the implementation looks awkward. https://github.com/aws/amazon-sagemaker-examples/blob/main/inference/generativeai/llm-workshop/lab6-stream-with-pagination/stream_pagination_lmi.ipynb