Closed sergiopaniego closed 1 month ago
I'm not sure if the file location is correct, and we might need to link the recipe in the README or somewhere else, so that may need updating as well. 😄
@sergiopaniego I really like the notebook! :fire:
Let's have the current notebook broken down into two parts:
CC: @Vaibhavs10 what do you think?
Thanks a lot for the feedback, @ariG23498!
I think it makes sense to split the notebook into two separate parts, as you suggested. Should I start working on it now, or do we need a double-check from @Vaibhavs10? 😄
Let' wait for @Vaibhavs10 to give it a look? I don't want you to work twice 😅
Thanks a lot for the in-depth review!
I'll update the PR accordingly 👨💻
I've updated the notebook covering all the possible improvements mentioned @pcuenca. 🤗
Open to any further suggestions or enhancements!
cc @ariG23498 @Vaibhavs10
Thank you for the review, @Vaibhavs10! 😊
I've updated the notebook with a new sentence: These models, along with many others, are efficiently deployed via TGI on the serverless API.
Additionally, I’ve renamed both the folder and the file: llama_tgi/tgi_recipe.ipynb -> llama_tgi_api_inference/tgi_api_inference_recipe.ipynb
I hope this aligns with the expectations!
Added @Vaibhavs10 😄! Thanks a lot for all the comments and suggestions!!
Perfect from myside to merge - I'd let @pcuenca take the final pass and merge! 🤗
Thanks for the review and suggestions @pcuenca! I've improved the notebook based on them 😄
Adding a TGI recipe that includes examples of how to serve and consume the 🦙 Llama models using different methods available in the HF ecosystem.
fix #45
@ariG23498