Closed deep-diver closed 6 months ago
@sayakpaul
addressed comments
: local LLM (gemma)'s batch inference with dataset.map()
approach.
: makes simultaneous gemini API calls on the batch generated outputs
@sayakpaul
addressed your comments!
Thanks for iterating!
this PR basically runs the programs written in a jupyter notebook for model evaluation #5
when it is run, the following is printed out
and it returns the following values for the downstream jobs
This code runs the text generation inference with batch size of 1. In the later follow-up PR, text generation with arbitrary batch sizes should be supported. To do this, we need the following code modifications: