ShishirPatil / gorilla

Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
https://gorilla.cs.berkeley.edu/
Apache License 2.0
11.53k stars 1.01k forks source link

Hi, I am going to share some data with you guys. #99

Open ValorLin opened 1 year ago

ValorLin commented 1 year ago

Hi, I am interested in this project and I have done some reasearch. I tried the evaluation on the TorchHub API senerio with some of the latest models and other fine-tuned models.

Below is a example of the training data that I used for fine-tuning on OpenAI models.

{"prompt":"I need an API that can estimate the relative depth of objects from a single image. The output should be an inverse depth map.\n\n###\n\n","completion":" {'domain': 'Classification', 'api_call': \"model = torch.hub.load('intel-isl\/MiDaS', 'DPT_Hybrid', pretrained=True)\", 'api_provider': 'PyTorch', 'explanation': \"Load the pretrained MiDaS model for estimating relative depth from a single image, which outputs an inverse depth map.\", 'code': 'import torch\nmodel = torch.hub.load(\"intel-isl\/MiDaS\", \"DPT_Hybrid\", pretrained=True)'}###"}

The full converted training data can be found here: https://gist.github.com/ValorLin/807ac5f75361775dfd2cc31f21a03e77

I hope that data is helpful. Its very interesting result, you should try it.

ShishirPatil commented 1 year ago

Hey @ValorLin thank's for sharing this! Super insightful. Yeah, one thing we might want to keep in mind is that OpenAI's models are slightly larger, vs the 7B Gorilla models - so that might explain the performance (?) Or potentially better pre-training data? Would be interesting for the open source community to explore.

BTW do you mind writing a blog or a write-up, with your pipeline including the data-generation, results, insights on where it does well, not so well, etc ! Would be a great read for the community - including and especially me 😄

Nice job!