OpenGVLab / LLaMA-Adapter

[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
GNU General Public License v3.0
5.75k stars 374 forks source link

Code for reproducing evaluation results on ScienceQA #6

Open TJKlein opened 1 year ago

TJKlein commented 1 year ago

Hi, For reasons of reproducibility, it would be great if you provided source code to reproduce the results on ScienceQA. Thanks.

gaopengpjlab commented 1 year ago

Sure, we are going to release all training/inference codes for all results reported in paper.

gaopengpjlab commented 1 year ago

@TJKlein Please the following multimodal inference code for image captioning. The code structure is very similar as ScienceQA.

https://huggingface.co/spaces/csuhan/LLaMA-Adapter/tree/main .

TJKlein commented 1 year ago

@TJKlein Please the following multimodal inference code for image captioning. The code structure is very similar as ScienceQA.

https://huggingface.co/spaces/csuhan/LLaMA-Adapter/tree/main .

Thanks for pointing at it. But I guess I would wait for the release of your scripts.

yushuinanrong commented 1 year ago

+1

basteran commented 1 year ago

Hi I am interested too in the integration of the visual adapter. When do you think the code will be released?

ZrrSkywalker commented 1 year ago

Thanks for your interest and waiting :). We are organizing the multi-modal code and will release it in one or two weeks.

basteran commented 1 year ago

Thank you for your quick response!

metemadi commented 1 year ago

hi there! any updates on this? would love to take a crack at fine-tuning on a visual instruction dataset (not just image-caption as in your amazing work in v2!).

dhyani15 commented 1 year ago

Hi, just wanted to check for any updates on this?