declare-lab / flan-alpaca

This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as Flan-T5.
Apache License 2.0
348 stars 38 forks source link

use gpt4all dataset #7

Closed Shiro836 closed 1 year ago

Shiro836 commented 1 year ago

I am impressed by the quality results and speed of the XL model. And it can very easily fit in my 24GB GPU. I found this dataset recently: https://github.com/nomic-ai/gpt4all But my GPU is unable to train the model due to insufficient memory. Can anyone train it, please?

Shiro836 commented 1 year ago

Awesome!!! The results are insanely good.

munish0838 commented 1 year ago

Hi @Shiro836, can you please share the training process!

Shiro836 commented 1 year ago

@munish0838 I didn't train it, you can find trained model in README.md

munish0838 commented 1 year ago

thank you @Shiro836