castorini / ura-projects

0 stars 1 forks source link

What's the largest open-source LLM we can run on Colab? #4

Open lintool opened 10 months ago

lintool commented 10 months ago

Can someone poke around ad find out?

Initial poking by me:

lintool commented 10 months ago

Ref: https://twitter.com/1littlecoder/status/1664368229232328709

Screen Shot 2023-09-04 at 1 15 59 PM
AndreSlavescu commented 3 months ago

https://huggingface.co/TheBloke/Wizard-Vicuna-30B-Uncensored-GPTQ

largest model by parameter size that fits on free tier colab

lintool commented 3 months ago

@AndreSlavescu is there a notebook demonstrating this?

AndreSlavescu commented 3 months ago

@AndreSlavescu is there a notebook demonstrating this?

Made a notebook here:

https://colab.research.google.com/drive/1LjWNiGyc331IyBogYi1q1FvYTYxLj-BF?usp=sharing

works well with T4 and takes roughly 15 mins end to end (installing deps + downloading weights + loading and inf)

AndreSlavescu commented 3 months ago

An observation I had though was that the T4 now seems to have reduced memory. I thought it used to have 16 gigs previously, but in the event its ever slightly too little, Kaggle's free tier K80 might be a good idea as well.