groq / mlagility

Machine Learning Agility (MLAgility) benchmark and benchmarking tools
MIT License
38 stars 11 forks source link

Create LLM corpus #314

Closed danielholanda closed 1 year ago

danielholanda commented 1 year ago

Issue description

Some LLMs are currently part of the popular_on_huggingface corpus of MLAgility. Those models are significantly large, causing our benchmarking infrastructure to take significantly longer when processing them.

Task

Create corpus composed only of LLM models, since those models must be treated differently than most models.

Suggested implementation

Example:

from transformers import OpenAIGPTModel, OpenAIGPTConfig
model = OpenAIGPTModel(config=OpenAIGPTConfig()) 

Instead of

from transformers import OpenAIGPTModel
model = OpenAIGPTModel.from_pretrained("openai-gpt")

Suggested models

Decoder-Only

Encoder-Decoder

Encoder-Only

Suggested list of mandatory labels for LLMs

danielholanda commented 1 year ago

@jeremyfowers @ramkrishna2910

ramkrishna2910 commented 1 year ago

Pointer to LLMs: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard

jeremyfowers commented 1 year ago

I will introduce two corpora:

jeremyfowers commented 1 year ago

I will also likely break out anything to do with sanitizing popular_on_huggingface into a separate issue.

danielholanda commented 1 year ago

I really like the idea of having both llm and llm_layer!

jeremyfowers commented 1 year ago

https://pypi.org/project/detoxify/ is a PyPI package required by 3 models in our popular_on_huggingface corpus. However, that package requires transformers==4.22.1, which is pretty old and doesn't have key LLMs like LLaMA. detoxify also appears to be abandoned, with no updates since 2021.

I plan to remove those 3 models from popular_on_huggingface to unblock the LLM work. I don't think those 3 models specifically are a big deal, but this does highlight a potentially bad trend of dep conflicts between our models. Those conflicts seem inevitable as we grow our set of models. We may need to have per-corpus deps or something.