Closed danielholanda closed 1 year ago
@jeremyfowers @ramkrishna2910
Pointer to LLMs: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
I will introduce two corpora:
llm
: collection of LLM modelsllm_layer
: same LLMs models as llm
, but each "model" is just a single layer of the original LLM. This is useful for faster handling, less disk/ram use, benchmarking on smaller units of HW, etc.I will also likely break out anything to do with sanitizing popular_on_huggingface
into a separate issue.
I really like the idea of having both llm
and llm_layer
!
https://pypi.org/project/detoxify/ is a PyPI package required by 3 models in our popular_on_huggingface corpus. However, that package requires transformers==4.22.1
, which is pretty old and doesn't have key LLMs like LLaMA. detoxify
also appears to be abandoned, with no updates since 2021.
I plan to remove those 3 models from popular_on_huggingface
to unblock the LLM work. I don't think those 3 models specifically are a big deal, but this does highlight a potentially bad trend of dep conflicts between our models. Those conflicts seem inevitable as we grow our set of models. We may need to have per-corpus deps or something.
Issue description
Some LLMs are currently part of the
popular_on_huggingface
corpus of MLAgility. Those models are significantly large, causing our benchmarking infrastructure to take significantly longer when processing them.Task
Create corpus composed only of LLM models, since those models must be treated differently than most models.
Suggested implementation
popular_on_huggingface
to a corpus calledLLMs
popular_on_huggingface
corpusconfigs
rather than loading pre-trained models to avoid long download times.Example:
Instead of
Suggested models
Decoder-Only
Encoder-Decoder
Encoder-Only
Suggested list of mandatory labels for LLMs
arch::encoder_only,decoder_only,encoder_decoder
author::<>
year::<>