huggingface / hub-docs

Docs of the Hugging Face Hub
http://hf.co/docs/hub
Apache License 2.0
271 stars 233 forks source link

Give users the ability to compare models' outputs for a given task. #56

Open dynamicwebpaige opened 2 years ago

dynamicwebpaige commented 2 years ago

🗒️ Motivation

When a user selects a specific task on the Hugging Face Hub - for example, image-to-text:

Screen Shot 2022-03-05 at 10 14 06 PM

That user is shown a series of models, with no guidance as to which model might be state of the art, or which might be the most performant for their use case.

To test the capabilities and behavior of each model, the user must:

🙏 Desired Behavior

The user should be able to:

gary149 commented 2 years ago

This would be so cool, I really like the user story you made with the side-by-side benchmark!

We were talking about integrating something like this but into tasks pages, so the workflow would be:

  1. Select a task
  2. Read a bit about the task and discover some SOTA models (we tried to editorialize that a bit by writing an explanatory text about each task and associating a note to each hand-curated model).
  3. Compute multiple widgets with a single input to compare the outputs (today you only get one curated model to test the task).

And maybe when you select a particular task on /models we could add a link to the task page:

image

This is probably the simplest way of doing it but I understand that it's not the same as having it directly integrated into the /models page.

So maybe we want to go further and do it exactly like you said: integrate it directly into the /models page: you drag/input a picture/audio/text and all the visible models on the page compute and it switches to "benchmark mode" (that could be gamechanger 🤯). That will of course be a lot of work and I'm not even sure that we can hold that many computations at the same time 👀 (edit: we will find a way 👍 ).