DKRZ-AIM / HAI-HI-unconference-2023

Topic collection for the Unconference at the Helmholtz AI/Helmholtz Imaging conference joint day, June 14, 2023, Hamburg
4 stars 0 forks source link

Open-source and open science for large-scale generalist learning and transferable foundation models #9

Open JeniaJitsev opened 1 year ago

JeniaJitsev commented 1 year ago

Description

Major recent breakthroughs in generalist, transferable learning were executed by training and using large-scale language, vision or language-vision foundation models like GPT, ViT, CLIP or Stable Diffusion. Strong transferability of these models in zero-shot and few-shot regimes makes it possible to adapt those to various tasks where only small amounts of training data or no training data is available, while the pre-training of such models require both large compute and large volumes of generic data.

However, the study of such foundation models at larger scales is still mostly conducted in few large industry labs (openAI, Google, Facebook). To counteract this trend of power concentration and allow broader research community to study these important model class, various grassroot organizations self-organized into networks of research labs, machine learning enthusiasts and citizen scientists to gain expertise and resources necessary for such studies. Some important examples are

  1. BigScience, France https://bigscience.huggingface.co/
  2. ELeutherAI, USA https://www.eleuther.ai/
  3. LAION, Germany https://laion.ai/

Following those examples, we would like to discuss what is necessary to bring back research on foundation models at scale to academia and what are the perspectives of open-source and open science for studying transferable foundation models at larger scale. Speficially, we would like to discuss collaborative efforts

Organizational

Organizer(s)

Jenia Jitsev (JSC, LAION), j.jitsev@fz-juelich.de Christoph Schuhmann (LAION), christoph.schuhmann@laion.ai

Format

Open discussion & world cafe with smaller groups according to topics (datasets, model training)

Timeframe

1.5 h

Number of participants

Min. 3 participants

SusanneWenzel commented 1 year ago

@JeniaJitsev Would you need a screen? We possibly don't have one for each session. Flipchart will be available

ilsenatorov commented 1 year ago

Would it be possible to start at 14:30?

SusanneWenzel commented 1 year ago

Sure, already noted

JeniaJitsev commented 1 year ago

@JeniaJitsev Would you need a screen? We possibly don't have one for each session. Flipchart will be available

Screen would be actually very helpful, yes

JeniaJitsev commented 1 year ago

We start 14:30 and are looking forward fruitful discussion and exchange! As a good starter point - if you have a story of using a pre-trained foundation model (LLM, vision, language-vision) for transferring to your particular research, feel free to share it.

SusanneWenzel commented 1 year ago

@JeniaJitsev if possible, please make a note here on the (rough) number of participants. Also don't forget to make a note here about the outcome of the session and, if applicable, future plans that came out of this session.