Open JeniaJitsev opened 1 year ago
@JeniaJitsev Would you need a screen? We possibly don't have one for each session. Flipchart will be available
Would it be possible to start at 14:30?
Sure, already noted
@JeniaJitsev Would you need a screen? We possibly don't have one for each session. Flipchart will be available
Screen would be actually very helpful, yes
We start 14:30 and are looking forward fruitful discussion and exchange! As a good starter point - if you have a story of using a pre-trained foundation model (LLM, vision, language-vision) for transferring to your particular research, feel free to share it.
@JeniaJitsev if possible, please make a note here on the (rough) number of participants. Also don't forget to make a note here about the outcome of the session and, if applicable, future plans that came out of this session.
Description
Major recent breakthroughs in generalist, transferable learning were executed by training and using large-scale language, vision or language-vision foundation models like GPT, ViT, CLIP or Stable Diffusion. Strong transferability of these models in zero-shot and few-shot regimes makes it possible to adapt those to various tasks where only small amounts of training data or no training data is available, while the pre-training of such models require both large compute and large volumes of generic data.
However, the study of such foundation models at larger scales is still mostly conducted in few large industry labs (openAI, Google, Facebook). To counteract this trend of power concentration and allow broader research community to study these important model class, various grassroot organizations self-organized into networks of research labs, machine learning enthusiasts and citizen scientists to gain expertise and resources necessary for such studies. Some important examples are
Following those examples, we would like to discuss what is necessary to bring back research on foundation models at scale to academia and what are the perspectives of open-source and open science for studying transferable foundation models at larger scale. Speficially, we would like to discuss collaborative efforts
Organizational
Organizer(s)
Jenia Jitsev (JSC, LAION), j.jitsev@fz-juelich.de Christoph Schuhmann (LAION), christoph.schuhmann@laion.ai
Format
Open discussion & world cafe with smaller groups according to topics (datasets, model training)
Timeframe
1.5 h
Number of participants
Min. 3 participants