In this session, we will discuss the possible impacts of generative AI on the Web. As it is an emerging topic and its consequences still need to be better understood, we will focus on technical and governance challenges in three main dimensions: data collection on the Web, Integration of LLMs in a search engine; production and publication of synthetic contents on the Web.
To facilitate the discussion, we will start from three main questions, which can be discussed in different groups depending on the number of participants:
What are the limits of scraping web data to train Generative AI and what technical measures should be implemented to ensure privacy, prevent copyright infringement, and effectively manage content creator consent?
What are the potential impacts of incorporating LLMs (Large Language Models) as chatbot interfaces into search engines? How would this change the web traffic and digital economy once the chatbot could respond directly to users instead of providing links?
How could Web technologies help detect AI-generated content posted on the Web, restrain the dissemination of sensitive content, and provide accountability?__
Session goal
The session aims to introduce the theme and discuss the possible consequences of Generative AI for the Web. We also want to stimulate a debate with people from different backgrounds on how Web technologies should respond to those changes. We also plan to elaborate and consolidate the result of the session to guide further discussion on the topic.
Additional session chairs (Optional)
My friend Ian
Another friend of mine
IRC channel (Optional)
GenerativeAI
Who can attend
Anyone may attend (Default)
Session duration
60 minutes (Default)
Other sessions where we should avoid scheduling conflicts (Optional)
Session description
In this session, we will discuss the possible impacts of generative AI on the Web. As it is an emerging topic and its consequences still need to be better understood, we will focus on technical and governance challenges in three main dimensions: data collection on the Web, Integration of LLMs in a search engine; production and publication of synthetic contents on the Web.
To facilitate the discussion, we will start from three main questions, which can be discussed in different groups depending on the number of participants:
What are the limits of scraping web data to train Generative AI and what technical measures should be implemented to ensure privacy, prevent copyright infringement, and effectively manage content creator consent?
What are the potential impacts of incorporating LLMs (Large Language Models) as chatbot interfaces into search engines? How would this change the web traffic and digital economy once the chatbot could respond directly to users instead of providing links?
How could Web technologies help detect AI-generated content posted on the Web, restrain the dissemination of sensitive content, and provide accountability?__
Session goal
The session aims to introduce the theme and discuss the possible consequences of Generative AI for the Web. We also want to stimulate a debate with people from different backgrounds on how Web technologies should respond to those changes. We also plan to elaborate and consolidate the result of the session to guide further discussion on the topic.
Additional session chairs (Optional)
IRC channel (Optional)
GenerativeAI
Who can attend
Anyone may attend (Default)
Session duration
60 minutes (Default)
Other sessions where we should avoid scheduling conflicts (Optional)
No response
Estimated number of in-person attendees
Don't know (Default)
Instructions for meeting planners (Optional)
No response
Agenda, minutes, slides, etc.