Green-Software-Foundation / scer

Software Carbon Efficiency Rating
Other
29 stars 5 forks source link

2024.02.21 #30

Closed seanmcilroy29 closed 6 months ago

seanmcilroy29 commented 6 months ago

2024.02.21 Agenda/Minutes


Time: Bi-weekly @ 1600 (GMT) - See the time in your timezone

Antitrust Policy

Joint Development Foundation meetings may involve participation by industry competitors, and the Joint Development Foundation intends to conduct all of its activities in accordance with applicable antitrust and competition laws. It is, therefore, extremely important that attendees adhere to meeting agendas and be aware of and not participate in any activities that are prohibited under applicable US state, federal or foreign antitrust and competition laws.

If you have questions about these matters, please contact your company counsel or counsel to the Joint Development Foundation, DLA Piper.

Recordings

WG agreed to record all Meetings. This meeting recording will be available until the next scheduled meeting. Meeting recording link

Roll Call

Please add 'Attended' to this issue during the meeting to denote attendance.

Any untracked attendees will be added by the GSF team below:

Agenda

Introductions

PR submissions

Review IMDA Email

Submitted Issues

Draft Timeline

Spec Review

Next Steps

AOB

Future meeting Agenda submissions

Next Meeting

Adjourn

Meeting Action Items / Standing Agenda / Future Agenda submissions

mvaltas commented 6 months ago

Attended

chrisxie-fw commented 6 months ago

attended.

chrisxie-fw commented 6 months ago

The team agreed on the following meeting minutes and next steps:

seanmcilroy29 commented 6 months ago

MoM

Energy consumption and carbon footprint of large language models. Chris has proposed a framework for evaluating and comparing community agencies using large language models. Additionally, he has raised concerns about the energy consumption and carbon footprint of high-end Nvidia chips, which consume the same energy as a small nation and have a larger carbon footprint by 2024. To address the issue of differences in profiles, life cycles, and efficiency tuning, Chris suggests separating large language models into production and interference phases. He also emphasizes the importance of adopting more efficient and sustainable practices in AI model training to reduce energy consumption and emissions.

Carbon footprint of AI training. Marco suggests that we should categorize the efficiency of training based on the amount of carbon emissions it produces, taking into account both the training and deployment phases. Sean O and Marco discuss the carbon emissions associated with AI models during the discussion, and Chris shares his insights on how Hugging Face accounts for these emissions. Chris also talks about a project that successfully reduced carbon emissions by 20 metric tonnes, using a French supercomputer powered by nuclear energy. The team also talks about Carbon AI, a tool that aims to estimate the carbon footprint of AI models, including national language models, by tracking the consumption of computer resources and energy mix. Chris discusses the most popular hugging rates and their downloads, likes, and environmental impact. Marco raises a question about why commonly visited regions are not included in the evaluation parameter and suggests some potential alternatives for providing value.

AI model rating and training costs. Chris: Algorithms become a software category, with training and operating distinct.

Standardizing AI model training and evaluation. Sean O and Chris discussed incorporating a new AI model into an existing specification. The group consider various factors, such as training data and hardware embodied carbon. Chris emphasizes the importance of standardizing AI models and their applications. He highlights the need to identify different phases of the model lifecycle.

AI model training costs and efficiency. Chris suggested that "inbody" should be defined as the cost of training a model. This could help justify the cost of training and spread it across millions of users. On the other hand, Marco proposed that the training cost should be separate from the user perspective. He suggested focusing on hardware requirements and tokens per kilowatt for efficient deployment. The group discussed the cost of training large language models (LLMs) and how it should be factored into the product itself. Marco also mentioned open-source projects for national model safety and workload testing and highlighted how Google and OpenAI advocate for a training phase to lower the carbon cost of software development. Chris expressed interest in reviewing and learning from the framework for setting up standards or practices. He aims to achieve a similar attitude towards lowering carbon costs in software development.

AI model benchmarking and rating. Marco leads a discussion about the Open AI leaderboard, which is a platform for benchmarking AI models. This includes metrics related to energy and safety. The group discusses how to obtain and rate data to benchmark these models. Marco suggests collaborating with a benchmarking agency to promote reading and standardize a linking algorithm. Sean O suggests preparing something to present to potential partners, such as visualizations of their data. Marco suggests creating a website that scrapes their data and provides value in return. Chris suggests investigating if potential partners' definition of integrity aligns with their own. Marco agrees and suggests drilling down a subset of their data to create a rating for a specific category.

Action Items [ ] Review the draft SCER spec for RMS submitted by Chris [ ] Review the IMDA email sent by Sean M. [ ] Review the issues added by Kitty [ ] Review Chris' updated AI use case study [ ] Review the Hugging Face benchmarks and process [ ] Gather data from Hugging Face and calculate ratings [ ] Build a demo site with ratings using Hugging Face data [ ] Align integrity definitions between Hugging Face and Green Software Foundation [ ] Identify a subset of LLM categories to focus on