Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

2024.06.26 #70

Closed
24 tasks
seanmcilroy29 opened this issue Jun 7, 2024 · 5 comments
Closed
24 tasks

2024.06.26 #70

seanmcilroy29 opened this issue Jun 7, 2024 · 5 comments
Assignees

Comments

@seanmcilroy29
Copy link
Contributor

seanmcilroy29 commented Jun 7, 2024


2024.06.26 Agenda/Minutes


Time: Bi-weekly @ 1700 (BST) - See the time in your timezone

  • Co-Chair - Chris Xie- (Futurewei)
  • Co-Chair - Vacant
  • Convener – Sean Mcilroy (Linux Foundation)

Antitrust Policy

Joint Development Foundation meetings may involve participation by industry competitors, and the Joint Development Foundation intends to conduct all of its activities in accordance with applicable antitrust and competition laws. It is, therefore, extremely important that attendees adhere to meeting agendas and be aware of and not participate in any activities that are prohibited under applicable US state, federal or foreign antitrust and competition laws.

If you have questions about these matters, please contact your company counsel or counsel to the Joint Development Foundation, DLA Piper.

Recordings

WG agreed to record all Meetings. This meeting recording will be available until the next scheduled meeting.
Meeting recording link

Roll Call

Please add 'Attended' to this issue during the meeting to denote attendance.

Any untracked attendees will be added by the GSF team below:

  • Full Name, Affiliation, (optional) GitHub username

Agenda

  • Approve agenda
  • Approve previous Meeting Minutes

Introductions

  • Any new members?

Standard update submissions

Project Dashboard

Future meeting Agenda submissions

Previous meeting AI's

Action Items

  • Chris - Incorporate suggestions around terminology and framing from Asim into the Share presentation
  • Project members - Explore benchmarking approaches and how to prescribe standardised tests

Next Meeting

  • 26th Jun @ 9 am P/T

Adjourn

  • Motion to adjourn

Meeting Action Items / Standing Agenda / Future Agenda submissions

  • Add here
@chrisxie-fw
Copy link
Contributor

attended

@seanmcilroy29 seanmcilroy29 changed the title 2024.06.12 2024.06.26 Jun 26, 2024
@metacosm
Copy link

attended

1 similar comment
@adityamanglik
Copy link
Contributor

attended

@seanmcilroy29
Copy link
Contributor Author

MoM

Chris opens the meeting at 1700 (BST)

Rating systems for AI models, including concerns about lack of transparency and standardization.
Chris discusses issues with the rating system and advocates for transparency and disclosure.
The group discussed the lack of disclosure in rating systems, using examples from Nutriscore and Hugging Face.
Chris emphasizes the importance of a standardized framework for rating AI language models.

Standardizing AI model benchmarking process.
Chris outlines four steps for evaluating large language models, including software categorization and benchmarking.
Tamar and the group discuss the number of tokens per kilowatt hour in the context of AI model benchmarking.
Chris clarifies that the standard framework does not define algorithms but provides a four-step framework.

Defining and benchmarking large language models.
Chris discusses the importance of defining a "profile" for language models, including model size, application type, and performance metrics. Chris discusses the importance of defining base class for publicization readings of large language models. Chris discusses using a base software framework for AI models, with music streaming and healthcare examples.

Standardizing benchmarking for AI models to compare their efficiency.
Tamar and group 1 discuss the challenge of defining categories and benchmarks for software evaluation. Chris proposes a framework for standardizing the evaluation process, with four common steps across all software. The group discussed how to use a shared standard process to measure green request emissions. Chris explains how to assess the efficiency of AI models using a rating system based on requests.

Standardizing carbon efficiency ratings for AI models.
Chris outlines Salesforce uses a community-driven standard, Sheer, to assess the carbon footprint of their services for customers. The group discussed concerns about the rating system's ability to compare companies offering large language models based on carbon efficiency. Chris: The company aims to create industry-wide standards for language models. The group discussed the goal of having standardized tests for language models across companies. Chris proposes defining a carbon-efficient rating for large language models open to other software categories. The member asks about focusing on LLM standards and potentially expanding to other software categories.

AI standards and frameworks for software development.
Tamar and the group discussed the complexity of AI and its unique challenges compared to other software products.
The group explores defining categories for AI, including depth-first and breadth-first approaches, and the need for experts in specific areas to define standards. Chris discusses default implementations and disclosure in AI models. Tamar shares updates on the green AI committee and its intersection with reinventions.

Action Items

  • Explore potential extensions of the standard framework to other software categories like databases
  • Participate in the new Green AI Committee to represent the interests of the SCER standard

@seanmcilroy29
Copy link
Contributor Author

Approved

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

5 participants