THE open source ai definition 1.0

We have released the first stable version of the Definition.

How was Open Source AI defined?

The Open Source Definition is a practical guide to judge if legal documents grant the four freedoms to software, following the principles of the GNU Manifesto. More than two decades passed between the GNU Manifesto and the writing of the Open Source Definition. For AI we cannot wait decades to produce a new document. The Open Source Initiative started coordinating in 2022 a global process to sharpen collective knowledge and identify the principles that lead to a widely adopted Open Source AI Definition (OSAID).

OSI brought together global experts to establish a shared set of principles that can recreate permissionless, pragmatic and simplified collaboration for AI practitioners, similar to that which the Open Source Definition has done for the software ecosystem.  The output of this work is a published definition document on the website of the Open Source Initiative.

The document will be used to validate whether an AI system is an Open Source AI, or not. The validation process will be similar to the evaluation of existing licenses for software: community-led, open and public.

The Process

The board requires by the in-person meeting of 2024 in Raleigh, an Open Source AI Definition that is supported by stakeholders that include deployers of AI, end users of AI and subjects (those affected by AI decisions), provides positive examples of AI systems, is rooted in current practice and provides a reference for interested parties.

Collaboration with multiple stakeholders

The Open Source Initiative (OSI) has guarded the Open Source Definition for over 25 years and has robust processes for developing, amending, and consulting on licenses. This authority position is recognised by a number of leaders in organizations who have agreed to co-design a new definition suited for AI and ML.

These leaders joined the co-design process in a personal capacity and with various degrees of direct involvement from their employers like Digital Public Goods Alliance, Mozilla Foundation, Open Knowledge Foundation, National Endowment for Democracy, Center for Tech and Civic Life, Code for America, Wikimedia Foundation, Creative Commons, Linux Foundation, MLCommons, EleutherAI, Open Future, GitHub, Microsoft, Google, DataStax, Amazon, Meta, Hugging Face, GIZ FAIR Forward – AI for All, OpenLLM France, Polytechnic Institute of Paris, Intel, Apache Software Foundation, Samsung, and the United Nations International Telecommunications Union (ITU).

Co-design process

Co-design, also called participatory or human-centered design, is a set of creative methods used to solve communal problems by sharing knowledge and power. The co-design methodology addresses the challenges of reaching an agreed definition within a diverse community (Costanza-Chock, 2020: Escobar, 2018: Creative Reaction Lab, 2018: Friedman et al., 2019). As noted in MIT Technology Review’s article about this project, “[t]he open-source community is a big tent… encompassing everything from hacktivists to Fortune 500 companies…. With so many competing interests to consider, finding a solution that satisfies everyone while ensuring that the biggest companies play along is no easy task.” (Gent, 2024). The co-design method allows us to integrate these diverging perspectives into one just, cohesive, and feasible standard. Support from such a significant and broad group of people also creates a tension to be managed between moving swiftly enough to deliver outputs that can be used operationally, and taking the time to consult widely to understand the big issues and garner community buy-in. 

The first step of the co-design process was to identify the freedoms needed for Open Source AI. After various online and in-person activities and discussions, including five workshops across the world, the community adopted the four freedoms for software, now adapted for AI systems.

The next step was to form four working groups to initially analyze four AI systems. To achieve better representation, special attention was given to diversity, equity and inclusion. Over 50% of the working group participants are people of color, 30% are black, 75% were born outside the US and 25% are women, trans and nonbinary.

These working groups discussed and voted on which AI system components should be required to satisfy the four freedoms for AI. The components we adopted are described in the Model Openness Framework developed by the Linux Foundation.

The vote compilation was performed based on the mean total votes per component (μ). Components which received over 2μ votes were marked as required and between 1.5μ and 2μ were marked likely required. Components that received between 0.5μ and μ were marked likely not required and less than 0.5μ as not required.

The working groups evaluated legal frameworks and legal documents for each component. Finally, each working group published a recommendation report. The end result is the OSAID with a comprehensive definition checklist encompassing a total of 17 components. More working groups are being formed to evaluate how well other AI systems align with the definition.

OSAID multi-stakeholder co-design process: from component list to a definition checklist

OSAID multi-stakeholder co-design process: from component list to a definition checklist

The Open Source AI Definition Process

We have released the first stable version

RC1

  • Published in early October
  • The draft is completed in all its parts
  • The draft is supported by at least 2 representatives for each of the 6 stakeholder groups

Stable version

  • Outcome of in-person and online meetings through the summer/early autumn
  • The document is endorsed by at least 5 reps for each of the stakeholder groups
  • Announced in late October
See the 2023 project activity

June 7, 2023

Open call for papers, webinar series

Collect ideas from the community describing precise problem areas in AI and collect suggestions for solutions.

June 21
San Francisco, CA

Kickoff meeting

A small gathering at the Mozilla HQ to set the ground rules, create the first working document of a Definition of Open Source AI.

July 14
Portland, OR

Community review #1

Meeting at birds of a feather in Portland during the FOSSY event from the Software Freedom Conservancy.

July 27-28
Geneva, Switzerland

Community review #2

Meeting at birds of a feather at the Linux Foundation Open Source Congress in Geneva.

September-November
Online

Deep Dive Webinars

Host the online presentations from experts approved by the Deep Dive program committee

September 19-21
Bilbao, Spain

Community review #3

Meeting at birds of a feather in Bilbao, Spain around the Linux Foundation Open Source Summit.

October 17
Raleigh, NC

Release candidate 1

Meeting at All Things Open at OSI’s day event. We’ll collect all the feedback and prepare a release candidate of the Open Source AI Definition.

Oct 24, 2023

Community review #4

Workshop to review the latest draft of the Definition at the Linux Foundation Members Summit in Monterey, CA.

Nov 14, 2023

Community review #5

Digital Public Goods Alliance annual members meeting in Addis Ababa, Ethiopia.

Dec 12, 2023

Community review #6

Workshop at the Linux Foundation’s AI.dev (Open Source GenAI & ML Summit) in Palo Alto, California.

February 2024

Call For Volunteers
+ Activity Feedback and Revision

FOSDEM talk (Brussels)

Bi-Weekly Virtual Public Town halls.

Draft 0.0.5

March

Virtual System Review Meetings Begin

Draft 0.0.6

April

Virtual System Review Meetings Continue

Open Source Summit North America workshop

The Free Software Legal and Licensing Event workshop

Draft 0.0.7

May

Virtual System Review Meetings END

PyCon workshop (Pittsburgh)

Draft 0.0.8

June

Feedback Informs Content of OSI In-Person Stakeholder Meeting

OW2 talk (Paris)

Open Expo Europe talk(Madrid)

RC 1

July

OSPOs for Good panel session (New York)


OSCA Community webinar (Virtual)

August

AI_dev talk (Hong Kong)

Open Source Congress talk (Beijing)

0.0.9

September

Deep Learning Indaba talk (Dakar)

India FOSS talk (Bangalore)

OSS Europe talk (Vienna)

Nerdearla talk (Buenos Aires)

Release Candidate 1

October

Data in OSAI workshop (Paris)

OCX talk (Mainz)

All Things Open Stable Version Presentation (Raleigh)

Release Stable Version

Ongoing

It doesn’t end with the Stable Version

We’ll need to define rules for maintenance and review of the Definition. The OSI board of directors approved the creation of a new committee to oversee the development of the Open Source AI Definition, approve the Stable Version and set rules for the maintenance of Definition.

Who is involved in this process?

🛠️ System Creators

Makes AI system and/or component that will be studied, used, modified, or shared through an open source license.

📃 License Creators

Writes or edits the open source license to be applied to the AI system or component; includes compliance.

🏛️ Regulators

Writes or edits rules governing licenses and systems (e.g. government policy-maker).

🎓 Licensees

Seeks to study, use modify, or share an open source AI system (e.g. AI engineer, health researcher, education researcher)

⌨️ End Users

Consumes a system output, but does not seek to study, use, modify, or share the system (e.g., student using a chatbot to write a report, artist creating an image)

🙇 Subjects

Affected upstream or downstream by a system output without interacting with it intentionally; includes advocates for this group (e.g. people with loan denied, or content creators).

Governance

Governance for the project is provided by the OSI Board of Directors. The OSI board members have expertise in business, legal, and open source software development, as well as experience across a range of commercial, public sector, and non-profit organizations. Formal progress reports including achievements, budget updates, and next steps are provided monthly by the Program Lead for advice and guidance as part of regular Board business. Additionally, informal updates on the outcomes of key meetings and milestones are provided via email to the Board as required. Details of the current Board, including profiles for each Director are available on here.

How to participate

The OSAID co-design process is open to everyone interested in collaborating. There are many ways to get involved:


Previous Work

Deep Dive AI Webinar Series 2023

Speakers from law, academia, enterprise, NGOs, and the OSS community presented webinars addressing pressing issues and potential solutions in our use and development of AI systems. 

All Things Open – Deep Dive AI 2023

After two community reviews and a first pass at comments, we released a new draft version. The base is a preamble to explain “why Open Source AI”, followed by the beginning of a formal definition.

2022 Deep Dive AI Podcasts

We released a series of 6 podcasts with experts on the matter to discuss all various aspects of Open Source AI.

2022 Panel Discussions

Four experts in Business, Society, Legal, Academia further dissect the issues posed by AI systems.

Deep Dive AI: The 2023 Report Update

By bringing together experts from various domains, the OSI is actively contributing to the discourse on Open Source AI, laying the groundwork for a future where the principles of openness, transparency, and collaboration continue to underpin the evolution of cutting-edge technologies for the benefit of society as a whole.

Deep Dive AI: The 2022 Report

What does it mean for an AI system to be Open Source?

This report summarizes the discussions above and underscores what we’ve learned about the challenges and opportunities for the Open Source movement posed by AI.

Deep Dive Report 2022

Supported by

Alfred P. Sloan Foundation

OSI’s efforts wouldn’t be possible without the support of our sponsors and thousands of individual members. Become a sponsor or join us today!