Overview
As part of this definitional co-design process, four working groups were convened in January and February to vote on which components should be required for an AI system to be considered “open source” according to the Open Source AI Definition.
To make the working groups as global and representative as possible, we conducted focused outreach to increase inclusion of Black, Indigenous, and other People of Color, particularly women and individuals from the Global South. Each working group also included either one or two creators or advisors on the system under discussion to provide technical expertise. The reports from those groups, including member lists and voting results, are below:
- Report from Llama 2 working group
- Report from Pythia working group
- Report from BLOOM working group
- Report from OpenCV working group
Working group members were invited to vote as to whether each component was required to study, use, modify, and/or share that AI system. The votes from all working groups were then compiled to create a total tally of votes. The compiled votes can be viewed on this spreadsheet. I then created a rubric, based on the mean number of votes per component (µ), to create a recommendation associated with each component. The details of that rubric are also in the spreadsheet, in column M.
The spreadsheet and recommendations were shared with all working group members via email. They were also shared publicly at last Friday’s townhall and on Tuesday in a Zoom meeting open to all working group members. The next step emerging from the Tuesday meeting was to share the recommendations in this forum for public comment.
Recommendations
The recommendations below respond to the question:
- Should X component be required for an AI system to be licensed as open?
Based on the number of votes for each component across all working groups, the results are as follows:
Required
- Training, validation, and testing code
- Inference code
- Model architecture
- Model parameters
- Supporting libraries & tools*
Likely Required
- Data preprocessing code
Maybe Required
- Training datasets
- Testing datasets
- Usage documentation
- Research paper
Likely Not Required
- Model card
- Evaluation code
- Validation datasets
- Benchmarking datasets
- All other data documentation
Not Required
- Data card
- Evaluation data
- Evaluation results
- Model metadata
- Sample model outputs
- Technical report
- Code used to perform inference for benchmark tests
*Includes other libraries or code artifacts that are part of the system, such as tokenizers and hyperparameter search code, if used.
We look forward to reading your thoughts and questions below. Thank you again for being part of this process.