Posted on

by

in

OSI participates in Columbia Convening on openness and AI; first readouts available

I was invited to join Mozilla and the Columbia Institute of Global Politics in an effort that explores what “open” should mean in the AI era. A cohort of 40 leading scholars and practitioners from Open Source AI startups and companies, non-profit AI labs, and civil society organizations came together on February 29 at the Columbia Convening to collaborate on ways to strengthen and leverage openness for the good of all. We believe openness can and must play a key role in the future of AI. The Columbia Convening took an important step toward developing a framework for openness in AI with the hope that open approaches can have a significant impact on AI, just as Open Source software did in the early days of the internet and World Wide Web. 

This effort is aligned and contributes valuable knowledge to the ongoing process to find the Open Source AI Definition. 

As a result of this first meeting of Columbia Convening, two readouts have been published; a technical memorandum for technical leaders and practitioners who are shaping the future of AI, and a policy memorandum for policymakers with a focus on openness in AI.

Technical readout

The Columbia Convening on Openness and AI Technical Readout was edited by Nik Marda with review contributions from myself, Deval Pandya, Irene Solaiman, and Victor Storchan.

The technical readout highlighted the challenges of understanding openness in AI. Approaches to openness are falling under three categories: gradient/spectrum, criteria scoring, and binary. The OSI is championing a binary approach to openness, where AI systems are either “open” or “closed” based on whether they meet a certain set of criteria.

The technical readout also provided a diagram that shows how the AI stack may be described by the different dimensions (AI artifacts, documentation, and distribution) of its various components and subcomponents.

Policy readout

The Columbia Convening on Openness and AI Policy Readout was edited by Udbhav Tiwari with review contributions from Kevin Klyman, Madhulika Srikumar, and myself.

The policy readout highlighted the benefits of openness, including:

Enhancing reproducible research and promoting innovation

Creating an open ecosystem of developers and makers

Promoting inclusion through open development culture and models

Facilitating accountability and supporting bias research

Fostering security through widespread scrutiny

Reducing costs and avoiding vendor lock-In

Equipping supervisory authorities with necessary tools

Making training and inference more resource-efficient, reducing environmental harm

Ensuring competition and dynamism

Providing recourse in decision-making

The policy readout also showcased a table with the potential benefits and drawbacks of each component of the AI stack, including the code, datasets, model weights, documentation, distribution, and guardrails.

Finally, the policy readout provided some policy recommendations:

Include standardized definitions of openness as part of AI standards

Promote agency, transparency and accountability

Facilitate innovation and mitigate monopolistic practices

Expand access to computational resources

Mandate risk assessment and management for certain AI applications

Hold independent audits and red teaming

Update privacy legislation to specifically address AI challenges

Updated legal framework to distinguish the responsibilities of different actors

Nurture AI research and development grounded in openness

Invest in education and specialized training programs

Adapt IP laws to support open licensing models

Engage the general public and stakeholders

You can follow along with the work of Columbia Convening at mozilla.org/research/cc and the work from the Open Source Initiative on the definition of Open Source AI at opensource.org/deepdive.

Source: opensource.org