Skip to main content

Andreas Liesenfeld, Assistant Professor, Radboud University

State of Open: The UK in 2024

Phase Two: The Open Manifesto Report

 

 

Thought Leadership: Towards meaningful open AI: key findings and societal implications

The past year has seen a steep rise in generative AI systems that claim to be open. But how open are they really? The question of what counts as open source in generative AI takes on particular importance in light of the EU AI Act that regulates ‘open source’ models differently, creating an urgent need for practical openness assessment. 

Recent work by Andreas Liesenfeld and Mark Dingemanse (Centre for Language Studies, Radboud University, The Netherlands) provides a stark view of the openness of current generative AI. In a systematic sweep of the landscape, they survey 45 text and text-to-image models that bill themselves as open. Key findings and contributions include the following:

  1. Open-washing is widespread. Open-washing is the use of terms like ‘open’ and ‘open source’ for marketing purposes without actually providing meaningful insight into source code, training data, fine-tuning data or architecture of systems. Corporations like Meta, Mistral and Microsoft regularly co-opt terms like ‘open’ and ‘open source’ while shielding most of their models from scrutiny.
  1. The EU AI Act puts legal weight on the term ‘open source; without clearly defining it, creating an incentive for open-washing, opening up a key pressure point for lobbying, and necessitating clarity about what constitutes openness in the domain of generative AI.
  1. Openness in Generative AI can be meaningfully measured if we think of it as composite (consistent of multiple elements) and gradient (it comes in degrees). The work distinguishes 14 dimensions of openness, from training datasets to scientific and technical documentation and from licensing to access methods.
  1. Meaningful openness is possible, and is exemplified by some of the smaller players in the field, who go the extra mile to document their systems and open them up to scrutiny. We identify organisations like AllenAI (with OLMo) and BigScience Workshop + HuggingFace (with BloomZ) as key players moving the needle of openness in generative AI.

Why this matters for everybody

Although the EU AI Act creates a particular sense of urgency, openness is in fact of key importance for innovation and science, as well as to society. Here are three key points why openness in generative AI is important to society at large:

  1. Openness helps build critical AI literacy by demystifying capabilities. If a company claims their AI can ‘pass the bar exam’, it helps to know exactly what was in the training data. The sheer magnitude of training data for GPT4 means that it can effectively take the bar exam ‘open book’, which is much less impressive. As AI applications are fast becoming ubiquitous, openness about training data and fine-tuning routines helps us equip the general public with a critical understanding.
  1. Openness grants agency to people subjected to automated decisions. A clear example is the LAION image dataset underlying a widely used image generator. The openness of this dataset enabled auditing by Dr. Birhane and others which brought to light the presence of problematic material and prompted revisions and retractions. pre that AI is often best seen as a shorthand for automation. Transparency about training data and model architecture makes models auditable, and can help to make biases visible.
  1. Openness provides real benefits on the ground. Corporations like to hand-wave at ‘AI safety’ as a reason to keep systems under wraps, but this is mostly a thinly disguised attempt to obscure clear and present harms already underway, including increasing spam content in search engines and the rapid spread of misinformation. Only open systems can be responsibly used in science and education and can support a healthy open source ecosystem.

Evidence-based openness assessment can help foster a generative AI landscape in which models can be effectively regulated, model providers can be held accountable, scientists can scrutinise generative AI, and people can make informed decisions for or against using generative AI.

First published by OpenUK in 2024 as part of State of Open: The UK in 2024 Phase Two “The Open Manifesto”

© OpenUK 2024

Download Thought Leadership

Download Report

View all

Scroll to top of the content