The Tragedy of AI (and what to do about it)

Image adapted from Tragedy and Comedy by Tim Green CC BY

A tragedy is characterized by four things:

  1. A heroic noble protagonist

  2. Fatal flaw(s)

  3. An inevitable downfall

  4. Catharsis

Heroic Noble AI image adapated fom Super Heros by 5chw4r7z CC BY
Fatal Flaws image adapted from They Always Focus On Our Flaws by Theilr CC BY-SA
Inevitable Downfall image adapted from The Downfall Of Machismo by Tman CC BY-NC-SA
Catharsis image adapted from Catharsis by Jason Jacobs CC BY

I’m doing a presentation titled “The Tragedy of AI (and what to do about it)” Monday June 23, 2025 at the Association of Learning Technology (ALT) Open Education Conference (OER25) in London England. The theme of this event is “Speaking Truth to Power: Open Education and AI in the Age of Populism”.

To complement and accompany my presentation I’ve written this essay blog post in two parts.

Part one, “The Tragedy of AI” establishes why AI is a tragedy. It reveals fatal flaws of AI. It shows how these fatal flaws risk leading to tragic inevitable downfall. This Tragedy of AI essay specifically focuses on the negative impact AI is having on the social agreements and fundamental principles that underlie open education, open science, and other open knowledge practices.

Part two, “Catharsis” explores the “(and what to do about it)” part of the Tragedy of AI. It puts forward options for taking action to address or prevent the tragedy of AI. It defines actions in ways that fit with open and deal with the fatal flaws of AI in ways that mitigate downfall.

PART ONE - The Tragedy of AI

AI as Heroic Noble Protagonist

 

Much has been written about the exceptional capabilities of AI. The OER25 conference program is chock a block full of such sessions. This post doesn’t delve into that, enough is being said.

The focus of this post is instead on the other three characteristics of a tragedy - fatal flaws, inevitable downfall, and catharsis.

The Fatal Flaws of AI

 

There are five fatal flaws of AI:

  1. Tragedy of the Capitalism over the Commons

  2. Taking without consent

  3. Demotivating sharing

  4. Breaking reciprocity, and

  5. Bleeding open dry

Lets dig in to each of them.

AI Fatal Flaw #1 - Tragedy of the Capitalism over the Commons

Everyone knows, as Darcy Norman puts it,

“GenAI tools, based on Large Language Models (LLMs), are built by extracting content from across the internet—often without clear consent or compensation. This extraction-based origin demands caution.”

I put it a bit more forcefully. AI as a robbing, a plunder, a pillaging of a public commons - all of the Internet. In creating their Large Language Models developers of AI have helped themselves to all the content you and I have on the Internet. AI as an appropriation of a commons with little to no regard for the commons itself or the social norms associated with it.

Robin Sloan in his essay “Is it okay?” says:

“How do you make a language model? Goes like this: erect a trellis of code, then allow the real program to grow, its development guided by a grueling training process, fueled by reams of text, mostly scraped from the internet. Now. I want to take a moment to think together about a question with no remaining practical importance, but persistent moral urgency:

Is that okay?

The question doesn’t have any practical importance because the AI companies — and not only the companies, but the enthusiasts, all over the world — are going to keep doing what they’re doing, no matter what.

The question does still have moral urgency because, at its heart, it’s a question about the things people all share together: the hows and the whys of humanity’s common inheritance. There’s hardly anything bigger.”

In 1968 Joseph Hardin wrote the “Tragedy of the Commons”. In it he asks us to:

“Picture a pasture open to all. It is to be expected that each herdsman will try to keep as many cattle as possible on the commons. Such an arrangement may work reasonably satisfactorily for centuries because tribal wars, poaching, and disease keep the numbers of both man and beast well below the carrying capacity of the land.”

But what happens when the capacity of the land is met? Hardin says;

“The rational herdsman concludes that the only sensible course for him to pursue is to add another animal to his herd. And another; and another.... But this is the conclusion reached by each and every rational herdsman sharing a commons. Therein is the tragedy. Each man is locked into a system that compels him to increase his herd without limit-in a world that is limited.”

The effects of this self-interest over group interest Hardin suggests are overgrazing and a destruction of the commons to the detriment of all.

Louis Villa in his "My commons roller coaster" says:

"Whether or not the web was a “commons” is complicated, so I don’t want to push that label too hard. (As just one example, there was not much boundary setting—one of Ostrom’s signals of a traditional common pool resource.) But whether or not it was, strictly speaking, a “commons”, it was inarguably the greatest repository of knowledge the world had ever seen. Among other reasons, this was in large part because the combination of fair use and technical accessibility had rendered it searchable. That accessibility enabled a lot of good things too—everything from language frequency analysis to the Wayback Machine, one of the great archives of human history.

But in any case it’s clear that those labels, if they ever applied, very much merit the past tense. Search is broken; paywalls are rising; and our collective ability to learn from this is declining. It’s a little much to say that this paper is like satellite photos of the Amazon burning... but it really does feel like a norm, and a resource, are being destroyed very quickly, and right before our eyes."

Image adapted from of Global Carbon Monoxide by NASA CC BY

AI is a tragedy based on AI developers assuming everything on the web is there for the taking, acting autonomously, out of pure self-interest without interaction or consideration of the commons or the people that make that commons. In doing so AI negatively impacts practices of sharing and openness risking the destruction of the very commons on which it relies.

Anna Tumadóttir in her article “Reciprocity in the Age of AI” says:

“What we ultimately want, and what we believe we need, is a commons that is strong, resilient, growing, useful (to machines and to humans)—all the good things, frankly. But as our open infrastructures mature they become increasingly taken for granted, and the feeling that “this is for all of us” is replaced with “everyone is entitled to this”. While this sounds the same, it really isn’t. Because with entitlement comes misuse, the social contract breaks, reciprocation evaporates, and ultimately the magic weakens. 

Hardin’s Tragedy of the Commons intuitively resonates with us. Air pollution, over grazing, deforestation, climate change and other global challenges align with his notion that people always act out of selfish self interest. But the tragedy of the commons is a concept, a theory, and as Elinor Ostrom’s Nobel prize winning work showed, it is not how a commons works in practice.

The reality is that AI is not a Tragedy of the Commons. It is a Tragedy of Capitalism. It is a tragedy of self-interest and the belief in competition over cooperation. It is a tragedy of the relentless pursuit of growth and profit no matter what the cost.

In part two “Catharsis” I’ll describe some of the ways Ostrom discovered people mitigate the potential for a tragedy of the commons and explore how we can apply these ideas to AI and our open work.

For those of you interested in exploring more about the differences between how a commons works and how capitalism works I highly recommend these resources:

The Serviceberry - Abundance and Reciprocity in the Natural World by Robin Wall Kimmerer. This beautiful book published in 2024 is an eloquent exploration of the issue based on lessons from the natural world.

The Gift - How the Creative Spirit Transforms the World by Lewis Hyde. This book explores old gift-giving cultures and their relation to commodity societies.

Doughnut Economics - Seven Ways to Think Like a 21st Century Economist by Kate Raworth. Doughnut economics, is a visual framework for sustainable development. The framework is shaped like a doughnut, hence its name. It explores the creation of an alternate economy that meets the needs of people without overshooting Earth's ecological ceiling re-framing economic problems in the process.

Made With Creative Commons by Paul Stacey and Sarah Hinchliff Pearson. While working at Creative Commons I co-wrote this book with Sarah. It is guide to sharing your knowledge and creativity with the world, and sustaining your operation while you do.

AI Fatal Flaw #2 - Taking Without Consent
As Daniel Campos notes,

AI is constructed non-consensually on the back of copyrighted material

This is one of the greatest stains on this technology, period.

People’s work was taken, without consent, and it is now being used to generate profit for parties visible and not.

Thing is, this is unsustainable for all parties. The technology requires ongoing training to remain relevant. Everyone who is part of its value chain, in the long term, must be represented in both its profits and the general framework for participation.

AI developers did not seek permission. They mined any works they wanted. No care was taken as to whether the works are under copyright or have an open license. Everything on the Internet, content of all kinds, posts, books, journal articles, music, … it didn’t matter, AI developers view all such resources as legitimate sources of data for its models. The more data mined the better the AI model.

As Johana Bhuiyan notes in “Companies building AI-powered tech are using your posts” reports:

“"Even if you haven’t opted in to letting them use your data to train their AI, some companies have opted you in by default. Figuring out how to stop them from hoovering up your data to train their AI isn’t exactly intuitive. Left to their own devices, more than 300,000 Instagram users have posted a message on their stories in recent days stating they do not give Meta permission to use any of their personal information to inform their AI. To be clear, just like the Facebook statuses of yore, a simple Instagram post will not do anything to stop Meta from using your data in this way.

This is not a new practice as Bhuiyan notes:

“Default opt-ins are an industry-wide issue. A recent report by the Federal Trade Commission (FTC) on the data practices of nine social media and streaming platforms including WhatsApp, Facebook, YouTube and Amazon found that nearly all of them fed people’s personal information into automated systems with no comprehensive or transparent way for users to opt out.

Overall, there was a lack of access, choice, control, transparency, explainability and interpretability relating to the companies’ use of automated systems,” the FTC report reads.”

AI developers are facing lawsuits over their use of content without consent. Creators are advocating for consent, credit and compensation. AI developers seem to be speaking out of two sides of their mouths on this issue. On the one side they are claiming their use of content is allowed under copyright fair use (in the US) or text and data mining rules. On the other they are simultaneously pursuing licensing deals with major media rights holders.

Consent to use is not just about data scraped from the web. It’s also about the data associated with your use of the AI itself. Sadly the business models of AI are further entrenching have’s vs have nots. If you are using a free version of an AI application your prompts and outputs are, by default, being used to train the AI. If you are using paid versions of an AI application your prompts and outputs are not being used to train the AI. For example Google’s Gemini AI terms of use says this:

“When you use Paid Services, including, for example, the paid quota of the Gemini API, Google doesn't use your prompts (including associated system instructions, cached content, and files such as images, videos, or documents) or responses to improve our products.”

“When you use Unpaid Services, including, for example, Google AI Studio and the unpaid quota on Gemini API, Google uses the content you submit to the Services and any generated responses to provide, improve, and develop Google products and services and machine learning technologies, including Google's enterprise features, products, and services.”

So if you can afford to pay to use AI your prompts and outputs aren’t used to train the AI. If you are less well off and can’t afford to pay then you pay by agreeing to have your data used.

AI Fatal Flaw #3 - Demotivating Sharing

Sharing is central to open practices whether that be open education, open science, open data, open culture or any of the many other forms of open. Creative Commons licenses have been a key means by which sharing of content has been enabled.

However, as Anna Tumadóttir, CEO of Creative Commons, says in her article "Reciprocity in the Age of AI":

“Since the inception of CC (Creative Commons), there have been two sides to the licenses. There’s the legal side, which describes in explicit and legally sound terms, what rights are granted for a particular item. But, equally there’s the social side, which is communicated when someone applies the CC icons. The icon acts as identification, a badge, a symbol that we are in this together, and that’s why we are sharing. Whether it’s scientific research, educational materials, or poetry, when it’s marked with a CC license it’s also accompanied by a social agreement which is anchored in reciprocity. This is for all of us.

But, with the mainstream emergence of generative AI, that social agreement has come into question and come under threat, with knock-on consequences for the greater commons. Current approaches to building commercial foundation models lack reciprocity. No one shares photos of ptarmigans to get rich, no one contributes to articles about Huldufólk​ seeking fame. It is about sharing knowledge. But when that shared knowledge is opaquely ingested, credit is not given, and the crawlers ramp up server activity (and fees) to the degree where the human experience is degraded, folks are demotivated to continue contributing.”

Sadly AI demotivates sharing.

Image adapted from Sharing With Others by Frank Homp CC BY

AI Fatal Flaw #4 - Breaking Reciprocity

At this point AI is a permissionless, acquisition and assertion of the right to use the works of others. It is a utilization of others works for profit without reciprocity. By breaking with the reciprocal social exchange principle that involves the giving and receiving of benefits or actions between individuals AI engages in negative reciprocity gaining more out of the exchange than the public and naturally people feel wronged.

Image adapted from Reciprocity by Keith Solomon CC BY-NC-SA

AI developers do not seek permission. They mine any works they want. They don’t care whether the works used are under copyright or have an open license. Everything on the Internet, books, journal articles, music, … it doesn’t matter, all such resources are considered legitimate sources of data for its models.

As Frank Pasquale and Haochen Sun note,

“Most AI firms are not compensating creative workers for composing the songs, drawing the images, and writing both the fiction and nonfiction books that their models need in order to function. AI thus threatens not only to undermine the livelihoods of authors, artists, and other creatives, but also to destabilize the very knowledge ecosystem it relies on.”

They go on to note that legislators agree,

“The situation strikes many policymakers as deeply unfair and undesirable. As the Communications and Digital Committee of the United Kingdom’s House of Lords has concluded, “[w]e do not believe it is fair for tech firms to use rightsholder data for commercial purposes without permission or compensation, and to gain vast financial rewards in the process.”

And this is not just an issue for proprietary works. Openly licensed and peer reviewed sources of data are considered high quality and AI prizes such data sources. But AI violates and degrades the integrity of open sharing by taking without giving. As Stephanie Decker notes in The Open Access – AI Conundrum: Does Free to Read Mean Free to Train?,

"The central issue remains that commercial AI companies extract significant economic value from OA content without necessarily returning value to the academic ecosystem that produced it, while at the same time disrupting academic incentive structures and attribution mechanisms."

Creative Commons CEO Anna Tumadóttir says,

“Reciprocity in the age of AI means fostering a mutually beneficial relationship between creators/data stewards and AI model builders. For AI model builders who disproportionately benefit from the commons, reciprocity is a way of giving back to the commons that is community and context specific.

(And in case it wasn’t already clear, this piece isn’t about policy or laws, but about centering people).

This is where our values need to enter the equation: we cannot sit neutrally by and allow “this is for everyone” to mean that grossly disproportionate benefits of the commons accrue to the few. That our shared knowledge pools get siphoned off and kept from us.

We believe reciprocity must be embedded in the AI ecosystem in order to uphold the social contract behind sharing. If you benefit from the commons, and (critically) if you are in a position to give back to the commons, you should. Because the commons are for everyone, which means we all need to uphold the value of the commons by contributing in whatever way is appropriate.

There never has been, nor should there be, a mandatory 1:1 exchange between each individual and the commons. What’s appropriate then, as a way to give back?”

We’ll explore answers to this question in the second part of this essay “Catharsis”.

AI Fatal Flaw #5 - Bleeding Open Dry

Molly White in her post “Wait, not like that”: Free and open access in the age of generative AI” says,

“The real threat isn’t AI using open knowledge — it’s AI companies killing the projects that make knowledge free. The true threat from AI is that they will stifle open knowledge repositories benefiting from the labor, money, and care that goes into supporting them while also bleeding them dry. It’s that trillion dollar companies become the sole arbiters of access to knowledge after subsuming the painstaking work of those who made knowledge free to all, killing those projects in the process.”

AI does not acknowledge, care for, or contribute to the stewardship of the commons on which it relies.

Drain hole in concrete with patched cracks

Bleeding Open Dry image adapted from Pavement Drain With Patched Cracks photo by Paul Stacey CC BY-SA

Molly goes on to say:

“Anyone at an AI company who stops to think for half a second should be able to recognize they have a vampiric relationship with the commons. While they rely on these repositories for their sustenance, their adversarial and disrespectful relationships with creators reduce the incentives for anyone to make their work publicly available going forward (freely licensed or otherwise). They drain resources from maintainers of those common repositories often without any compensation. They reduce the visibility of the original sources, leaving people unaware that they can or should contribute towards maintaining such valuable projects. AI companies should want a thriving open access ecosystem, ensuring that the models they trained on Wikipedia in 2020 can be continually expanded and updated. Even if AI companies don’t care about the benefit to the common good, it shouldn’t be hard for them to understand that by bleeding these projects dry, they are destroying their own food supply.

And yet many AI companies seem to give very little thought to this, seemingly looking only at the months in front of them rather than operating on years-long timescales.”

AI is vampiric text overlay on artistic rendering of a vampire

Vampire by Alvaro Tapia licensed CC BY-NC-ND

AI is vampiric sucking the lifeblood out of the corpus of all human knowledge and selling it back to us as a monthly subscription.

Techbro: “We stole your data to train our new AI model” User: 🤨  Techbro: “would you like to rent it back from us?” User: 😡  Variation of TWONK comic by: UK based cartoonist and illustrator Steve Nelson

Image adapted from Hacker Memes based on original Twonks comic by Steve Nelson

The Inevitable Downfall

 

The fatal flaws of AI lead to an inevitable downfall. The very things AI relies on:

  • The Commons

  • Consent and willing participation

  • Sharing of high quality, thriving, evolving data

  • Reciprocity

  • Openness

are its fatal flaws.

  1. Tragedy of Capitalism over the Commons

  2. Taking without consent

  3. Demotivating sharing

  4. Breaking reciprocity, and

  5. Bleeding open dry

The fatal flaws not only have the potential for an inevitable downfall of AI but the downfall of open education, open science and other open practices too.

Every time I use AI I feel like I am aiding and abetting this tragedy and downfall. In my use I condone its taking.

The current issues around ChatGPT and Studio Ghibli describe it well. Shanti Escalante-De Mattei says in her article “The ChatGPT Studio Ghibli Trend Weaponizes “Harmlessness” to Manufacture Consent”:

“Every time a user transforms a selfie, family photo or cat pic into a Ghibli-esque image, they normalize the ability of AI to steal aesthetics, associations, and affinities that artists spend a lifetime building. Participatory propaganda doesn’t require users to understand the philosophical debates about artists’ rights, or legal ones around copyright and consent. Simply by participating they help Altman and his competitors win the battle of public opinion. For many, the Ghibli images will be their first contact with generative AI. For more informed users, the flood of images reinforces the inevitability that AI will re-shape our world. It’s particularly cruel that Miyazaki’s style has become the AI vanguard, given that he famously stuck to laborious hand-drawn animation even as the industry shifted to computer-generated animation.”

She concludes:

"It is laborious, loving, useful work that imparts self-knowledge and the understanding of shared struggle and humanity. You cannot skip to the end. You cannot just generate the artwork or the essay and get the learning and the satisfaction that the work imparts. When you forgo that labor, you let fascists, corporations, and all manner of elite actors make the act of expressing yourself so easy that you feel free, even agentive, as they take your rights from you."

I have used AI. But this is not the kind of relationship I want. I’m not a fan of the enshitification of the Internet and much of AI is following the enshitification model flooding us with slop (deepfakes, misinformation, personality appropriation, bias, … - need I go on?) and exploitive business models.

Oh sure there have been publications about the ethics of AI and attempts to regulate (at least in the EU) but to date there has been no effective repercussion to the tragedy of AI. There is no change in behaviour no adjustment of course. If anything, there is a race to commercialization, economic growth, and global dominance - regulations, guardrails, social agreements be damned.

AI’s trampling of the commons puts it at odds with open.

Should our response simply be one of acceptance?

Everybody Knows but nobody does anything?

Is there no course of action to address or prevent the tragedy of AI?

I think there are lots of emerging options.

I have hope.

So lets move on to “Catharsis”.

PART Two - Catharsis

 

In this Catharsis section we shift from the Tragedy of AI to the “(and what to do about it)” part. What are options for taking action to address or prevent the tragedy of AI? What actions can the open community take to deal with the fatal flaws of AI in ways that mitigate downfall?

Anna Tumadóttir, CEO of Creative Commons says:

“There never has been, nor should there be, a mandatory 1:1 exchange between each individual and the commons. What’s appropriate then, as a way to give back? So many possibilities come to mind, including:

  • Increasing agency as a means to achieve reciprocity by allowing data holders to signal their preferences for AI training 

  • Credit, in the form of attribution, when possible

  • Open infrastructure support

  • Cooperative dataset development

  • Putting model weights or other components into the commons

Catharsis can take many forms.

Building on Anna’s list here are five catharsis actions I think the open community can pursue to resolve the fatal flaws of AI and avoid the tragedy:

  1. Social Agreements & Regulations

  2. Preference Signals

  3. Reciprocity

  4. Public AI

  5. Collaborative open data set creation

Lets explore each of these.

AI Catharsis #1 - Social Agreements & Regulations

Samuel Moore in "Governing the Scholarly Commons" says:

"The question, then, is not whether AI is inherently good or bad but more concerning who controls it and with what motivation. If it is answerable to affected communities, ethicists and technological experts, AI may develop in a more productive way than if it is governed by the needs of shareholders and profit-seeking companies. The problem of AI governance – in the context of academic knowledge production – is the focus of this report."

He goes on to suggest:

"The governance of knowledge production is currently weighted heavily in favour of the market, which is to say that decisions to implement a technology or business model are determined by how profitable they are, how much labour they save, or how financially efficient they are. I am therefore interested in ways to keep the power of the market in check in the service of more responsible AI development."

Social Agreements and Regulations image adapted from Agreement by Takashi Toyooka CC BY-NC

Moore's report looks at a range of different strategies for governing the implementation of AI as a scholarly commons. Options for good governance of AI in academic knowledge production include:

  • Commons-based approaches

  • Academic governance

  • Academic "citizen" assemblies

These all have merit.

Commons-based approaches have a high fit with open education and open science.

Nobel prize winning Elinor Ostrom studied many different commons. She found that the tragedy of the commons is not inevitable, as Hardin thought. Instead, if commons participants decide to cooperate with one another, communicate, monitor each other’s use, and enforce rules for managing it, they can avoid the tragedy.

Open education and open science have been functioning in just such a cooperative manner but AI disrupts that practice.

It is in the interest of AI developers and the open community to work together in resolving the fatal flaws and avoid a shared tragedy.

What we need is dialogue, communication and relationship building between the open community and AI developers. There are lots of AI applications targeting education. Their development efforts are proceeding with little to no engagement with the open education community. Open education could step up and take an active leading role in seeking to communicate and influence how AI development for education takes place. But we can’t do that in isolation from the AI developers themselves.

We could for example, seek to work with efforts like Current AI who have a focus area specifically relating to Open supporting open standards and tools that make AI accessible and adaptable for everyone.

Alek Tarkowski in his article "Data Commons Can Save Open AI" says:

"We must do everything in our power to ensure that future datasets are built upon a data commons with stewardship and control."

His recently released report, “Data Governance in Open Source AI- Enabling Responsible and Systemic Access” argues that collective action is needed to release more data and improve data governance to balance open sharing with responsible release.

Consortia of open organizations could collaborate and propose solutions to the fatal flaws. Everyone acknowledges the need for AI guardrails. Clearly rules and agreements are needed. But these can come in many forms ranging from social norms, to policy, and regulations. How would the open community define these?

Over one hundred countries signed on to the "Statement on Inclusive and Sustainable Artificial Intelligence for People and the Planet" committing to:

"Initiate a Global Dialogue on AI governance and the Independent International Scientific Panel on AI and to align on-going governance efforts, ensuring complementarity and avoiding duplication."

How might the open community participate and provide input into such efforts?

Governments are responsible for regulations with the EU AI regulation being the prime example to date. Perhaps AI, developed with public funds, ought to be open to the public that paid for it in a similar manner as with open science and open education?

Over time the legality related to copyright lawsuits AI faces will emerge. But it will take a long time and it is highly politicized as highlighted in Copyright Office head fired after reporting AI training isn’t always fair use. Nevertheless it seems likely there will be new rules that seek to strike some kind of effective balance between the public interests, maintaining a thriving creative community and allowing technological innovation to flourish. It seems prudent that we make an effort to define what that looks like for open education and open science.

One option is for us to define norms and social agreements. We've seen some related action around efforts to define what "open source" means in the context of AI models. But to date there is no similar effort around defining what open means in the context of AI in open education and open science. There are lots of efforts around ethics in relation to AI but not so much around open.

In Governing the Commons, Elinor Ostrom summarized eight design principles associated with creating a sustainable commons:

1. Clearly defined boundaries

2. Congruence between appropriation and provision rules and local conditions

3. Collective-choice arrangements

4. Monitoring

5. Graduated sanctions

6. Conflict-resolution mechanisms

7. Minimal recognition of rights to organize

8. Nested enterprises

These design principles are worth considering as a framework for devising open norms and agreements related commons based approaches to AI.

But there are many more options.

Molly White in her article “Wait, not like that”: Free and open access in the age of generative AI, says,

“It would be very wise for these [AI] companies to immediately begin prioritizing the ongoing health of the commons, so that they do not wind up strangling their golden goose. It would also be very wise for the rest of us to not rely on AI companies to suddenly, miraculously come to their senses or develop a conscience en masse.

Instead, we must ensure that mechanisms are in place to force AI companies to engage with these repositories on their creators' terms.

There are ways to do it: models like Wikimedia Enterprise, which welcomes AI companies to use Wikimedia-hosted data, but requires them to do so using paid, high-volume pipes to ensure that they do not clog up the system for everyone else and to make them financially support the extra load they’re placing on the project’s infrastructure. Creative Commons is experimenting with the idea of “preference signals” — a non-copyright-based model by which to communicate to AI companies and other entities the terms on which they may or may not reuse CC licensed work. Everyday people need to be given the tools — both legal and technical — to enforce their own preferences around how their works are used.

Some might argue that if AI companies are already ignoring copyright and training on all-rights-reserved works, they'll simply ignore these mechanisms too. But there's a crucial difference: rather than relying on murky copyright claims or threatening to expand copyright in ways that would ultimately harm creators, we can establish clear legal frameworks around consent and compensation that build on existing labor and contract law. Just as unions have successfully negotiated terms of use, ethical engagement, and fair compensation in the past, collective bargaining can help establish enforceable agreements between AI companies, those freely licensing their works, and communities maintaining open knowledge repositories. These agreements would cover not just financial compensation for infrastructure costs, but also requirements around attribution, ethical use, and reinvestment in the commons.

The future of free and open access isn't about saying “wait, not like that” — it’s about saying "yes, like that, but under fair terms”. With fair compensation for infrastructure costs. With attribution and avenues by which new people can discover and give back to the underlying commons. With deep respect for the communities that make the commons — and the tools that build off them — possible. Only then can we truly build that world where every single human being can freely share in the sum of all knowledge.”

I think this entire list of suggestions is a good harbinger of what is to come. Of particular note is the shift from using copyright as the primary legal framework for open to one that uses labor and contract law to enact AI social agreements and regulations.

AI Catharsis #2 - Signal Preferences

I think signal preferences have a lot of potential.

Abstact art representing smoke signals

Signal Preferences image adapted from Smoke Signals by Ashok Boghani CC BY-NC

As Molly noted, Creative Commons (CC) is experimenting with the idea of preference signals. In 2024 a Creative Commons Position Paper on Preference Signals was released that gives a good overview of what is being considered. This paper notes:

What is new with generative AI is that the unanticipated uses are happening at scale. With concentrated power there is a risk of concentrated benefits and creators are questioning anew whether the bargain is worth it. The creative works are being used outside of their original context in a way that does not distribute any of the usual rewards back to the creator, either financial or reputational.”

One signal preferencing approach focuses on making it possible to signal opt-out vs opt-in. By default AI developers automatically opted us all in to their use of our data. This could be changed by regulation. But in the interim there is interest in a technical opt-out signalling preference that allows creators to prevent their content from being used for AI training.

The most promising solution seems to be Robots.txt. Audrey Hingle and Mallory Knodel note in their article "Robots.txt Is Having a Moment: Here's Why We Should Care" note:

“Robots.txt remains important as a foundational tool due to its widespread adoption and familiarity among website owners and developers. It provides a straightforward mechanism for declaring basic crawling permissions, offering a common starting point from which more advanced and specific solutions can evolve.”

However, they go on to note challenges:

“Robots.txt is primarily useful for website owners and publishers who control their own domains and can easily specify crawling rules. It doesn't effectively address content shared across multiple platforms or websites, nor does it give individual content creators, such as artists, musicians, writers, and other creative professionals, a way to easily communicate their consent preferences when they publish their work on third party sites, or when their work is used by others.”

Opt-in vs opt-out has largely focused on providing a way for websites and end users to opt out of having AI use their data for training. I’d love to see opt-in options that give users options for opting in to having their data used for AI that serves the public interest rather than for big tech for-profit shareholders.

Creative Commons is taking a different approach from opt-in vs opt-out. In their Position Paper on Preference Signals they say:

“CC’s approach is to reject the all-or-nothing framework and create options for sharing that reflect a more generous and collaborative spirit than default copyright. The CC licenses exist on a spectrum of permissiveness, all underpinned by the goal of enabling access to and sharing of knowledge and creativity as part of a global commons, built on mutual cooperation and shared values.”

They go on to note:

“We have uncovered many of the limitations of using instruments such as robots.txt as an indicator of opt-in or opt-out for generative AI training. In many cases, robots.txt and a website’s terms of service are inconsistent, and robots.txt is a limiting protocol when it comes to creator content in the commons as a public good (including, but not limited to art, culture, science, journalism, scientific data). Further, approaches that propagate the limiting binary of blunt instruments of opt-out do not take into consideration the values and social norms embedded in sharing content on the web. CC’s approach is to develop and advocate for tools that empower creators and contribute to a healthy and ethical commons for the public good.”

Creative Commons has made a lot of progress on this and on June 25, 2025 they are hosting a CC Signals Kickoff event. There invitation to this event says:

“We are building a standardized, global mechanism called CC signals that aims to increase the agency of those creating and stewarding the content that is relied upon for AI training. We invite you to join us as we officially kick off the CC signals project. During this kickoff event, you’ll hear from members of the CC team and community who will share an outline of the first public proposal of the CC signals framework. We will also provide resources that give background information and explore early thinking on legal and technical implementation of CC signals. Our goal is to set up members of the CC global community to engage with this proposal so that you can provide input and recommendations that will strengthen the initiative as we build toward pilot implementation later this year. This is a shared challenge, and a shared opportunity. Whether you're a funder, developer, policymaker, educator, platform operator, or creator, your participation matters. Join us.”

Anna CEO of Creative Commons says:

“Part of CC being louder about our values is also taking action in the form of a social protocol that is built on preference signals, a simple pact between those stewarding data and those reusing it for generative AI. Like CC licenses, they are aimed at well-meaning actors and designed to establish new social norms around sharing and access based on reciprocity. We’re actively working alongside values-aligned partners to pilot a framework that makes reciprocity actionable when shared knowledge is used to train generative AI.”

A key emphasis is that these preference signals are focused on reciprocity. I think we’ll see preferences address reciprocity in ways that involve direct or indirect financial subsidization, mutual exchange non-monetary contribution, and credit/attribution recognition.

One of the best things about existing CC licenses is their use of icons and human readable deeds. I expect these signal preferences to follow that successful format with icons that symbolize preferences and easy to read deeds that clearly specify the terms of the social agreement.

AI Catharsis #3 - Reciprocity

This is, perhaps, the single biggest fatal flaw of AI. I applaud CC’s efforts to address it.

Given the volume of data used to train AI models the value of any one creators data is likely very small. One precedent for thinking about this is music streaming where creators get a small fraction of a cent per stream. However, as seen by AI licensing deals, AI developers want to work more with large collection holders rather than individual creators. Reciprocity at the aggregate level rather than at the individual level.

Paul Keller in "AI, the Commons, and the limits of copyright" has a novel suggestion:

“We should look for a new social contract, such as the United Nations Global Digital Compact, to determine how to spend the surplus generated from the digital commons. A social contract would require any commercial deployment of generative AI systems trained on large amounts of publicly available content to pay a levy. The proceeds of such a levy system should then support the digital commons or contribute to other efforts that benefit humanity, for example, by paying into a global climate adaptation fund. Such a system would ensure that commercial actors who benefit disproportionately from access to the “sum of human knowledge in digital, scrapable form” can only do so under the condition that they also contribute back to the commons.”

Network image showing dots connected in reciprocal exchange

Image adapted from Reciprocity by Keith Solomon CC BY-NC-SA

If AI is going to take and benefit from the commons it surely ought to contribute back to it? Do we, the open community, have ideas for how AI can contribute back to the commons?

The open community tends to openly license and share resources not with the expectation of personal financial gain but with the aim of enabling a commons that benefits all. What can AI developers contribute to a commons benefitting us all?

One consideration is credit and attribution. Creators of all types value reputation gain through acknowledgement and recognition. This is especially so in open where open science and open education communities use Creative Commons licenses that all mandate attribution. More broadly, citations and references are a core and essential practice in education and research.

By not providing credit and attribution AI breaks the reciprocal social norms of academia. This is one area where academia is well positioned to push back and advocate for fully referenced AI outputs. Credit, in the form of attribution, should be the default. At the very least the datasets on which the model is trained should be fully disclosed.

AI Catharsis #4 - Public AI

Another form reciprocity could take is support for open infrastructure that supports the commons more broadly. Open infrastructures are technologies, services, and resources provided and managed by non-commercial organizations. A big issue with AI is the complexity of the technology stack involved with developing and using AI applications. Reciprocity could entail creation of alternatives to for-profit AI in the form of open infrastructure for Public AI.

Open infrastructures are technologies, services, and resources provided and managed by non-commercial organisations. A big issue with AI is the complexity of the technology stack involved with developing and using AI applications. Open infrastructure can establish an alternative to For-profit AI in the form of Public AI.

In a recently published "Public AI - A Public Alternative to Private AI Dominance" blog post and paper Alek Tarkowski, Felix Sieker, Lea Gimpel, and Cailean Osborne write:

"Today’s most advanced AI systems and foundation models are largely proprietary and controlled by a small number of companies. There is a striking lack of viable public or open alternatives. This gap means that cutting-edge AI remains in the hands of a select few, with limited orientation toward the public interest, accountability or oversight.

Public AI is a vision of AI systems that are meaningful alternatives to the status quo. In order to serve the public interest, they are developed under transparent governance, with public accountability, equitable access to core components (such as data and models), and a clear focus on public-purpose functions."

Public AI serves the public interest. This seems common ground with open education and open science.

Public AI image adapted from What Is For the Public Good? photo by Paul Stacey CC BY-SA

The report goes on to say:

A vision for public AI needs to take into account today’s constraints at the compute, data and model layers of the AI stack, and offer actionable steps to overcome these limitations. This white paper offers a clear overview of AI systems and infrastructures conceptualized as a stack of interdependent elements, with compute, data and models as its core layers.

It also identifies critical bottlenecks and dependencies in today’s AI ecosystem, where dependency on dominant or even monopolistic commercial solutions constrains development of public alternatives. It highlights the need for policy approaches that can orchestrate resources and various actors across layers, rather than attempting complete vertical integration of a publicly owned solution.

To achieve this, it proposes three core policy recommendations:

  1. Develop and/or strengthen fully open source models and the broader open source ecosystem

  2. Provide public compute infrastructure to support the development and use of open models

  3. Scale investments in AI capabilities to ensure that sufficient talent is developing and adopting these models”

All these ideas seem pertinent to our open work. I also think Public AI gives users an opt-in option alternative to big tech. We should be thinking about how to have our work be Digital Public Goods.

Reciprocity remains an AI fatal flaw but there are options. How might all of us working in open education and open science pursue and shape these options?

AI Catharsis #5 - Collaborative open data set creation

One of the ways the open community can take action is around collaborative open data set creations as a form of Digital Public Goods.

Dew drops on sunlit spider web

Sunrise in Dew Drops on Spider Web by Matthew Paulson CC BY-NC-ND

A fascinating example is described in “Common Corpus: building AI as Commons” where Alek Tarkowski and Alicja Peszkowska share a profile of Pierre-Carl Langlais, a digital humanities researcher, Wikipedian, and a passionate advocate for open science. Pierre is also the co-founder of a French AI startup, Pleias, and the coordinator of Common Corpus, a public domain dataset for training LLMs.

They note:

“As the largest training data set for language models based on open content to date, Common Corpus is built with open data, including administrative data as well as cultural and open-science resources – like CC-licensed YouTube videos, 21 million digitized newspapers, and millions of books, among others. With 180 billion words, it is currently the largest English-speaking data set, but it is also multilingual and leads in terms of open data sets in French (110 billion words), German (30 billion words), Spanish, Dutch, and Italian.

Developing Common Corpus was an international effort involving a spectrum of stakeholders from the French Ministry of Culture to digital heritage researchers and open science LLM community, including companies such as HuggingFace, Occiglot, Eleuther, and Nomic AI. The collaborative effort behind building the data set reflects a vision of fostering a culture of openness and accessibility in AI research. Releasing Common Corpus is an attempt at democratizing access to large, quality data sets, which can be used for LLM training.”

Openly licensed, peer-reviewed data is generally considered more valuable by AI developers and researchers due to its accessibility, transparency, and potential for reduced bias. It has been fascinating to watch Wikimedia devise its strategy for AI. They state:

“As the internet continues to change and the use of AI increases, we expect that the knowledge ecosystem will become increasingly polluted with low-quality content, misinformation, and disinformation. We hope that people will continue to care about high quality, verifiable knowledge and they will continue to seek the truth, and we are betting that they will want to rely on real people to be the arbiters of knowledge. Made by human volunteers, we believe that Wikipedia can be that backbone of truth that people will want to turn to, either on the Wikimedia projects or through third party reuse.”

The popularity of Wikipedia as a data source resulted in such an increase in traffic from AI data scrapers that performance was being jeopardized for regular users. This led them to create a set of enterprise grade APIs in Wikimedia Enterprise that provide access to datasets of Wikipedia and sister projects while being supported by robust contracts, expert services, and unwavering support. Wikimedia Enterprise is a paid service targeted at commercial users of Wikimedia. Christie Rixford notes:

“At first glance, this initiative seems to be mainly about securing new revenue sources and thus improving sustainability of this civic platform. But in reality it is a milestone in developing Wikimedia as an access to knowledge infrastructure, and a strategy for adjusting to ongoing changes in the online ecosystem.”

They go on to say:

“The case of the Enterprise API is fascinating, as it shifts focus from the most visible part of the Wikimedia project: the production of encyclopaedic content by the community of volunteers. Instead, it focuses on code and infrastructure as tools for increasing access to knowledge. In doing so, it shows the limits of enabling reuse solely through legal means.” 

I think there are opportunities to create not just Large Language Models but smaller more customized openly licensed data sets for various academic domains using open science and open education sources. Higher education institutions or national systems could embark on the creation of specialized AI that starts with an existing open data set to which are added institution faculty endorsed open data sources and even student created open data, toward providing a unique and localized AI experience.

At the OEGlobal 2024 conference in Melbourne last year Martin Dougiamas proposed an OER Dataset for AI project.

Diagram showing steps associated with creating a curated OER Dataset for AI.

Another good example is the Open Datasets Initiative which aims to curate high-fidelity AI ready open datasets in biology and the life sciences. You can imagine something similar for climate action and other specific areas such as the United Nations SDGs or other similar targeted goals.

Conclusion: Tragedy Foretold? or Tragedy Avoided?

I have argued that to date AI is a tragedy caused by five fatal flaws:

  1. Tragedy of Capitalism over the Commons

  2. Taking without consent

  3. Demotivating sharing

  4. Breaking reciprocity, and

  5. Bleeding open dry

The fatal flaws conflict with the very things AI relies on risking an inevitable downfall of AI and associated downfall of open education, open science and other open practices.

I have also argued that the fatal flaws can be changed and suggested five ways the open community can take action to do so:

  1. Social Agreements & Regulations

  2. Preference Signals

  3. Reciprocity

  4. Public AI

  5. Collaborative open data set creation

There are other AI flaws and options for action but these, I think, are the key ones. The extent to which AI really does end up being a tragedy or, alternatively, an important and thriving addition to open knowledge creation depends on our response to it.

I hope this post helps make the fatal flaws of AI visible and inspires action to resolve them in ways that contribute to a flourishing commons.

In the words of Anna Tumadóttir:

“When we talk about defending the commons, it involves sustaining them, growing them, and making sure that the social contract remains intact for future generations of humans. And for that to happen, it’s time for some reciprocity.”

Next
Next

Reimagining Open At The Crossroads