“Do we need new laws for AI?” asks Jamie Dobson, founder of Container Solutions and author of ‘Visionaries, Rebels and Machines’
The current regulatory vacuum around Artificial Intelligence‘s use of data cannot persist indefinitely. Whether through government regulation, industry self-regulation, or landmark legal cases, new frameworks for managing AI’s relationship with human creativity must emerge.
The AI Revolution is different
The Industrial Revolution—and the pivotal role played by James Watt’s steam engine—is often cited when discussing fundamental changes to our economic and social fabric. Since then, we’ve had a number of other “revolutions”.
The harnessing of electricity ushered in mass production, electrified cities, and a communications revolution driven by the telegraph and later the telephone. Cars existed prior to Henry Ford’s moving assembly line, but this automotive revolution made them affordable, fundamentally changing global transportation, urban planning, and industry. The Space Age—beginning in 1957 with the launch of Sputnik—transformed communication and navigation, and the microprocessor, the Internet, DNA technology, and mobile phones all kicked off their own revolutions.
But the Industrial Revolution is still the biggie – the most bang-for-your-buck we’ve had in terms of shifting human civilisation.
Until now.
Artificial Intelligence is coming for James Watt’s crown. And AI’s revolution will be very different. Unlike previous technological revolutions that primarily transformed industries reliant on physical labour, AI’s impact extends to intellectual and creative domains previously considered uniquely human.
AI’s appetite for data
Modern AI systems learn by digesting vast quantities of human-created content. ChatGPT, Copilot, and DALL-E are sophisticated pattern-recognition systems trained on billions of examples of human creativity and knowledge.
Initially, tech companies trained these models on publicly available data, but as models grew more sophisticated, they required ever more data. Companies expanded their harvesting to include copyrighted content, paywalled articles, and private repositories. And that’s a problem for creators relying on compensation for their efforts, skill and talent.
Many AI companies delete their training data after models are built, making it nearly impossible to audit models for bias or track the use of copyrighted material. But even if this were not the case, currently, most jurisdictions have no specific regulations governing how companies can use publicly available data for AI training. This regulatory vacuum has allowed AI developers to operate under a take-first-ask-later approach, creating multi-billion-dollar technology platforms using content they didn’t create or license.
Undermining creative industries
The creative sector faces a unique double threat: not only might their jobs be automated, but their existing work could be used to train the very systems replacing them. When AI systems can freely reproduce the style and substance of human creators without compensation, we’re looking at a potentially destructive cycle:
- AI systems train on human-created content without compensation
- Economic incentives for human creation diminish
- New content production declines in quantity or quality
- AI systems have less novel material to learn from
- AI outputs become increasingly derivative and homogenised
So, it’s in everyone’s best interests for creators to continue to produce original content by being properly compensated for their contribution to AI learning. But will this happen without regulation?
As governments worldwide grapple with these challenges, several regulatory approaches are emerging:
Opt-in or Opt-out Models
The simplest solution could be to create a system for opting content in or out of AI training models. In theory, this could be quick to implement with minimum complexity. Yet, given that some models are already being trained on copyrighted content (which should already be a legal “opt-out”), it might not be particularly effective.
- Opt-Out Model (UK Proposal): Content creators must explicitly mark their work as not available for AI training. This places the burden on creators to protect their content.
- Opt-In Model (Advocated by Creator Organisations): AI companies must obtain permission before using copyrighted material, similar to how music licensing works.
For businesses, an opt-out system offers fewer obstacles to AI development but creates long-term legal uncertainty. An opt-in system provides clearer legal boundaries but potentially slower access to training data.
The UK’s proposed opt-out mechanism is particularly contentious. It’s essentially telling creators that someone can take their property unless they explicitly post a “No Trespassing” sign – in a language that hasn’t been invented yet. Critics argue this approach heavily favours large tech companies, as creators could easily lose rights to their work by simply forgetting to check a box or failing to implement technical measures they may not even understand.
Another issue is policing this and ensuring opted out data is not inadvertently or deliberately used.
Data Rights and Compensation Models
Similar to how music and literary rights work, content creators could receive compensation when their work is used for AI training. This could be done on an ad-hoc basis, like music streaming, or through government distribution via a digital tax.
- Collective licensing: Creators register with collecting societies that negotiate with AI companies and distribute payments based on usage. This model exists in music with performing rights organisations such as PRS in the UK, ASCAP and BMI in the USA, GEMA in Germany or SACEM in France.
- Data dividend: A tax or fee on AI companies based on their data usage, with proceeds distributed to creators. This resembles public lending rights systems in countries like the UK, Canada, and Australia, where authors receive payments when libraries lend their books.
- Direct licensing: Individual negotiations between major content producers and AI companies, with standardised terms for smaller creators.
AI as a Public Resource
Some experts advocate treating advanced AI systems like public utilities or natural monopolies. This would work similarly to electricity companies, for example, where the national grid is seen as a natural monopoly and the government implements certain standards and expectations for managing it as a public resource.
- Private companies would continue developing AI, but under enhanced regulatory oversight
- Transparency requirements would include regular audits and public reporting
- Universal access provisions would ensure broad distribution of benefits
- Price controls or licensing requirements would prevent monopolistic practices
This approach draws from how telecommunications, electricity, and other essential services are regulated in many countries. It acknowledges both the innovation potential of private enterprise and the public interest in fair, accessible AI systems.
Transparency and Technical Safeguards
Any potential regulatory framework will require some level of transparency and technical safeguards to ensure that AI is not operating as a black box. We need to know how the algorithms are fed and on what data to ensure that creators are fairly compensated and we aren’t introducing systemic biases into what will become ubiquitous technology. In other words, whatever system is chosen, it needs to be tracked and policed to ensure compliance. And this ‘policing’ needs to be paid for.
This will entail:
- Provenance tracking: Requiring AI companies to maintain comprehensive records of their training data sources, making it possible to audit systems and verify compliance.
- Attribution systems: Building technical capabilities into AI systems that credit original creators when their work influences outputs.
- Content authentication: Developing standards for watermarking or otherwise identifying AI-generated content to distinguish it from human-created work.
These technical measures would address what many legal experts consider a fundamental flaw in current AI development: the lack of transparency about what data is being used and how it influences outputs.
The publishing industry could offer a useful approach. Copyright registration systems and ISBN standards create a framework for tracking and attributing written works. Similar systems could be developed for AI training data, creating both accountability and the technical infrastructure for fair compensation.
The International Regulatory Landscape
Different regions are taking distinctly different approaches to AI regulation:
The European Union: Comprehensive Protection
The EU’s AI Act takes a risk-based approach, imposing stringent requirements for high-risk applications while allowing more flexibility for lower-risk uses. Regarding training data, the EU leans toward an opt-in model that requires permission before using copyrighted material.
For businesses operating in Europe, this means higher compliance costs but also a clearer regulatory landscape and potentially greater consumer trust.
The United Kingdom: Innovation First
The UK is positioning itself as a pro-innovation hub, with proposed regulations that favour AI developers. Their controversial opt-out model places the burden on creators to explicitly mark their work as not available for AI training.
Critics argue this approach unfairly disadvantages individual creators who may lack the technical knowledge or resources to implement protection measures. It essentially tells creators that someone can use their property unless they implement specific (as yet not developed or available) technical measures – measures many creators may not understand.
The United States: Sector-Specific Approach
The US has thus far avoided comprehensive AI regulation, instead focusing on sector-specific guidelines and voluntary frameworks. This creates a more fragmented landscape where rules for using data may vary significantly across industries.
This approach offers flexibility but creates regulatory uncertainty for businesses, who may need to navigate different requirements across sectors and jurisdiction, which in turn increases costs and risks.
Conclusion
The AI businesses that thrive won’t be those extracting maximum short-term value from unregulated data harvesting, but those building sustainable models that respect and reinforce the creative ecosystem upon which AI ultimately depends.
After all, AI systems—no matter how advanced—remain fundamentally derivative of human creativity and knowledge. Ensuring the continued flourishing of that human creativity isn’t just an ethical imperative, it’s essential for AI’s own future development.
For businesses navigating this rapidly evolving landscape, the key is neither blind resistance nor uncritical embrace of AI, but thoughtful engagement with both the possibilities these technologies offer and the responsibilities they entail.
ABOUT THE AUTHOR
Jamie Dobson is the founder of Container Solutions, and has been helping companies, across industries, move to cloud native ways of working for over ten years. Container Solutions develops a strategy, a clear plan and step by step implementation helping companies achieve a smooth digital transformation. With services including Internal Developer Platform Enablement, Cloud Modernisation, DevOps/DevSecOps, Site Reliability Engineering (SRE) Consultancy, Cloud Optimisation and creating a full Cloud Native Strategy, companies get much more than just engineering know-how. Jamie is also author of ‘The Cloud Native Attitude’ and the recently published ‘Visionaries, Rebels and Machines: The story of humanity’s extraordinary journey from electrification to cloudification’. Both are available from Amazon and good bookstores.
https://www.container-solutions.com/