Should AI companies be allowed to mine creative content to train models?

· Thomas Wood
Should AI companies be allowed to mine creative content to train models?

When AI companies train models on text, video and images from creative industries, the end product is an AI model which can create near-human quality visual designs or copywriting. Many artists have argued the AI companies are exploiting creatives by profiting from their work, while the original creators do not receive any compensation. By flooding the market with low-cost content, AI has driven down rates for human artists, which is threatening creative careers.

Artists are also distressed by style mimicry, where AI models can create instant artworks “in the style of” Banksy, Van Gogh, Frida Kahlo, or Tracy Emin.

In the recent Getty Images v Stability AI case, a judge ruled that Stability AI’s image generation model, which was trained on images belonging to Getty Images, didn’t infringe Getty Images’ copyright because an AI model doesn’t store the original photos and is not a copy in the sense of copyright law. It was also difficult for Getty Images to enforce their copyright because Stability’s training took place outside the UK - meaning that jurisdiction appears to only apply to the location that the model training took place in. The Getty ruling revealed gaps in copyright law that the UK government is now trying to fix.

The UK government has proposed changing copyright rules to make it easier for AI companies to mine content, by assuming that creative works are fair game for training AI models, unless the creators choose to “opt out” of such arrangements. In other words, all your artwork can be used to train an AI unless you take proactive steps to prevent it. TV companies and film makers have claimed that this opt-out system is unworkable. Eric Fellner, co-chair of Working Title Films, called the proposal an “existential threat” to creative industries.

In February 2025, over 1000 musicians including Kate Bush, Annie Lennox, and Damon Albarn released a silent protest album titled Is This What We Want?, which had 12 silent tracks had titles that collectively spelled out: “The British Government Must Not Legalise Music Theft To Benefit AI Companies.”

In the last week, British artists were granted a small victory, as the House of Lords released a landmark report titled AI, Copyright and the Creative Industries[1], which criticised all of the government’s opt-in models. However, the report has been criticised as “kicking [copyright] down the road” in the Financial Times.[2]

The House of Lords recommended that the UK government abandons the “opt-out” models for creatives, introduce protections against “in the style of” AI outputs, and force companies to be transparent about where their training data came from. In other words, the UK should move towards a more regulated approach to AI training.

The Lords also recommended the UK should build its own sovereign AI using licensed data.

What does this mean for artists and AI companies?

We heard that one of the most acute harms arising from generative AI outputs is outputs ‘in the style of’ an artist. These can imitate their recognisable style, voice or personality in a way that displaces commissions or devalues the distinctive appeal of their work without reproducing a substantial part of any particular work.

  • House of Lords, AI, copyright and the creative industries[1]

Unfortunately, I don’t think the House of Lords’ recommendations will have any effect at all.

AI training and model use can take place anywhere in the world, so whatever the UK does to enforce certain restrictions will be easily bypassed.

I don’t see how AI can be prevented from creating works “in the style of” an artist. It is currently legal for a human to manually create an artwork in the style of an artist, and I struggle to see an effective way that we can forbid AI models from doing this without granting artists a monopoly over a particular creative style, which ultimately harms other artists. If “painting like David Hockney” becomes a legal infringement, it could lead to a wave of frivolous lawsuits against human artists.

UK based AI companies have argued that under the government’s proposals or the proposals in the Lords’ report, it will be hard to compete with US and Chinese companies, who operate under more permissive rules, and the UK would miss out on the AI boost to the economy. The idea of a sovereign AI has been criticised as unrealistic. It would be hard for the UK to train a publicly-owned “clean” model which would be able to compete with GPT-5 or Gemini.

I do not foresee any large changes coming quickly off the back of the Lords’ report, and even with the best of intentions, any legislative measures can be bypassed by tech companies using more permissive jurisdictions for training their models.

References

  1. House of Lords Communications and Digital Committee, AI, copyright and the creative industries, 2026.
  2. Anna Gross and Daniel Thomas, UK to delay difficult decisions on AI copyright rules, Financial Times, 7 March 2026.

Unlock Your Future in NLP!

Dive into the world of Natural Language Processing! Explore cutting-edge NLP roles that match your skills and passions.

Explore NLP Jobs

Are we in an AI bubble?
Generative ai

Are we in an AI bubble?

None of the content of this article constitutes investment advice. Few would dispute that the global economy is currently in an AI boom. However, opinions are divided over whether this boom also constitutes a bubble.

Launching Harmony Meta
Ai in research

Launching Harmony Meta

We are excited to introduce the new Harmony Meta platform, which we have developed over the past year. Harmony Meta connects many of the existing study catalogues and registers.

The Ethics of AI in Healthcare: Opportunities and Risks
Generative ai

The Ethics of AI in Healthcare: Opportunities and Risks

Guest post by Jay Dugad Artificial intelligence has become one of the most talked-about forces shaping modern healthcare. Machines detecting disease, systems predicting patient deterioration, and algorithms recommending personalised treatments all once sounded like science fiction but now sit inside hospitals, research labs, and GP practices across the world.

What we can do for you

Transform Unstructured Data into Actionable Insights

Contact us