Is generative AI a threat to creative industries?

· Thomas Wood
Is generative AI a threat to creative industries?

Elevate Your Team with NLP Specialists

Unleash the potential of your NLP projects with the right talent. Post your job with us and attract candidates who are as passionate about natural language processing.

Hire NLP Experts

The AI ethics debate: does generative AI harm creative industries? What role will AI play in creative industries in the future?

Artists are worried that DALL-E could put them out of a job. Creative writers are raising the alarm about ChatGPT. Hollywood actors are worried about their voices being cloned. Before long, a similar debate could arise about AI-generated music.

What is not clear at the moment, is whether large generative AI models are harming those in creative professions, or if they are merely the latest iteration of tools that creative humans have at their disposal. There are a number of other widely discussed topics in AI ethics, from AI bias to autonomous weapons, but the advent of generative models in the last year has already begun to reshape the creative industry.

At the touch of a button, ChatGPT can generate a blog post, or DALL-E can create a graphic. Those who would otherwise have hired a freelancer can now use natural language processing and AI for free. A recent investigation by NewsGuard found 49 “news” sites that appear to be almost entirely written by AI.

The book illustrator Chris Mould pulled out of the Bradford Literature Festival when it emerged that AI was used to generate its publicity images.

…how can I stand under their roof and tell people they can go to art school and work in these disciplines, if the material used to publicise that event is generated at the push of a button?

Chris Mould

Mould is objecting to AI art on the grounds that it cheapens art, that is, it replaces human labour with an automated process.

The Australian artist Jazza has posted an interesting dissection and critique on his YouTube channel of an art book which he deduced was AI generated:

But there is another more subtle issue with AI generated art. Generative models do not exist in a vacuum of pure algorithms. They need human artists to work. GPT-3 was trained on a dataset of nearly a trillion words, all written by… humans! DALL-E was trained on a similar enormous dataset of images including original artwork. OpenAI acknowledge this on their website:

“reproducing training images verbatim can raise legal questions around copyright infringement, ownership, and privacy (if people’s photos were present in training data).” Source: OpenAI blog

What we can do for you

Concerned about Generative AI in creative industries?

We use AI across a multitude of industries on a daily basis. If you need data science or AI consulting, or want to discuss or learn more about generative AI in creative industries, talk to Fast Data Science

Generative AI for creative vision

Generative AI isn’t just about churning out realistic images. It can be used for powerful creative purposes:

  • The World Wildlife Fund used generative AI to imagine a world without nature, with animals extinct. Generative AI excels here because it can depict scenarios that don’t exist in reality.
  • A company used generative AI to create images of Pride celebrations in Thai cities. While Pride celebrations only exist in reality in Bangkok, generative AI could envision what they might look like elsewhere.

This showcases the potential of generative AI to not just replicate, but to inspire and create entirely new visual concepts.

The debate around generative AI and creative industries is only just beginning. While some fear it as a job killer, others see it as a powerful tool for creative expression. As with any new technology, the key will be using it responsibly and ethically to augment human creativity, not replace it.

OpenAI and Scarlett Johansson’s voice

In 2024, OpenAI said it would remove one of the voices used by ChatGPT, after the Hollywood actress Scarlett Johansson complained that it was too similar to her voice. Johansson claimed that OpenAI had contacted her, asking her permission to use her voice in the product, and she declined, and OpenAI included it anyway.

In this case, Johansson hired legal counsel, and OpenAI withdrew the voice from their product. It is unclear exactly whether Johansson’s voice was used to train OpenAI’s model, or whether the resemblance was coincidental, but it is precisely this plausible deniability which creatives are worried about.

In a time when we are all grappling with deepfakes and the protection of our own likeness, our own work, our own identities, I believe these are questions that deserve absolute clarity

Scarlett Johansson

The ethical question about generative AI and creative industries

The crux of the ethical question we face around generative models should not be the effect they have on supply and demand. In any case, when has it ever been possible to restrict a new technology because it puts people out of work? Even if one country legislates about this, companies will find a way.

We should be focusing on the right of the artist to control the use of their artwork for training AI models. Since the beginning of the digital age, artists have had to worry about others stealing their work. They can mitigate the risk of this happening by watermarking work, restricting access through technology, or through copyright laws.

So, if an artist can restrict others from copying their work without consent, what about using it to train an AI without their consent? Unfortunately this appears to be a legal grey area. The tech company Stability AI created an AI art tool called Stable Diffusion, which was trained on images taken from the internet, including from stock photo libraries such as Getty Images.

Getty Images has commenced legal proceedings against Stability AI in the High Court of Justice in London, alleging copyright violation and violation of their website’s terms of service.

Stock photo libraries embracing generative AI in creative industries

In 2023, I went to the Financial Times Future of AI Summit in London, and I was interested to watch a presentation by a speaker from one of the stock photo libraries, who said that AI image generation is simply “the latest disruptor in our industry”. Analogue photography to digitisation was a previous disruptor, and generative AI is the next wave. Stock content providers specialising in original content had to adapt to previous disruptive technologies and this opinion leader believed that they can survive this change.

The visual media companies are not sitting on the sidelines. Many formerly traditional content providers have made their own generative AI models, with some limitations that set them aside from the best-known big players in generative AI:

  • No celebrities
  • No brand names
  • No protected content

This ensures they avoid copyright issues and stay clear of ethically murky territory. The training data underlying these generative models wasn’t scraped from the internet, but was already obtained commercially by the content providers.

AI vs creative industries: How will the lawsuits pan out?

Even IP lawyers are unsure of how the AI/creative industry lawsuits will develop. On both sides of the Atlantic, the outcomes of the initial cases will affect the future landscape of AI art and generative AI’s impact on creative industries. In particular, copyright law defines fair use, derivative works and transformative works.

If you ask an AI to generate music in the style of the Rolling Stones, when their music is in its training dataset, would this be considered a derivative work? After all, depending on how it is calibrated, AIs can and often do spit out exact instances from their training dataset.

I asked DALL-E to generate me an image of Mickey Mouse in the style of Banksy, and it gave me the following, indicating that both Disney-copyrighted images, and Banksy’s artwork, must be in its training set.

Mickey Mouse in the style of Banksy

There are some lower-key alternatives to DALL-E which market themselves as taking a more ethical approach to artists’ rights. However, I do not envisage grassroots startups and non-profits making a significant change to the landscape without the budget of the big tech giants (the cost of training GPT-3 has been estimated to be between $4.6 million and $12 million).

Perhaps artists need to have the option to attach a license to their work, stating “I do not consent for this work to be used for training an AI model”. Or the default for any creative work could be an opt-in system, where only explicit consent counts.

I expect that in the next few years, after the first big civil cases, the dust will settle and the issue will become clearer to everyone what constitutes fair use of an artistic work and what role AI will play in creative industries.

The history of intellectual property

Interestingly, the concept of copyright did not begin in the digital age, but began much earlier, with the British Statute of Anne 1710, titled “An Act for the Encouragement of Learning, by vesting the Copies of Printed Books in the Authors or purchasers of such Copies, during the Times therein mentioned”. In essence, the arrival of the printing press in Europe allowed popular books to be easily copied.

Now, generative AI has created a new challenge for artists and writers, and legislation will need to catch up.

What’s next for generative AI and creative industries?

Percentage of papers on Pubmed containing delve

GPT has attracted attention because it is quite trigger happy with certain words such as ‘delve’. By 2024, nearly one in 200 papers on PubMed contained the word ‘delve’.

I have tried to use generative AI for content writing. Its output is mediocre at best, a distillation of the content of the internet at the date its training data was harvested. I view machine learning models as a generalisation of a kind of averaging (the simplest machine learning model possible is to take the mean of a set of numbers - this is a regression model with one single parameter).

In my view, a generative AI does not really create content, but synthesises and averages the existing content. I think for that reason a portion of the hysteria around generative models is somewhat unjustified.

Elevate Your Team with NLP Specialists

Unleash the potential of your NLP projects with the right talent. Post your job with us and attract candidates who are as passionate about natural language processing.

Hire NLP Experts

Look up company data from names (video)
Ai for business

Look up company data from names (video)

How to look up UK company data from company names (video) Imagine you have a clients list, suppliers list, or investment portfolio…

Unstructured data
Big dataNatural language processing

Unstructured data

Unstructured Data in Healthcare with NLP Introduction In today’s digital healthcare landscape, data plays a pivotal role. However, while medical records, patient feedback, and clinical research generate vast amounts of information, not all of it is easy to manage or analyze.

How to train your own AI: Fine tune an LLM for mental health data
Generative aiAi in research

How to train your own AI: Fine tune an LLM for mental health data

Fine tuning a large language model refers to taking a model that has already been developed, and training it on more data.

What we can do for you

Transform Unstructured Data into Actionable Insights

Contact us