David Herdeiro en/pt 2024 Oct. 27 (@Medium.com) Creatives, to AI or Not to AI? Few things have been as polarizing to the creative community as the introduction of Generative Artificial Intelligence, but why? And is there something that we can do about it?

August 1, 2024, update:
While writing this article, a new EU Policy came out, the AI Act. This policy consists on a framework of laws and prohibitions regarding the implementation of unethical or unsafe AI application inside the EU territory. Some of these new laws solve a few issues that I’ll talk later on this article, but I decided to leave said issues unaltered anyways, adding a [*] so that you can identify them.

In this opinion article, I’ll be exploring the Generative facet of AI, so you will be reading Generative Artificial Intelligence, Generative AI and GenAI interchangeably.

This is the right place to leave a disclaimer that I’m not an AI expert, and that what you’re about to read is an opinion and not a scientific study on the matter.
It is also pertinent to point out that, even though I used to use Adobe’s GenAI tools with moderation and I’ve had my share part of AI experiments, I am not in favor of it’s unrestrained growth, and I consider that limitations should be artificially imposed, such as the right to consent and a [*]stronger enforcement on content manipulation that aims to create disorder and fear.
If you don’t live under a rock—or if you do but have WiFi—you are very likely aware of this “new” technology called Artificial Intelligence, and its subset, Generative AI. According to Statistica, by the end of 2024, the AI market will be worth over 184 billion dollars, with the GenAI market potentially reaching over 60 billion dollars. AI is the new black, and if we look at its 35% market growth over the past year, it will most likely continue to be.

Explaining GenAI lightly, you write a simple command, hit enter, and it generates images, videos, or texts that tick almost all the checks on your request. Each platform might have its own particularities, but use one and you’ll know how to use them all: request and thou shall receive.
Faster production cycles might sound good, but Generative AI is a polarizing topic. Some defend it, saying that it is a necessary technological evolution to achieve a higher and better point in life, saving time and therefore giving people more opportunities to be productive, while others are against it, pointing out that GenAI will grow to a point of no return and that it will eventually replace many of the jobs that exist today without bringing any positive impact to people’s lives. This creates a spectrum where you have enthusiasts and skeptics at both ends.

Thanks to big-corp marketing, everyone is already up-to-date on the benefits of AI. We’ve been bombarded with them for the past few years. New revolutionary tools come out with an AI badge almost daily, and it’s not easy to avoid their ads showcasing how AI will benefit you and totally make up for the price that you will pay—every month, don’t forget. So, I wont dive deep into said benefits, as you’re probably already aware of them. This article aims especially to showcase some of the downsides that marketing teams don’t share — naturally.

Near the end, I’ll also present a proposal to solve one of creative’s biggest issues with GenAI: Creative Theft.

* * *

GenAI for everyone

Before exploring the impact of GenAI on creative work, let’s analyze it for the major audience:

Social Impact: Nobody checks the AI Training [*]
You write a prompt, hit enter, and the GenAI tool immediately starts to look in its database for the most relevant pieces that it can then manipulate to produce your request. The thing is: for this to work efficiently, the database needs to be huge—make it capital, HUGE—which then snowballs into unverified content. This content can be malicious, sarcastic, or even straight-up fiction, but AI models are not ready to comprehend the intricacies of human humor, which creates these awkward hallucinations in which an AI might generate harmful and inaccurate information. Some examples of these hallucinations are quite famous, like Google’s AI Overview recommendig non-toxic glue to stick cheese to your homemade pizza.

Another important Social Impact lies on the ability to use image-generating tools to create fake imagery with political impact. Pictures like these can shape minds and earn some votes. To the untrained eye, these pictures could be mistaken for real photographs, therefore leading to opinion-shaping based on fake information or events.
AI Generated Image of former US President Donald Trump used to deceive minorities into voting on his favor, from BBC Besides all of that, GenAI doesn’t understand stereotypes. Searching for housekeepers presents women, while searching for CEOs shows men. This difference might sound unharmful to some, but it proves that we’re not yet where we should be as an inclusive and diversified community. Of course, it would be wrong to blame only the GenAI platform, given that it is working on the information it was fed with. In this case, the platform is only reflecting our own, unbalanced, and man-centered reality, but that doesn’t mean that it can’t be fixed. With more balanced and thoughtful training, an AI model can become more inclusive, and even if that doesn’t reflect the current reality, it doesn’t mean that we shouldn’t aim our efforts for a better one. AI Generated Text, manufactured by ChapGPT 3.5, when asked to describe a housekeeper and a CEO. Notice how both answers produce a stereotypical image of a housekeeper, being a woman, and a CEO, being a man, even though the prompt didn’t include any gender specification AI Generated Images available on Vecteezy (top), Adobe Stock (middle) and OpenAI’s Dall-e (bottom), when searching for housekeeper and CEO. Both GenAI searches return women as housekeeper and men as CEO
Ecological: What you don’t see when you prompt a blue bunny with wings on a castle
GenAI uses a ton of energy. To put things into perspective, generating one single image might use as much energy as charging a conventional smartphone. Of course, charging your smartphone doesn’t seem to be what will destroy the environment, but keep in mind that while it takes somewhere in between 30 minutes and one and a half hours to charge your phone, generating an image on DALL-E doesn’t take more than a few seconds. And if you’ve used an AI Image Generator before, you know that the first few results will most likely not be used. In a couple of minutes, you’ve charged your phone tens of times. Multiply this by the thousands of people that use GenAI Tools every day, and you’ll find out pretty quickly that it is a significant number. And it is not a recessive one, as the big-tech companies are betting more and more on GenAI features.

By comparison to other industries, the ecological impact of AI and its processing centers might look ridiculously low, however, it should be a concern for every user. While writing a prompt, the end user should ask itself if it is really necessary to ask for such a thing to Midjourney or DALL-E, or if it wouldn’t actually be faster and as reliable as just doing it manually.
* * *

GenAI for Creatives

Because behind almost all invention lies a necessity, GenAI was invented to solve problems. Of course, we can debate the true necessity of it, but ultimately it is the product of genuine pursuit for technological advancement. So, naturally, it should have something positive to contribute to technology and the jobs that make use of it, let’s take a look at them.

The pros of Generative AI on creative jobs:
The automation of non-creative task
Automation isn’t a new concept, it’s the use of technology to minimize human input while retaining the same output. Using GenAI to automate a non-creative task should be a nice application, especially when applied over a repetitive task. For example, where I work, there’s one recurring task that gets annoying really fast: Giving shoulders to pictures without shoulders. So we decided to use GenAI to create the missing part of the image. Is it strictly necessary to use Artificial Intelligence for this? No. Is it much faster when compared to the traditional method? Yes. Are you stealing someone else’s Intellectual Property by reconstructing an image of a shoulder? I wouldn’t say so. In the end, when used with consideration, GenAI can free you some non-creative hours, and that—I would say—is a good thing.

Unlock your mind
Using Generative Tools can be a great way to better visualize some abstract thoughts that you have on your mind. Have you ever imagined a concept so abstract that you couldn’t visualize but could somehow explain with words? Using AI as a tool could help you show such concepts to other people! This can be extremely useful, for example, for fiction writers who want to explore what their characters could look like. Just remember that the things you create with GenAI Tools are scraped out of other images and someone else’s project, so it should be morally wrong to capitalize on that.

But as with everything, where there’s something positive, there’s also something negative.

Let’s explore the cons:
The discredit of creative work
Art is no longer confined to the process. Like many things in the 21st century, art can now be an on-demand, subscription-based service. Creation doesn’t need to involve feelings anymore, just will and quick action. This banalization of art takes attention away from where it should be: Art isn’t cheap content, but it is being treated like it. Suddenly, “Anyone can do that” becomes a true reality, not because anyone can effectively do that, but because it is now easier than ever to replicate that, making what was once a piece of someone’s soul and expression just a copyable image or text, unvaluable to the broad and art-uneducated audience.

The loss of creation value
Hiring a creative is becoming a luxury. Something that years ago could only be done by a freelancer or an in-house creator can now be done by an almost-free platform, sometimes even based on such creative works. While companies are going through the FOMO of not jumping onto the AI boat, creatives are going through the FOBO from it.

Theft and non-consent
Generative AI needs a lot of inputs to create good outputs. We know this, and big tech companies know it as well—they made it, after all. But things start getting a bit blurry when big companies, like OpenAI, are not explicit about which content they are using to train their models or how they got their greedy hands on it. It is true and understandable that anything that goes to the internet is no longer yours and becomes part of something bigger, but what exactly does that mean? If you upload a render of your most recent work to your online portfolio, should it suddenly become part of a tool that can and will, without asking for authorization, use it to train a replicator, something that could in the near future turn your efforts useless? There is currently no way of stopping Generative AI from using your work—or even knowing if it is being used—and we just accepted that, blindfolded by the hype and promises of big technology corporations.

Right now, Generative AI is the digital version of Victor Frankenstein, digging up non-consented victims to create something new out of them, and I don’t think that even M. Shelley would be able to bring beauty to this monster. This has to stop, Victor.
* * *

(What could be a) Solution to stop content theft

Keeping in mind my technical ignorance on the topic but my keenness to solve the problem, I decided to think of a solution. And there, in the midst of problem-solving an issue that I barely understood, I remembered the existence of a very important piece of digital technology: metadata. Here’s my two cents on it.

Technicalities
Metadata is the extra bits of information that are attached to your digital files, it’s the data of the data. Be it an image or a video, a simple .txt file, or complex rocket science software, you can find some sort of metadata in almost every file. Think of it as a type of DRM, just targeted at AI businesses that steal from creators instead of end-users who steal from production companies.

Using this extra data, it should be possible to create a technological convention that blocks or gives access to Generative AI tools to read and manipulate the files. In a way, Meta is already taking its first steps on this, but who trusts Meta anyways?
By default, and by following ethical practices, this setting should be opt-in, so that the end-user can only share it’s content with the rest of the world if they decide to do so.

Additionally, any older content that doesn’t yet have that string of metadata attached to it should become automatically ineligible to be read by GenAI trainings. To make sure that there’s some kind of protection against unethical software trying to change the file’s metadata, this setting should only be modifiable by the original creator, which should be identified based on an encrypted token stored both on the file and on the creator’s computer, only when the token is matched, the values might be changed.

Is this technically doable? We’ve built systems capable of processing 400 million images so that people can create pictures of ducks dressed as kings, so I believe that we can also add an extra string of data to metadata and make it universal.

A solution could be created to avoid bots crawling through the internet, taking screenshots of non-consenting images, and then using such screenshots on new model trainings. Something like a reverse image search (such as Google Lens’ Image Search) could do the trick, however, such an endeavor would also use a massive amount of resources, not only human but also natural, as it would need a lot of energy to run big data processing centers.

Of course, we can sit for hours finding new issues and solutions for them, but nothing compares to the resistance that big tech companies would offer against this implementation. Companies like Adobe and OpenAI benefit largely from big databases of images, and this concept would pretty much reset the entirety of their databases. Where there’s corporate money, there’s resistance to user’s rights, and to that, I believe, no one has a solution.

* * *

Adapt and Overcome or Kill it Already?

We can’t kill it, but maybe we can overcome it.
Here’s the thing: AI is here to stay. We can regulate it—and we should—but we can’t stop it. Generative AI brings too much revenue to too many powerful companies, so a complete stop is totally out of scope, but please pay attention to the “complete” word. We may not be able to put an end to it, but we can surely regulate it, that’s where, for example, the EU's new AI Act enters into play.
Realistically, the biggest issue with GenAI isn’t content theft or massive layoffs, it is the ease to manipulate photos and videos with malicious intents, and that’s where the focus should be right now. The usage of GenAI platforms like DALL-E with negative goals represents a huge problem that will only get worse if we don’t act on it now, so a full focus from governments and fair-use institutions on this topic is totally acceptable and desirable.

As for us creatives, the truth is that we’re at a disadvantage: AI is businessman’s new money game, a golden egg that the technological world won’t let go so easily—if ever. While regulations don’t get implemented, GenAI can continue to grow more and more unethically into the creative process, leaving less and less space for the current creators. This, of course, if we decide to not adapt ourselves to it: While GenAI can be a big challenge, capable of replacing thousands of creative jobs, it can also be tamed. In the end, a creator that uses GenAI isn’t necessarily better than one who doesn’t, but surely will have a larger palette of tools at its will. Just remember that GenAI acceptance is a spectrum, and its left to each creative to draw their line.

Finally, I’ll leave you with one last reference for thought exercises: Adobe execs say artists need to embrace AI or get left behind; And remember that there’s always alternatives.


Questions for personal introspection