Toggle Menu

How Much of Our Humanity Are We Willing to Outsource to AI?

Tech companies claim artificial general intelligence systems will propel our society forward. But the cost to our humanity may not be worth the risk.

Sage Cammers-Goodwin and Rosalie Waelen

March 27, 2024

OpenAI claims “to ensure AGI benefits all of humanity,” but in reality it threatens the very concept of humanity.(Photo: CFOTO / Future Publishing via Getty Images)

OpenAI’s latest system, Sora, creates videos on the basis of text prompts. You write a prompt and Sora does the rest. The result: highly photorealistic videos and endearing animations. From mammoths sauntering through a snowy meadow to a market full of people in Lagos, Nigeria, in the year 2056. Nothing seems too wild for Sora’s imagination.

Sora is not yet available to the wider public, because OpenAI is first letting “red teamers” assess the product in regard to its potential for abuse. OpenAI also asked creative professionals for feedback on how Sora can benefit them, framing Sora as a tool that could enhance their output.

At the same time, however, technologies like Sora may render many creatives irrelevant. Just last month, famed director and producer Tyler Perry halted an $800 million studio expansion after seeing what Sora is capable of. It would eliminate the need for him to build sets or travel to film on-location. “I can sit in an office and do this with a computer, which is shocking to me,” he told The Hollywood Reporter.

It should come as no surprise that advances in AI have sparked debates about the future of work as well as concerns about misinformation, bias, and copyright infringements. Responding to the backlash, tech companies have (virtue-)signaled consideration for these issues by stressing the importance of Ethical AI, Responsible AI, or Trustworthy AI and developing guidelines for these noble goals. Indeed, OpenAI’s website includes the disclaimer that despite extensive testing, it cannot predict all possible benefits and harms of its technology, and that (ironically) it is critical to release (potentially harmful) AI into the real world to increase AI’s safety over time.

Current Issue

View our current issue

Subscribe today and Save up to $129.

Policymakers are also actively discussing how to regulate AI in order to prevent or, at least, limit the aforementioned problems. In fact, earlier this month the EU passed the landmark AI Act, which includes bans and restrictions related to biometric identification systems, manipulative uses of AI, and artificial general intelligence (AGI, or AI with human-level or above intelligence and the ability to self-teach).

While the widespread attention for AI ethics is to be applauded, it is guiding attention away from a deeper issue. Focusing on ways to regulate and improve AI confirms a techno-deterministic narrative, one that does not question the overall desirability of technological advancements and assumes that AI and AGI are inevitable. However, we as a society must question whether generative AI systems like Sora really bring about progress and whether these systems should be welcomed in the first place.

OpenAI claims “to ensure AGI benefits all of humanity,” but in reality it threatens the very concept of humanity. We are beyond AI just beating us at chess and Go, optimizing pricing, and recognizing faces. Although these previous milestones were impressive, shocking, and maybe even demeaning to those with a strong sense of human superiority, that technology “merely” automated routine and rule-based tasks. Now, as our preemptive concerns about super intelligence have slowly dampened, AI has begun to master a realm many long thought was uniquely human—the realm of creativity and imagination.

Creativity and imagination are related but different phenomena. To create or be creative means to materialize a vision or idea into a painting, a movie, a song, and so on. To imagine goes a step further. Imagination requires the ability to develop that vision in the first place—to see what has not yet been manifested. Sora fills the gaps in our prompts. The model not only understands prompts (a point OpenAI repeatedly emphasizes on Sora’s site), but also imagines what the scenes described could and should look like. Yes, that imagination is based on training data, but so too is ours.

For humans, imagination takes time and space to develop. But in our capitalist society, which determines our value based on productivity and volume, artists are told that AI will be helpful to them because it will allow them to create more content so that they can work faster. These promises of productivity make it easier for people to overlook the fact that AGI-driven intelligent systems are on track to have capacities equal or better than humans’. AI’s imagination will likely become more grandiose in years to come.

We should not let the promise of productivity or narrow debates about AI’s ethical implications distract us from the bigger picture. Under the guise of improving humanity by increasing productivity, we risk releasing our ultimate replacement.

The Nation Weekly
Fridays. A weekly digest of the best of our coverage.
By signing up, you confirm that you are over the age of 16 and agree to receive occasional promotional offers for programs that support The Nation’s journalism. You may unsubscribe or adjust your preferences at any time. You can read our Privacy Policy here.

We should not overestimate the durability of human skills in the face of technological advancements. Calculators decreased the need for mental math. GPS has limited independent exploration and map navigation. While some might be glad to let these skills go, there must be a limit to the qualities we are willing to outsource.

Thank you for reading The Nation!

We hope you enjoyed the story you just read, just one of the many incisive, deeply-reported articles we publish daily. Now more than ever, we need fearless journalism that shifts the needle on important issues, uncovers malfeasance and corruption, and uplifts voices and perspectives that often go unheard in mainstream media.

Throughout this critical election year and a time of media austerity and renewed campus activism and rising labor organizing, independent journalism that gets to the heart of the matter is more critical than ever before. Donate right now and help us hold the powerful accountable, shine a light on issues that would otherwise be swept under the rug, and build a more just and equitable future.

For nearly 160 years, The Nation has stood for truth, justice, and moral clarity. As a reader-supported publication, we are not beholden to the whims of advertisers or a corporate owner. But it does take financial resources to report on stories that may take weeks or months to properly investigate, thoroughly edit and fact-check articles, and get our stories into the hands of readers.

Donate today and stand with us for a better future. Thank you for being a supporter of independent journalism.

Thank you for your generosity.

Given AI’s current skills and capacity for imagination, it seems plausible that AI will not propel humanity forward, as OpenAI claims, but threaten humanity instead, by de-skilling us and rendering our unique features superfluous. That is why it’s past time that we discuss not only ethical implications like copyright breaches and algorithmic biases, but also what AI’s ability to imagine means for humanity. What does it mean to be human when our distinctive and characteristic skills and features are no longer uniquely ours?

It’s also crucial that we not allow ourselves to be misled by narratives about how ChatGPT and Sora will save humanity by making us more productive in work and everyday life, when they will likely primarily boost earnings for the one percent. Ultimately, as AI continues to develop, we may be left skill-less, uninspired, and dependent on our replacement. At that point, it will be too late to ask ourselves if building AI was worth the loss of our humanity.

Sage Cammers-GoodwinDr. Sage Cammers-Goodwin is a philosophy of technology researcher at the University of Twente. Her interests and expertise span corporate social responsibility, smart cities, and emerging technologies, including artificial intelligence. Her publications appear across multiple platforms including Oxford University Press.


Rosalie WaelenDr. Rosalie Waelen is currently working at the Sustainable AI Lab, which is a part of the Institute for Science and Ethics at the University of Bonn, Germany. Rosalie has previously published on the importance of critical theory perspectives in the AI debate and on the ethical and societal issues related to computer vision applications.


Latest from the nation