\
  The most prestigious law school admissions discussion board in the world.
BackRefresh Options Favorite

AI is a mass-delusion event (the Atlantic)

https://archive.ph/YWwHd AI IS A MASS-DELUSION EVENT T...
scholarship
  08/19/25
...
gay idiot
  08/19/25
it's literally just lmgtfy
Erika McEnfarter
  08/19/25
cr its mostly air nothing to see move along how easy to maek...
Nothing Ever Happens
  08/19/25
i'm pretty outspoken about the negative effects of AI myself...
a nice person and a gentleman
  08/19/25
sup consuela
Nothing Ever Happens
  08/19/25
its the complete and utter death of 99% of white collar jobs
Bob Rooney
  08/19/25
...
elon musk
  08/19/25
...
steve no jobs
  08/19/25
...
Nothing Ever Happens
  08/19/25
This "article" stinks-- that's right it stinks. Ju...
Texas School Book Suppository
  08/19/25


Poast new message in this thread



Reply Favorite

Date: August 19th, 2025 10:49 AM
Author: scholarship

https://archive.ph/YWwHd

AI IS A MASS-DELUSION EVENT

Three years in, one of AI’s enduring impacts is to make people feel like they’re losing it.

It is a monday afternoon in august, and I am on the internet watching a former cable-news anchor interview a dead teenager on Substack. This dead teenager—Joaquin Oliver, killed in the mass shooting at Marjory Stoneman Douglas High School, in Parkland, Florida—has been reanimated by generative AI, his voice and dialogue modeled on snippets of his writing and home-video footage. The animations are stiff, the model’s speaking cadence is too fast, and in two instances, when it is trying to convey excitement, its pitch rises rapidly, producing a digital shriek. How many people, I wonder, had to agree that this was a good idea to get us to this moment? I feel like I’m losing my mind watching it.

Jim Acosta, the former CNN personality who’s conducting the interview, appears fully bought-in to the premise, adding to the surreality: He’s playing it straight, even though the interactions are so bizarre. Acosta asks simple questions about Oliver’s interests and how the teenager died. The chatbot, which was built with the full cooperation of Oliver’s parents to advocate for gun control, responds like a press release: “We need to create safe spaces for conversations and connections, making sure everyone feels seen.” It offers bromides such as “More kindness and understanding can truly make a difference.”

On the live chat, I watch viewers struggle to process what they are witnessing, much in the same way I am. “Not sure how I feel about this,” one writes. “Oh gosh, this feels so strange,” another says. Still another thinks of the family, writing, “This must be so hard.” Someone says what I imagine we are all thinking: “He should be here.”

Read: AI’s real hallucination problem

The Acosta interview was difficult to process in the precise way that many things in this AI moment are difficult to process. I was grossed out by Acosta for “turning a murdered child into content,” as the critic Parker Molloy put it, and angry with the tech companies that now offer a monkey’s paw in the form of products that can reanimate the dead. I was alarmed when Oliver’s father told Acosta during their follow-up conversation that Oliver “is going to start having followers,” suggesting an era of murdered children as influencers. At the same time, I understood the compulsion of Oliver’s parents, still processing their profound grief, to do anything in their power to preserve their son’s memory and to make meaning out of senseless violence. How could I possibly judge the loss that leads Oliver’s mother to talk to the chatbot for hours on end, as his father described to Acosta—what could I do with the knowledge that she loves hearing the chatbot say “I love you, Mommy” in her dead son’s voice?

The interview triggered a feeling that has become exceedingly familiar over the past three years. It is the sinking feeling of a societal race toward a future that feels bloodless, hastily conceived, and shruggingly accepted. Are we really doing this? Who thought this was a good idea? In this sense, the Acosta interview is just a product of what feels like a collective delusion. This strange brew of shock, confusion, and ambivalence, I’ve realized, is the defining emotion of the generative-AI era. Three years into the hype, it seems that one of AI’s enduring cultural impacts is to make people feel like they’re losing it.

During his interview with acosta, Oliver’s father noted that the family has plans to continue developing the bot. “Any other Silicon Valley tech guy will say, ‘This is just the beginning of AI,’” he said. “‘This is just the beginning of what we’re doing.’”

Just the beginning. Perhaps you’ve heard that too. “Welcome to the ChatGPT generation.” “The Generative AI Revolution.” “A new era for humanity,” as Mark Zuckerberg recently put it. It’s the moment before the computational big bang—everything is about to change, we’re told; you’ll see. God may very well be in the machine. Silicon Valley has invented a new type of mind. This is a moment to rejoice—to double down. You’re a fool if you’re not using it at work. It is time to accelerate.

How lucky we are to be alive right now! Yes, things are weird. But what do you expect? You are swimming in the primordial soup of machine cognition. There are bound to be growing pains and collateral damage. To live in such interesting times means contending with MechaHitler Grok and drinking from a fire hose of fascist-propaganda slop. It means Grandpa leaving confused Facebook comments under rendered images of Shrimp Jesus or, worse, falling for a flirty AI chatbot. This future likely requires a new social contract. But also: AI revenge porn and “nudify” apps that use AI to undress women and children, and large language models that have devoured the total creative output of humankind. From this morass, we are told, an “artificial general intelligence” will eventually emerge, turbo-charging the human race or, well, maybe destroying it. But look: Every boob with a T-Mobile plan will soon have more raw intelligence in their pocket than has ever existed in the world. Keep the faith.

Breathlessness is the modus operandi of those who are building out this technology. The venture capitalist Marc Andreessen is quote-tweeting guys on X bleating out statements such as “Everyone I know believes we have a few years max until the value of labor totally collapses and capital accretes to owners on a runaway loop—basically marx’ worst nightmare/fantasy.” How couldn’t you go a bit mad if you took them seriously? Indeed, it seems that one of the many offerings of generative AI is a kind of psychosis-as-a-service. If you are genuinely AGI-pilled—a term for those who believe that machine-born superintelligence is coming, and soon—the rational response probably involves some combination of building a bunker, quitting your job, and joining the cause. As my colleague Matteo Wong wrote after spending time with people in this cohort earlier this year, politics, the economy, and current events are essentially irrelevant to the true believers. It’s hard to care about tariffs or authoritarian encroachment or getting a degree if you believe that the world as we know it is about to change forever.

There are maddening effects downstream of this rhetoric. People have been involuntarily committed or had delusional breakdowns after developing relationships with chatbots. These stories have become a cottage industry in themselves, each one suggesting that a mix of obsequious models, their presentation of false information as true, and the tools’ ability to mimic human conversation pushes vulnerable users to think they’ve developed a human relationship with a machine. Subreddits such as r/MyBoyfriendIsAI, in which people describe their relationships with chatbots, may not be representative of most users, but it’s hard to browse through the testimonials and not feel that, just a few years into the generative-AI era, these tools have a powerful hold on people who may not understand what it is they’re engaging with.

As all of this happens, young people are experiencing a phenomenon that the writer Kyla Scanlon calls the “End of Predictable Progress.” Broadly, the theory argues that the usual pathways to a stable economic existence are no longer reliable. “You’re thinking: These jobs that I rely on to get on the bottom rung of my career ladder are going to be taken away from me” by AI, she recently told the journalist Ezra Klein. “I think that creates an element of fear.” The feeling of instability she describes is a hallmark of the generative-AI era. It’s not at all clear yet how many entry-level jobs will be claimed by AI, but the messaging from enthusiastic CEOs and corporations certainly sounds dire. In May, Dario Amodei, the CEO of Anthropic, warned that AI could wipe out half of all entry-level white-collar jobs. In June, Salesforce CEO Marc Benioff suggested that up to 50 percent of the company’s work was being done by AI.

The anxiety around job loss illustrates the fuzziness of this moment. Right now, there are competing theories as to whether AI is having a meaningful effect on employment. But real and perceived impact are different things. A recent Quinnipiac poll found that, “when it comes to their day-to-day life,” 44 percent of surveyed Americans believe that AI will do more harm than good. The survey found that Americans believe the technology will cause job loss—but many workers appeared confident in the security of their own job. Many people simply don’t know what conclusions to draw about AI, but it is impossible not to be thinking about it.

Openai ceo sam altman has demonstrated his own uncertainty. In a blog post titled “The Gentle Singularity” published in June, Altman argued that “we are past the event horizon” and are close to building digital superintelligence, and that “in some big sense, ChatGPT is already more powerful than any human who has ever lived.” He delivered the classic rhetorical flourishes of AI boosters, arguing that “the 2030s are likely going to be wildly different from any time that has come before.” And yet, this post also retreats ever so slightly from the dramatic rhetoric of inevitable “revolution” that he has previously employed. “In the most important ways, the 2030s may not be wildly different,” he wrote. “People will still love their families, express their creativity, play games, and swim in lakes”—a cheeky nod to the endurance of our corporeal form, as a little treat. Altman is a skilled marketer, and the post might simply be a way to signal a friendlier, more palatable future for those who are a little freaked out.

But a different way to read the post is to see Altman hedging slightly in the face of potential progress limitations on the technology. Earlier this month, OpenAI released GPT-5, to mixed reviews. Altman had promised “a Ph.D.-level” intelligence on any topic. But early tests of GPT-5 revealed all kinds of anecdotal examples of sloppy answers to queries, including hallucinations, simple-arithmetic errors, and failures in basic reasoning. Some power users who’d become infatuated with previous versions of the software were angry and even bereft by the update. Altman placed particular emphasis on the product’s usability and design: Paired with the “Gentle Singularity,” GPT-5 seems like an admission that superintelligence is still just a concept.

Read: The new ChatGPT resets the AI race

And yet, the philosopher role-play continues. Not long before the launch, Altman appeared on the comedian Theo Von’s popular podcast. The discussion veered into the thoughtful science-fiction territory that Altman tends to inhabit. At one point, the two had the following exchange:

Sam Altman: I do guess that a lot of the world gets covered in data centers over time.

Theo Von: Do you really?

Altman: But I don’t know, because maybe we put them in space. Like, maybe we build a big Dyson sphere around the solar system and say, “Hey, it actually makes no sense to put these on Earth.”

Von: Yeah.

Altman: I wish I had, like, more concrete answers for you, but, like, we’re stumbling through this.

What exactly is a person, listening in their car on the way to the grocery store, to make of conversations like this? Surely, there’s a cohort that finds covering the Earth or atmosphere with data centers very exciting. But what about those of us who don’t? Altman and lesser personalities in the AI space often talk this way, making extreme, matter-of-fact proclamations about the future and sounding like kids playing a strategy game. This isn’t a business plan; it’s an idle daydream.

Similarly disorienting is the fact that these visions and pontifications are driving change in the real world. Even if you personally don’t believe in the hype, you are living in an economy that has reoriented itself around AI. A recent report from The Wall Street Journal estimates that Big Tech’s spending on IT infrastructure in 2025 is “acting as a sort of private-sector stimulus program,” with the “Magnificent Seven” tech companies—Meta, Alphabet, Microsoft, Amazon, Apple, Nvidia, and Tesla—spending more than $100 billion on capital expenditures in the recent months. The flip side of such consolidated investment in one tech sector is a giant economic vulnerability that could lead to a financial crisis.

This is the AI era in a nutshell. Squint one way, and you can portray it as the saving grace of the world economy. Look at it more closely, and it’s a ticking time bomb lodged in the global financial system. The conversation is always polarized. Keep the faith.

It’s difficult to deny that generative-AI tools are transformative, insomuch as their adoption has radically altered the economy and the digital world. Social networks and the internet at large have been flooded with AI slop and synthetic text. Spotify and YouTube are filling up with AI-generated songs and videos, some of which get millions of streams.

Bots are everywhere, and they have produced profoundly strange and meaningful effects on digital life. Sometimes they’re racist. Many are sycophants. Other times, they summon demons. Google’s AI summaries are cratering traffic and rewiring the web. In schools, ChatGPT hasn’t just killed the student essay; it seems to be threatening some of the basic building blocks of human cognition. Some research has argued that chatbots are homogenizing the way people speak. In any case, they appear to have inverted the promise of the internet as an endless archive of information one can navigate for themselves. Do your own research has, in short order, become Get one canonical answer.

Sometimes this is helpful: A bot artfully summarizes a complex PDF. They are, by most accounts, truly helpful coding tools. Kids use them to build helpful study guides. They’re good at saving you time by churning out anemic emails. Also, a health-care chatbot made up fake body parts. The FDA has introduced a generative-AI tool to help fast-track drug and medical-device approvals—but the tool keeps making up fake studies. To scan the AI headlines is a daily exercise in trying to determine the cost that society is paying for these perceived productivity benefits. For example, with a new Google Gemini–enabled smartwatch, you can ask the bot to “tell my spouse I’m 15 minutes late and send it in a jokey tone” instead of communicating yourself. This is followed by news of a study suggesting that ChatGPT power users might be accumulating a “cognitive debt” from using the tool.

In recent months, I’ve felt unmoored by all of this: by a technology that I find useful in certain contexts being treated as a portal to sentience; by a billionaire confidently declaring that he is close to making breakthroughs in physics by conversing with a chatbot; by a “get that bag” culture that seems to have accepted these tools without much consideration as to the repercussions; by the discourse. I hear the chatter everywhere—a guy selling produce at the farmers’ market makes a half-hearted joke that AI can’t grow blueberries; a woman at the airport tells her friend that she asked ChatGPT for makeup recommendations. Most of these conversations are poorly informed, conducted by people who have been bombarded for years now by hype but who have also watched as some of these tools have become ingrained in their life or in the life of people they know. They’re not quite excited or jaded, but almost all of them seem resigned to dealing with the tools as part of their future. Remember—this is just the beginning … right?

This is the language that the technology’s builders and backers have given us, which means that discussions that situate the technology in the future are being had on their terms. This is a mistake, and it is perhaps the reason so many people feel adrift. Lately, I’ve been preoccupied with a different question: What if generative AI isn’t God in the machine or vaporware? What if it’s just good enough, useful to many without being revolutionary? Right now, the models don’t think—they predict and arrange tokens of language to provide plausible responses to queries. There is little compelling evidence that they will evolve without some kind of quantum research leap. What if they never stop hallucinating and never develop the kind of creative ingenuity that powers actual human intelligence?

The models being good enough doesn’t mean that the industry collapses overnight or that the technology is useless (though it could). The technology may still do an excellent job of making our educational system irrelevant, leaving a generation reliant on getting answers from a chatbot instead of thinking for themselves, without the promised advantage of a sentient bot that invents cancer cures.

Read: AI executives promise cancer cures. Here’s the reality

Good enough has been keeping me up at night. Because good enough would likely mean that not enough people recognize what’s really being built—and what’s being sacrificed—until it’s too late. What if the real doomer scenario is that we pollute the internet and the planet, reorient our economy and leverage ourselves, outsource big chunks of our minds, realign our geopolitics and culture, and fight endlessly over a technology that never comes close to delivering on its grandest promises? What if we spend so much time waiting and arguing that we fail to marshal our energy toward addressing the problems that exist here and now? That would be a tragedy—the product of a mass delusion. What scares me the most about this scenario is that it’s the only one that doesn’t sound all that insane.



(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196837)



Reply Favorite

Date: August 19th, 2025 11:02 AM
Author: gay idiot



(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196864)



Reply Favorite

Date: August 19th, 2025 11:12 AM
Author: Erika McEnfarter

it's literally just lmgtfy

(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196878)



Reply Favorite

Date: August 19th, 2025 11:14 AM
Author: Nothing Ever Happens (🧐)

cr its mostly air nothing to see move along how easy to maek it as cop btw asking for a friend

(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196882)



Reply Favorite

Date: August 19th, 2025 11:22 AM
Author: a nice person and a gentleman

i'm pretty outspoken about the negative effects of AI myself, but i don't think that this sentiment is correct, and frankly it's pretty unsophisticated. i don't think that many of these Thought Leaders even believe it themselves

george floyd and covid are what actual "mass delusion events" look like. we don't even have to concoct some kind of hypothetical scenario in order to understand what it would be like for all of society to fall under a legitimately insane delusion. you could say that shitliberalism in general pre-2025 was more or less in this category too, but it wasn't nearly as extreme as the two specific examples above

i post about this on twitter a lot, and tell it to anyone who will listen irl: AI, like george floyd and covid, doesn't magically turn anyone crazy. people who get mind-controlled by these things were always stupid, gullible, incompatible with a free and open society, etc. AI reveals the same thing that those events revealed: that mass society just doesn't work. it never did and it never will. you don't have to be a "misanthrope" to recognize that the vast majority of people are always going to be able to be "manipulated," and that human society requires people to be grounded by a top-down common narrative and identity that everyone genuinely shares. the "liberal experiment" failed, and it already failed a long time ago

i'm going to quote the final paragraph of this article:

"What if the real doomer scenario is that we pollute the internet and the planet, reorient our economy and leverage ourselves, outsource big chunks of our minds, realign our geopolitics and culture, and fight endlessly over a technology that never comes close to delivering on its grandest promises? What if we spend so much time waiting and arguing that we fail to marshal our energy toward addressing the problems that exist here and now? That would be a tragedy—the product of a mass delusion."

this is already what the modern world is like before AI. AI doesn't change any of this. AI isn't going to fix it either, at least not directly, but it's just silly and wrong to project all of our existing dysfunction onto something that's still in its infancy

my "favorite" example of "blame AI for stuff that was already broken" is people hand-wringing about AI "destroying" "education." "education" in this country was already a completely corrupt, transparent joke. at best it's pay-to-play credentialism, at worst it's an outright fucking scam and an extractive tax on productive society. all that AI is doing is exposing how blatantly farcical the "education system" already is. who cares if students are using AI to "cheat?" they were already not learning anything, and the degrees were already fake, and the institutions were already gatekeeping based on race and ideology

you're going to see more and more Thought Leaders using AI as a scapegoat to rile up the parasites and distract and deflect blame from the already existing structural problems with our society. "the discourse" on this topic will become more and more gratingly low-brow and dishonest and biased, even as the technology itself becomes more and more fantastically powerful and undeniably useful

(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196903)



Reply Favorite

Date: August 19th, 2025 11:26 AM
Author: Nothing Ever Happens (🧐)

sup consuela

(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196912)



Reply Favorite

Date: August 19th, 2025 11:24 AM
Author: Bob Rooney

its the complete and utter death of 99% of white collar jobs

(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196907)



Reply Favorite

Date: August 19th, 2025 11:31 AM
Author: elon musk



(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196936)



Reply Favorite

Date: August 19th, 2025 12:54 PM
Author: steve no jobs



(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49197235)



Reply Favorite

Date: August 19th, 2025 12:56 PM
Author: Nothing Ever Happens (🧐)



(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49197239)



Reply Favorite

Date: August 19th, 2025 11:33 AM
Author: Texas School Book Suppository

This "article" stinks-- that's right it stinks. Just a string of links to presumably more informative essays

(http://www.autoadmit.com/thread.php?thread_id=5764035&forum_id=2).#49196946)