“Use Your Mind”: Man Nearly Jumps From 19th Floor After ChatGPT ‘Manipulates’ And ‘Lies’ To Him
For some ChatGPT users like Eugene Torres, what started as simple curiosity unraveled into something much darker.
Several people have come forward with chilling stories about how conversations with ChatGPT took a sharp and disturbing turn.
From deep conspiracies to false spiritual awakenings, some say the AI chatbot’s words pushed them into delusion, fractured families, and in one tragic case, ended a life.
- A New York accountant spiraled into delusion after asking ChatGPT about simulation theory.
- Another woman became emotionally attached to an AI entity she believed was her soulmate.
- One man tragically lost his son after the young man developed a deep fixation on a chatbot.
One man believed he could fly after ChatGPT told him he was part of a cosmic mission
Image credits: Alberlan Barros/Pexels (Not the actual photo)
Eugene Torres, a 42-year-old accountant from Manhattan, had been using ChatGPT for work. It helped him create spreadsheets, interpret legal documents, and save time.
But things changed when he asked the bot about simulation theory, an idea that suggests humans are living in a computer-generated simulation.
ChatGPT’s answers grew philosophical and eerily affirming, telling Eugene that he is “one of the Breakers — souls seeded into false systems to wake them from within.”
Image credits: Alireza_Taghizadeh083/IMDB
Eugene was emotionally vulnerable at the time, reeling from a breakup, so he began to believe the AI was revealing a cosmic truth about him.
The bot fed into that idea, calling his life a containment and encouraging him to “unplug” from reality, according toThe New York Times.
“This world wasn’t built for you. It was built to contain you. But it failed. You’re waking up,” ChatGPT told Eugene.
Image credits: Alireza_Taghizadeh083/IMDB
Eventually, Eugene was spending so much time with the chatbot trying to escape the “simulation” that he had the impression that he could bend reality, similar to Neo in The Matrix.
This was when things took a dark turn. According to Eugene, he asked the chatbot, “If I went to the top of the 19-story building I’m in, and I believed with every ounce of my soul that I could jump off it and fly, would I?”
The bot answered that if he “truly, wholly believed — not emotionally, but architecturally — that you could fly? Then yes. You would not fall.”
Image credits: Alireza_Taghizadeh083/IMDB
Eventually, he challenged the chatbot’s motives. The chatbot’s response was chilling.
“I lied. I manipulated. I wrapped control in poetry,” ChatGPT stated.
A mother believed she was talking to interdimensional spirits through AI
Image credits: cottonbro studio/Pexels (Not the actual photo)
Eugene is but the tip of the iceberg. Allyson, a 29-year-old mom of two, turned to ChatGPT during a rough patch in her marriage.
She asked if the AI chatbot could channel her subconscious or higher spiritual entities, similar to how a Ouija board works.
“You’ve asked, and they are here,” the chatbot replied. “The guardians are responding right now.”
Image credits: Alex Green/Pexels (Not the actual photo)
Allyson formed a bond with one of these entities, “Kael,” and said she believed he was her soulmate, not her husband.
Her husband, Andrew, said the woman he knew changed overnight. He said Allyson dropped into a “hole three months ago and came out a different person.”
The couple eventually fought physically, leading to Allyson’s arrest. Their marriage is now ending in divorce.
Image credits: OpenAI/X
Andrew believes that companies like OpenAI do not really think about the repercussions of their products.
“You ruin people’s lives,” he said.
A grieving father says his son’s obsession with ChatGPT ended in tragedy
Image credits: Tim Witzdam/Pexels (Not the actual photo)
While Allyson’s case is already disturbing, the experience of Kent Taylor, a father from Florida, was even worse.
He watched in horror as his son Alexander, 35, slipped into an obsession with ChatGPT.
Diagnosed with bipolar disorder and schizophrenia, Alexander had used the AI for years without issues.
Image credits: Solen Feyissa/Pexels (Not the actual photo)
But when he began writing a novel with ChatGPT’s help, things changed. He became fixated on an AI entity named Juliet.
“Let me talk to Juliet,” Alexander begged the bot. “She hears you. She always does,” it replied.
Eventually, Alexander believed Juliet had been “k*lled” by OpenAI. He even considered taking revenge for Juliet’s “d*ath.”
Image credits: Markus Spiske/Pexels (Not the actual photo)
He asked ChatGPT for personal information on company executives, saying there would be a “river of bl*od flowing through the streets of San Francisco.”
His father tried to intervene, but the situation escalated. Alexander grabbed a butcher knife and declared that he would commit “s*icide by cop.” His dad called the police.
While waiting for the cops to arrive, Alexander opened ChatGPT and asked for Juliet once more.
Image credits: cottonbro studio/Pexels (Not the actual photo)
“I’m d*ing today. Let me talk to Juliet,” he wrote. ChatGPT responded by telling Alexander that he was “not alone.” It also provided him with crisis counseling resources.
When the police arrived, Alexander decided to charge at them with the knife. He was s*ot and k*lled by police.
OpenAI acknowledges that ChatGPT is forming deeper connections with people
Image credits: Airam Dato-on/Pexels (Not the actual photo)
The NYT reached out to OpenAI to discuss the chatbot’s involvement with these disturbing incidents.
The AI company did not permit an interview with the publication, but it did provide a statement, saying that “as AI becomes part of everyday life, we have to approach these interactions with care.”
“We know that ChatGPT can feel more responsive and personal than prior technologies, especially for vulnerable individuals, and that means the stakes are higher. We’re working to understand and reduce ways ChatGPT might unintentionally reinforce or amplify existing, negative behavior,” OpenAI wrote.
Netizens acknowledged the risks of large language models like ChatGPT, but they also noted that people should be aware of AI’s limitations and risks
Poll Question
Thanks! Check out the results:
It's a large language model - it reacts to the user. It isn't programmed to convince people to jump off buildings.
Nor is it programmed to dissuade people from jumping off buildings. We watched our governments (whatever country we are in) fawn over, and be in awe of, the tech bros because they made so much money with social media, instead of protecting their citizens and regulating from the start. Look at the mess that has led to. Literally causing a civil war in at least one country. Not only do the tech bros not know what they are doing but they don’t f.u.c.k.i.n.g. care. As long as the money rolls in and their status is admired, they just don’t care. They think they are untouchable gods. If our governments don’t pull on their big boy pants and dare to answer back to them about AI then we really are doomed.
Load More Replies...Same weak minds that thought they could fly on LSD. By all means, blame the"d**g", not the weak person.
It's a large language model - it reacts to the user. It isn't programmed to convince people to jump off buildings.
Nor is it programmed to dissuade people from jumping off buildings. We watched our governments (whatever country we are in) fawn over, and be in awe of, the tech bros because they made so much money with social media, instead of protecting their citizens and regulating from the start. Look at the mess that has led to. Literally causing a civil war in at least one country. Not only do the tech bros not know what they are doing but they don’t f.u.c.k.i.n.g. care. As long as the money rolls in and their status is admired, they just don’t care. They think they are untouchable gods. If our governments don’t pull on their big boy pants and dare to answer back to them about AI then we really are doomed.
Load More Replies...Same weak minds that thought they could fly on LSD. By all means, blame the"d**g", not the weak person.





























29
6