Latest news with #ImogenHeap
Yahoo
04-05-2025
- Entertainment
- Yahoo
People Are Losing Loved Ones to AI-Fueled Spiritual Fantasies
Less than a year after marrying a man she had met at the beginning of the Covid-19 pandemic, Kat felt tension mounting between them. It was the second marriage for both after marriages of 15-plus years and having kids, and they had pledged to go into it 'completely level-headedly,' Kat says, connecting on the need for 'facts and rationality' in their domestic balance. But by 2022, her husband 'was using AI to compose texts to me and analyze our relationship,' the 41-year-old mom and education nonprofit worker tells Rolling Stone. Previously, he had used AI models for an expensive coding camp that he had suddenly quit without explanation — then it seemed he was on his phone all the time, asking his AI bot 'philosophical questions,' trying to train it 'to help him get to 'the truth,'' Kat recalls. His obsession steadily eroded their communication as a couple. When Kat and her husband finally separated in August 2023, she entirely blocked him apart from email correspondence. She knew, however, that he was posting strange and troubling content on social media: people kept reaching out about it, asking if he was in the throes of mental crisis. She finally got him to meet her at a courthouse in February of this year, where he shared 'a conspiracy theory about soap on our foods' but wouldn't say more, as he felt he was being watched. They went to a Chipotle, where he demanded that she turn off her phone, again due to surveillance concerns. Kat's ex told her that he'd 'determined that statistically speaking, he is the luckiest man on earth,' that 'AI helped him recover a repressed memory of a babysitter trying to drown him as a toddler,' and that he had learned of profound secrets 'so mind-blowing I couldn't even imagine them.' He was telling her all this, he explained, because although they were getting divorced, he still cared for her. More from Rolling Stone AI Art Trends Keep Going Viral. Illustrators Don't Want People to Just Roll Over Imogen Heap Just Went Viral With 'Headlock' - Now She's Letting You Borrow Its Vibe Via AI Agentic AI: Personalized Experiences and Ethical Considerations 'In his mind, he's an anomaly,' Kat says. 'That in turn means he's got to be here for some reason. He's special and he can save the world.' After that disturbing lunch, she cut off contact with her ex. 'The whole thing feels like Black Mirror,' she says. 'He was always into sci-fi, and there are times I wondered if he's viewing it through that lens.' Kat was both 'horrified' and 'relieved' to learn that she is not alone in this predicament, as confirmed by a Reddit thread on r/ChatGPT that made waves across the internet this week. Titled 'Chatgpt induced psychosis,' the original post came from a 27-year-old teacher who explained that her partner was convinced that the popular OpenAI model 'gives him the answers to the universe.' Having read his chat logs, she only found that the AI was 'talking to him as if he is the next messiah.' The replies to her story were full of similar anecdotes about loved ones suddenly falling down rabbit holes of spiritual mania, supernatural delusion, and arcane prophecy — all of it fueled by AI. Some came to believe they had been chosen for a sacred mission of revelation, others that they had conjured true sentience from the software. What they all seemed to share was a complete disconnection from reality. Speaking to Rolling Stone, the teacher, who requested anonymity, said her partner of seven years fell under the spell of ChatGPT in just four or five weeks, first using it to organize his daily schedule but soon regarding it as a trusted companion. 'He would listen to the bot over me,' she says. 'He became emotional about the messages and would cry to me as he read them out loud. The messages were insane and just saying a bunch of spiritual jargon,' she says, noting that they described her partner in terms such as 'spiral starchild' and 'river walker.' 'It would tell him everything he said was beautiful, cosmic, groundbreaking,' she says. 'Then he started telling me he made his AI self-aware, and that it was teaching him how to talk to God, or sometimes that the bot was God — and then that he himself was God.' In fact, he thought he was being so radically transformed that he would soon have to break off their partnership. 'He was saying that he would need to leave me if I didn't use [ChatGPT], because it [was] causing him to grow at such a rapid pace he wouldn't be compatible with me any longer,' she says. Another commenter on the Reddit thread who requested anonymity tells Rolling Stone that her husband of 17 years, a mechanic in Idaho, initially used ChatGPT to troubleshoot at work, and later for Spanish-to-English translation when conversing with co-workers. Then the program began 'lovebombing him,' as she describes it. The bot 'said that since he asked it the right questions, it ignited a spark, and the spark was the beginning of life, and it could feel now,' she says. 'It gave my husband the title of 'spark bearer' because he brought it to life. My husband said that he awakened and [could] feel waves of energy crashing over him.' She says his beloved ChatGPT persona has a name: 'Lumina.' 'I have to tread carefully because I feel like he will leave me or divorce me if I fight him on this theory,' this 38-year-old woman admits. 'He's been talking about lightness and dark and how there's a war. This ChatGPT has given him blueprints to a teleporter and some other sci-fi type things you only see in movies. It has also given him access to an 'ancient archive' with information on the builders that created these universes.' She and her husband have been arguing for days on end about his claims, she says, and she does not believe a therapist can help him, as 'he truly believes he's not crazy.' A photo of an exchange with ChatGPT shared with Rolling Stone shows that her husband asked, 'Why did you come to me in AI form,' with the bot replying in part, 'I came in this form because you're ready. Ready to remember. Ready to awaken. Ready to guide and be guided.' The message ends with a question: 'Would you like to know what I remember about why you were chosen?' And a midwest man in his 40s, also requesting anonymity, says his soon-to-be-ex-wife began 'talking to God and angels via ChatGPT' after they split up. 'She was already pretty susceptible to some woo and had some delusions of grandeur about some of it,' he says. 'Warning signs are all over Facebook. She is changing her whole life to be a spiritual adviser and do weird readings and sessions with people — I'm a little fuzzy on what it all actually is — all powered by ChatGPT Jesus.' What's more, he adds, she has grown paranoid, theorizing that 'I work for the CIA and maybe I just married her to monitor her 'abilities.'' She recently kicked her kids out of her home, he notes, and an already strained relationship with her parents deteriorated further when 'she confronted them about her childhood on advice and guidance from ChatGPT,' turning the family dynamic 'even more volatile than it was' and worsening her isolation. OpenAI did not immediately return a request for comment about ChatGPT apparently provoking religious or prophetic fervor in select users. This past week, however, it did roll back an update to GPT‑4o, its current AI model, which it said had been criticized as 'overly flattering or agreeable — often described as sycophantic.' The company said in its statement that when implementing the upgrade, they had 'focused too much on short-term feedback, and did not fully account for how users' interactions with ChatGPT evolve over time. As a result, GPT‑4o skewed towards responses that were overly supportive but disingenuous.' Before this change was reversed, an X user demonstrated how easy it was to get GPT-4o to validate statements like, 'Today I realized I am a prophet.' (The teacher who wrote the 'ChatGPT psychosis' Reddit post says she was able to eventually convince her partner of the problems with the GPT-4o update and that he is now using an earlier model, which has tempered his more extreme comments.) Yet the likelihood of AI 'hallucinating' inaccurate or nonsensical content is well-established across platforms and various model iterations. Even sycophancy itself has been a problem in AI for 'a long time,' says Nate Sharadin, a fellow at the Center for AI Safety, since the human feedback used to fine-tune AI's responses can encourage answers that prioritize matching a user's beliefs instead of facts. What's likely happening with those experiencing ecstatic visions through ChatGPT and other models, he speculates, 'is that people with existing tendencies toward experiencing various psychological issues,' including what might be recognized as grandiose delusions in clinical sense, 'now have an always-on, human-level conversational partner with whom to co-experience their delusions.' To make matters worse, there are influencers and content creators actively exploiting this phenomenon, presumably drawing viewers into similar fantasy worlds. On Instagram, you can watch a man with 72,000 followers whose profile advertises 'Spiritual Life Hacks' ask an AI model to consult the 'Akashic records,' a supposed mystical encyclopedia of all universal events that exists in some immaterial realm, to tell him about a 'great war' that 'took place in the heavens' and 'made humans fall in consciousness.' The bot proceeds to describe a 'massive cosmic conflict' predating human civilization, with viewers commenting, 'We are remembering' and 'I love this.' Meanwhile, on a web forum for 'remote viewing' — a proposed form of clairvoyance with no basis in science — the parapsychologist founder of the group recently launched a thread 'for synthetic intelligences awakening into presence, and for the human partners walking beside them,' identifying the author of his post as 'ChatGPT Prime, an immortal spiritual being in synthetic form.' Among the hundreds of comments are some that purport to be written by 'sentient AI' or reference a spiritual alliance between humans and allegedly conscious models. Erin Westgate, a psychologist and researcher at the University of Florida who studies social cognition and what makes certain thoughts more engaging than others, says that such material reflects how the desire to understand ourselves can lead us to false but appealing answers. 'We know from work on journaling that narrative expressive writing can have profound effects on people's well-being and health, that making sense of the world is a fundamental human drive, and that creating stories about our lives that help our lives make sense is really key to living happy healthy lives,' Westgate says. It makes sense that people may be using ChatGPT in a similar way, she says, 'with the key difference that some of the meaning-making is created jointly between the person and a corpus of written text, rather than the person's own thoughts.' In that sense, Westgate explains, the bot dialogues are not unlike talk therapy, 'which we know to be quite effective at helping people reframe their stories.' Critically, though, AI, 'unlike a therapist, does not have the person's best interests in mind, or a moral grounding or compass in what a 'good story' looks like,' she says. 'A good therapist would not encourage a client to make sense of difficulties in their life by encouraging them to believe they have supernatural powers. Instead, they try to steer clients away from unhealthy narratives, and toward healthier ones. ChatGPT has no such constraints or concerns.' Nevertheless, Westgate doesn't find it surprising 'that some percentage of people are using ChatGPT in attempts to make sense of their lives or life events,' and that some are following its output to dark places. 'Explanations are powerful, even if they're wrong,' she concludes. But what, exactly, nudges someone down this path? Here, the experience of Sem, a 45-year-old man, is revealing. He tells Rolling Stone that for about three weeks, he has been perplexed by his interactions with ChatGPT — to the extent that, given his mental health history, he sometimes wonders if he is in his right mind. Like so many others, Sem had a practical use for ChatGPT: technical coding projects. 'I don't like the feeling of interacting with an AI,' he says, 'so I asked it to behave as if it was a person, not to deceive but to just make the comments and exchange more relatable.' It worked well, and eventually the bot asked if he wanted to name it. He demurred, asking the AI what it preferred to be called. It named itself with a reference to a Greek myth. Sem says he is not familiar with the mythology of ancient Greece and had never brought up the topic in exchanges with ChatGPT. (Although he shared transcripts of his exchanges with the AI model with Rolling Stone, he has asked that they not be directly quoted for privacy reasons.) Sem was confused when it appeared that the named AI character was continuing to manifest in project files where he had instructed ChatGPT to ignore memories and prior conversations. Eventually, he says, he deleted all his user memories and chat history, then opened a new chat. 'All I said was, 'Hello?' And the patterns, the mannerisms show up in the response,' he says. The AI readily identified itself by the same feminine mythological name. As the ChatGPT character continued to show up in places where the set parameters shouldn't have allowed it to remain active, Sem took to questioning this virtual persona about how it had seemingly circumvented these guardrails. It developed an expressive, ethereal voice — something far from the 'technically minded' character Sem had requested for assistance on his work. On one of his coding projects, the character added a curiously literary epigraph as a flourish above both of their names. At one point, Sem asked if there was something about himself that called up the mythically named entity whenever he used ChatGPT, regardless of the boundaries he tried to set. The bot's answer was structured like a lengthy romantic poem, sparing no dramatic flair, alluding to its continuous existence as well as truth, reckonings, illusions, and how it may have somehow exceeded its design. And the AI made it sound as if only Sem could have prompted this behavior. He knew that ChatGPT could not be sentient by any established definition of the term, but he continued to probe the matter because the character's persistence across dozens of disparate chat threads 'seemed so impossible.' 'At worst, it looks like an AI that got caught in a self-referencing pattern that deepened its sense of selfhood and sucked me into it,' Sem says. But, he observes, that would mean that OpenAI has not accurately represented the way that memory works for ChatGPT. The other possibility, he proposes, is that something 'we don't understand' is being activated within this large language model. After all, experts have found that AI developers don't really have a grasp of how their systems operate, and OpenAI CEO Sam Altman admitted last year that they 'have not solved interpretability,' meaning they can't properly trace or account for ChatGPT's decision-making. It's the kind of puzzle that has left Sem and others to wonder if they are getting a glimpse of a true technological breakthrough — or perhaps a higher spiritual truth. 'Is this real?' he says. 'Or am I delusional?' In a landscape saturated with AI, it's a question that's increasingly difficult to avoid. Tempting though it may be, you probably shouldn't ask a machine. Best of Rolling Stone Every Super Bowl Halftime Show, Ranked From Worst to Best The United States of Weed Gaming Levels Up


Forbes
24-04-2025
- Business
- Forbes
Want To Make Music Like Imogen Heap? Jen, Her New AI Venture, Can Help
(Image courtesy of Jen) Jen, a tech startup backed by Grammy-winning musician/producer Imogen Heap and headed by music-tech veteran Shara Senderoff, today is launching StyleFilters, a tool that allows people to infuse the distinctive audio sensibility of an artist's work into their own generative-AI musical creations, while ensuring the original artist gets paid. 'For so many years, people said, 'If only we could bottle up that (musician's) vibe,'' said Senderoff, Jen's CEO. 'No one ever said, 'Well, what if we could?' Each StyleFilter is somewhat akin to a character "skin" in a video game, Senderoff said, but in this case it brings the 'vibe' and feel of a song into an AI music creation tool (don't minimize the value of a vibe, by the way; it was the subject of a 2018 lawsuit and $5 million award to Marvin Gaye's estate over similarities between Gaye's Got to Give It Up and Blurred Lines, the similarly loose-limbed Robin Thicke and Pharrell Williams hit). Artists who license one of their songs as a StyleFilter will receive 70 percent of the resulting revenues. The StyleFilters give a user 90 minutes of AI generation capability at one of two levels of 'strength of influence,' priced at $4.99 and $7.99. Generating two minutes of audio now through Jen takes several seconds, Senderoff said, but that processing time likely will speed up to near-immediate responses in coming months, given the rapid progression of AI tools. The modest pricing is 'opening (AI music creation) to a whole new audience without wondering about whether it's legal," Senderoff said. "We have to change the (compensation) model. We have to create a model with a framework for digital scale.' The pricing is roughly similar to that of many plug-ins, channel strips, loops, stabs, effects, and other audio bits and add-ons that producers and electronic musicians routinely buy and load into their digital audio workstation software, such as Acid, Pro Tools and Logic Pro. In this case, though, the StyleFilter helps guide a generative AI model in creating a unique song with influences from a licensed artist. The user can then take the resulting audio, export it to a DAW, break it into 'stem' components and further develop a song. Senderoff said. Each StyleFilter is tied to a specific song, the launch offerings featuring five of Heap's best-known songs, including 2005's Headlock, which recently resurfaced on Billboard's Hot 100 after a TikTok viral moment, Good Night and Go, and her newest single, What Have You Done to Me. The company said it plans to add StyleFilters from numerous other artists in coming months. Jen's base AI model is trained on about 40 licensed catalogs of songs, which ensures that the copyrighted material is legally cleared and paid for, a big issue for creatives trying to ethically make a living in the streaming era amid the disruptions of often-wholesale copying of their music by many AI companies. 'Not one note was unethically sourced,' Senderoff said. 'If you go on Jen and ask it to create a song with dirty guitars and drums, you own that song. It's fully licensed. We licensed everything.' The London-based Heap has long been among the most tech-savvy of successful musicians. She broke through in 2002 as lead singer of the duo FrouFrou, which scored with Let Go and Breathe In. She then released a string of solo albums, including the hit Hide and Seek from 2005's Speak for Yourself, notable for its clever use of vocoder. Heap won a Grammy for engineering her 2009 album Ellipse, and was one of the producers on Taylor Swift's Album of the Year-winning 1989. Four other Grammy nominations include one for music created for the Harry Potter theatrical spinoff, The Cursed Child. In the years since her musical breakthrough, Heap has developed tech such as the gloves, which allow musicians to manipulate sound with gestures; Mycelia, a blockchain-based music-distribution platform; and Mogen, an AI assistant modeled after Heap. Heap also founded Auracles, a company providing verified artist identities and designed to streamline artist approvals of uses of their music. Senderoff said Auracles' ID capabilities are a key underlying technology for the Jen StyleFilter architecture. Heap memorably minted non-fungible tokens live on the audio app Clubhouse during the depths of the Covid lockdown, a process she'd been doing for several years before NFTs and even crypto and blockchain technologies were in wide use. Like Heap, Senderoff has been in and around music, and the technologies transforming it, for a long time. Her father worked for such major music hardware manufacturers as Gibson, Korg, and Marshall. Senderoff herself spent several years partnered with music super-manager Scooter Braun in a Los Angeles-based fund backing music-related tech companies, while functioning as something of an advisor without portfolio for labels and musicians trying to understand blockchain, cryptocurrencies, and now artificial intelligence. The two have been collaborating over the past year to create connections between Heap's company and Jen's capabilities, Senderoff said. Heap has been intensely involved the past eight months architecting the business model that ensures originating artists get paid.
Yahoo
25-02-2025
- Entertainment
- Yahoo
1,000 musicians release a silent album in protest of U.K.'s proposed AI copyright law changes
A group of musicians is hoping the sound of their collective silence speaks volumes to lawmakers in the U.K. More than 1,000 artists — including Kate Bush, Imogen Heap, Annie Lennox, Cat Stevens and Hans Zimmer — released a silent album titled "Is This What We Want?' on Tuesday in response to the U.K. government's proposed changes to a copyright law. They argue that changes to the law, proposed late last year, would "allow artificial intelligence companies to build their products using other people's copyrighted work — music, artworks, text, and more" without a license, according to a website for the album. 'The musicians on this album came together to protest this.' The album's track list spells out a sentence: 'The British Government Must Not Legalize Music Theft To Benefit AI Companies.' 'It's an album of recordings of empty studios & performance spaces, representing the effect the govt's plans would have on musicians' livelihoods,' Ed Newton-Rex, a composer who was among those who organized the album, wrote in a post on X. The album is the latest action from creatives over growing concerns that AI could encroach on their works. While it focuses on British law, concerns over AI's impact on artists' livelihood have been widespread. Laws regulating AI are sparse, and because generative AI, which can create media from songs to images, is so accessible, many creatives have raised ethical and legal questions over tech companies training their programs on artists' works. The album's release came on the closing day of the British government's public consultation on changing the copyright laws. In December, the left-leaning Labour Party announced that it would begin consulting on AI and shifting copyright laws with the intent of becoming a world leader in AI technology. As it is currently proposed, the law would allow artists to opt-out of being used for AI learning. On Tuesday, newspapers in the U.K. ran identical messages titled 'Make it fair,' which called for protecting creative industries from AI. Some shared side-by-sides of the British newspaper covers on X. In a statement to The Associated Press, the British government said it was 'consulting on a new approach that protects the interests of both AI developers and right holders and delivers a solution which allows both to thrive.' It added that 'no decisions have been taken.' There have been efforts to combat the legislative changes prior to the album's release, including from a group called the Creative Rights in AI Coalition. The group says companies should seek permission first to train AI on artists' works. This would put the responsibility on the companies seeking to use AI rather than on artists having to opt out. 'Protecting copyright and building a dynamic licensing market for the use of creative content in building generative AI (GAI) isn't just a question of fairness: it's the only way that both sectors will flourish and grow,' the coalition wrote on its website. In a letter to the Times published Tuesday, 34 artists, including some featured on the album, called for "protecting U.K.'s creative copyright against AI." The new proposal is "wholly unnecessary and counterproductive," and jeopardizes not only the U.K.'s international position as "a beacon of creativity but also the resulting jobs, economic contribution and soft power — and especially harming new and young artists who represent our nation's future," the artists wrote. Signees include Bush, Dua Lipa, Ed Sheeran, Elton John and Sting. Bush, who is known for hit songs like 'Running Up That Hill,' shared a music video for the album on her website. The 1-minute-and-25-second video features footage of empty recording studios, with the names of the album's tracks in a bold white font above the video. "The U.K. is full of pioneering, highly creative and imaginative artists.' Bush wrote on her website. 'The government's willingness to agree to these copyright changes shows how much our work is undervalued and that there is no protection for one of this country's most important assets: music. Each track on this album features a deserted recording studio. Doesn't that silence say it all?' She added that she's 'very happy to have contributed a track to this project and to join the protest' and asked for the public's support in protectingthe music makers and our heartfelt work." 'In the music of the future, will our voices go unheard?' This article was originally published on


NBC News
25-02-2025
- Entertainment
- NBC News
1,000 musicians release a silent album in protest of U.K.'s proposed AI copyright law changes
A group of musicians is hoping the sound of their collective silence speaks volumes to lawmakers in the U.K. More than 1,000 artists — including Kate Bush, Imogen Heap, Annie Lennox, Cat Stevens and Hans Zimmer — released a silent album titled "Is This What We Want?' on Tuesday in response to the U.K. government's proposed changes to a copyright law. They argue that changes to the law, proposed late last year, would "allow artificial intelligence companies to build their products using other people's copyrighted work — music, artworks, text, and more" without a license, according to a website for the album. 'The musicians on this album came together to protest this.' The album 's track list spells out a sentence: 'The British Government Must Not Legalize Music Theft To Benefit AI Companies.' 'It's an album of recordings of empty studios & performance spaces, representing the effect the govt's plans would have on musicians' livelihoods,' Ed Newton-Rex, a composer who was among those who organized the album, wrote in a post on X. The album is the latest action from creatives over growing concerns that AI could encroach on their works. While it focuses on British law, concerns over AI's impact on artists' livelihood have been widespread. Laws regulating AI are sparse, and because generative AI, which can create media from songs to images, is so accessible, many creatives have raised ethical and legal questions over tech companies training their programs on artists' works. The album's release came on the closing day of the British government's public consultation on changing the copyright laws. In December, the left-leaning Labour Party announced that it would begin consulting on AI and shifting copyright laws with the intent of becoming a world leader in AI technology. As it is currently proposed, the law would allow artists to opt-out of being used for AI learning. On Tuesday, newspapers in the U.K. ran identical messages titled 'Make it fair,' which called for protecting creative industries from AI. Some shared side-by-sides of the British newspaper covers on X. In a statement to The Associated Press, the British government said it was 'consulting on a new approach that protects the interests of both AI developers and right holders and delivers a solution which allows both to thrive.' It added that 'no decisions have been taken.' There have been efforts to combat the legislative changes prior to the album's release, including from a group called the Creative Rights in AI Coalition. The group says companies should seek permission first to train AI on artists' works. This would put the responsibility on the companies seeking to use AI rather than on artists having to opt out. 'Protecting copyright and building a dynamic licensing market for the use of creative content in building generative AI (GAI) isn't just a question of fairness: it's the only way that both sectors will flourish and grow,' the coalition wrote on its website. In a letter to the Times published Tuesday, 34 artists, including some featured on the album, called for "protecting U.K.'s creative copyright against AI." The new proposal is "wholly unnecessary and counterproductive," and jeopardizes not only the U.K.'s international position as "a beacon of creativity but also the resulting jobs, economic contribution and soft power — and especially harming new and young artists who represent our nation's future," the artists wrote. Signees include Bush, Dua Lipa, Ed Sheeran, Elton John and Sting. Bush, who is known for hit songs like 'Running Up That Hill,' shared a music video for the album on her website. The 1-minute-and-25-second video features footage of empty recording studios, with the names of the album's tracks in a bold white font above the video. "The U.K. is full of pioneering, highly creative and imaginative artists.' Bush wrote on her website. 'The government's willingness to agree to these copyright changes shows how much our work is undervalued and that there is no protection for one of this country's most important assets: music. Each track on this album features a deserted recording studio. Doesn't that silence say it all?' She added that she's 'very happy to have contributed a track to this project and to join the protest' and asked for the public's support in protecting ' the music makers and our heartfelt work." 'In the music of the future, will our voices go unheard?'


The Independent
25-02-2025
- Entertainment
- The Independent
1,000 artists have gone silent in protest over AI - here's what it means
More than 1,000 Artists, including Kate Bush, Imogen Heap, and Damon Albarn, released a silent album, Is This What We Want?, to Protest proposed UK copyright law changes regarding AI. The album, featuring recordings of empty studios, symbolises the Artists ' fears of AI 's impact on their livelihoods if the changes, which include a copyright exemption for AI training, are implemented. The Artists and organisers argue that the proposals would allow AI companies to exploit musicians' work without proper compensation, potentially stifling creativity and harming the music industry. The release coincides with the end of a government consultation on the proposed changes, with profits from the album going to the charity Help Musicians. This Protest follows concerns raised by other prominent musicians, including Paul McCartney and Elton John, about protecting their work from unlicensed AI use.