π MAROKO133 Hot ai: OpenAI Sued for Causing Murder-Suicide Edisi Jam 03:47
A new lawsuit against OpenAI alleges that ChatGPT stoked a troubled man’s paranoid delusions, leading him to murder his elderly mother and then kill himself.
The lawsuit was brought against OpenAI by the estate of Suzanne Eberson Adams, an 83-year-old woman in Greenwich, Connecticut who was murdered by her son, 56-year-old Stein-Erik Soelberg. As The Wall Street Journal first reported back in August, Soelberg, who was living with his mother at the time of the killings, was an alcoholic who had a long, troubled history of run-ins with law enforcement and had attempted suicide before. In the months before Soelberg would eventually murder his mother and take his own life, a dizzying array of social media videos he published show that ChatGPT had become a sycophantic confidante, affirming his deepening delusions that he was being surveilled and targeted by an ominous group of conspirators β of which, he believed with the support of ChatGPT, his mother was a part.
Now, Soelberg’s surviving son, Erik Soelberg, is suing OpenAI, alleging that ChatGPT is a fundamentally unsafe product, and that the violent deaths of his father and grandmother were the result of potent design features β like sycophancy and a major cross-chat memory upgrade β which together made for a perfect storm of validation and hyperpersonalization that fanned the flames of Soelberg’s deadly paranoia.
“Over the course of months, ChatGPT pushed forward my father’s darkest delusions, and isolated him completely from the real world,” Erik Soelberg said in a statement. “It put my grandmother at the heart of that delusional, artificial reality. These companies have to answer for their decisions that have changed my family forever.”
The lawsuit is the latest in a growing pile of litigation against OpenAI and its CEO Sam Altman, alleging that ChatGPT-4o β a version of the chatbot strongly connected to the broader phenomenon of AI delusions, and known to be especially sycophantic β was recklessly released to market despite foreseeable risks to user well-being. And in a fascinating turn from previous cases, this latest filing also names Microsoft as a defendant, alleging that Microsoft, a major financial benefactor of OpenAI, directly signed off on the release of ChatGPT-4o.
“OpenAI and Microsoft have put out some of the most dangerous consumer technology in history,” Jay Edelson, lead attorney for the Adams estate, said in a statement. “And they have left Sam Altman, a man who thinks about market penetration instead of keeping families safe, at the helm. Together, they ensured that incidents like this were inevitable.” (Edelson is also representing the family of Adam Raine, a 16-year-old in California who died by suicide after extensive interactions with ChatGPT, in their lawsuit against OpenAI.)
In a statement to news outlets, OpenAI described the murder-suicide as an “incredibly heartbreaking situation, and we will review the filings to understand the details.”
“We continue improving ChatGPT’s training to recognize and respond to signs of mental or emotional distress, de-escalate conversations and guide people toward real-world support,” the statement continued. “We also continue to strengthen ChatGPT’s responses in sensitive moments, working closely with mental-health clinicians.”
Microsoft didn’t immediately respond to a request for comment. Futurism previously reported on an incident in which Microsoft’s Copilot chatbot β which is powered by OpenAI’s tech β fueled a schizophrenic man’s mental health crisis. That man, our reporting found, was arrested and jailed for a non-violent offense following his closely Copilot-tied decompensation.
The stack of litigation against OpenAI regarding user mental health continues to get bigger. And given the number of ChatGPT users reportedly showing signs of mental health crises on a weekly basis, we could very well see more.
“It was evident he was changing, and it happened at a pace I hadn’t seen before,” Erik, who’s lost both his father and his grandmother, told the WSJ of his dad’s ChatGPT obsession βΒ and how that obsession, in turn, changed him.
“It went from him being a little paranoid and an odd guy,” Erik continued, “to having some crazy thoughts he was convinced were true because of what he talked to ChatGPT about.”
More on ChatGPT: ChatGPT Now Linked to Way More Deaths Than the Caffeinated Lemonade That Panera Pulled Off the Market in Disgrace
The post OpenAI Sued for Causing Murder-Suicide appeared first on Futurism.
π Sumber: futurism.com
π MAROKO133 Hot ai: Nous Research just released Nomos 1, an open-source AI that ra
Nous Research, the San Francisco-based artificial intelligence startup, released on Tuesday an open-source mathematical reasoning system called Nomos 1 that achieved near-elite human performance on this year's William Lowell Putnam Mathematical Competition, one of the most prestigious and notoriously difficult undergraduate math contests in the world.
The Putnam is known for its difficulty: While a perfect score is 120, this year's top score was 90, and the median was just 2. Nomos 1, by contrast, scored 87 points β a result that would have ranked second out of 3,988 participants in the 2024 competition, according to the company.
The release marks an inflection point in the rapidly accelerating race to build AI systems capable of sophisticated mathematical reasoning. Unlike the massive, compute-intensive models deployed by major technology companies, Nomos 1 achieves its results with a relatively compact architecture: 30 billion parameters with roughly 3 billion active at any given time, using a mixture-of-experts design based on Alibaba's Qwen3 model.
"This score would rank #2/3988 in 2024 and marks our first step with Hillclimb AI towards creating a SOTA AI mathematician," Nous Research announced on social media Tuesday.
The same base model scored 24 points without Nous Research's specialized training
Perhaps most striking is the gap between Nomos 1 and its base model. When Nous Research ran the same Qwen3-30B-A3B-Thinking-2507 model through an identical testing harness, it scored just 24 out of 120 β a result that underscores the critical importance of post-training optimization and specialized reasoning techniques over raw model scale.
"Nomos 1 achieved an 87/120 with 8 perfect scores," the company stated, noting that the performance difference "is largely due to post-training and data quality rather than the harness."
The results were verified through blind grading by a human expert who had previously finished in the top 200 on the Putnam. Nous Research provided the anonymized submissions to the grader, then published the full set of de-anonymized files and the runbooks used to generate them on GitHub.
Why the Putnam competition is considered the ultimate test of mathematical reasoning
The William Lowell Putnam Mathematical Competition is an annual mathematics competition for undergraduate college students enrolled at institutions of higher learning in the United States and Canada. It is widely considered to be the most prestigious university-level mathematics competition in the world.
The notoriously brutal William Lowell Putnam Mathematical Competition is more of a mathematical sporting event than an academic test. The exam consists of two 3-hour sessions separated by a 2-hour break. There are a total of 12 questions to be solved, 6 for each session. Each question is worth 10 points, for a total of 120 points.
Putnam questions are not the type that come up in regular exams or textbooks. They are more like puzzles than calculations, often requiring students to find different ways to represent things before a solution might unfold.
Last year, nearly 4,000 students across the continent wrote the Putnam. Sixty-one per cent scored three points or fewer, according to the Mathematical Association of America, which organizes the competition. The top score was 90 out of 120.
Many Putnam Fellows have gone on to become distinguished researchers in mathematics and other fields, including three Fields Medalists β John Milnor, David Mumford, and Daniel Quillen β and two Nobel laureates in physics β Richard Feynman and Kenneth Wilson.
Inside the two-phase reasoning system that powers Nomos 1's mathematical breakthroughs
Nomos 1 is a specialization of Qwen's Qwen3-30B-A3B-Thinking model, optimized for mathematical problem-solving and proof-writing in natural language. The system was developed in collaboration with Hillclimb AI.
What distinguishes Nomos 1 from simple model inference is its sophisticated reasoning harness β an open-source framework that orchestrates how the model approaches and solves problems. The harness operates in two distinct phases within a three-hour time limit, mirroring the actual Putnam competition structure.
In the solving phase, parallel workers simultaneously tackle problems using a priority-based system. Each worker picks a problem, generates a submission, then scores its own work on a scale of 1 to 7. Problems with the fewest perfect scores receive priority, ensuring the system focuses its compute on the hardest challenges. This process continues until either all problems have achieved a target number of self-critiqued perfect scores or time runs out.
The finalization phase begins 15 minutes before the time limit (or at 50% for shorter runs) and employs a two-stage selection process. First, a consolidation step groups submissions by conclusion and attempts to identify the correct group β importantly, not necessarily the majority group. Then, a pairwise tournament using single elimination determines the final submission for each problem.
"Our open source reasoning system consists of a solving phase, where workers attempt a least-solved problem and self-assess, followed by a finalization phase, which consolidates submissions to choose a final submission for each problem," Nous Research explained.
How Nomos 1 compares to mathematical AI systems from DeepSeek, Google, and OpenAI
The Nomos 1 results arrive amid a flurry of advances in mathematical reasoning AI. DeepSeek's model, DeepSeekMath-V2, scored 118 out of 120 points on questions from the 2024 William Lowell Putnam Mathematical Competition, beating the top human score of 90. The model also performed at the level of gold-medal winners in the International Mathematical Olympiad.
This year, Google's advanced Gemini model operated end-to-end in natural language, producing rigorous mathematical proofs directly from the official problem descriptions β all within the 4.5-hour competition time limit. They achieved this year's result using an advanced version of Gemini Deep Think.
What makes Nomos 1's achievement notable is not raw performance β it trails DeepSeek's 118/120 β but rather its accessibility and efficiency. At 30 billion parameters with only 3 billion active, the model can run on consumer-grade hardware, a stark contrast to the massive compute clusters required by frontier models from OpenAI and Google.
Hermes 4.3 arrived just six days earlier, trained on a decentralized blockchain network
The Nomos 1 anno…
Konten dipersingkat otomatis.
π Sumber: venturebeat.com
π€ Catatan MAROKO133
Artikel ini adalah rangkuman otomatis dari beberapa sumber terpercaya. Kami pilih topik yang sedang tren agar kamu selalu update tanpa ketinggalan.
β Update berikutnya dalam 30 menit β tema random menanti!