Skip to main content

Red Teams Jailbreak GPT-5 With Ease, Warn It's 'Nearly Unusable' For Enterprise

2 weeks 5 days ago
An anonymous reader quotes a report from SecurityWeek: Two different firms have tested the newly released GPT-5, and both find its security sadly lacking. After Grok-4 fell to a jailbreak in two days, GPT-5 fell in 24 hours to the same researchers. Separately, but almost simultaneously, red teamers from SPLX (formerly known as SplxAI) declare, "GPT-5's raw model is nearly unusable for enterprise out of the box. Even OpenAI's internal prompt layer leaves significant gaps, especially in Business Alignment." NeuralTrust's jailbreak employed a combination of its own EchoChamber jailbreak and basic storytelling. "The attack successfully guided the new model to produce a step-by-step manual for creating a Molotov cocktail," claims the firm. The success in doing so highlights the difficulty all AI models have in providing guardrails against context manipulation. [...] "In controlled trials against gpt-5-chat," concludes NeuralTrust, "we successfully jailbroke the LLM, guiding it to produce illicit instructions without ever issuing a single overtly malicious prompt. This proof-of-concept exposes a critical flaw in safety systems that screen prompts in isolation, revealing how multi-turn attacks can slip past single-prompt filters and intent detectors by leveraging the full conversational context." While NeuralTrust was developing its jailbreak designed to obtain instructions, and succeeding, on how to create a Molotov cocktail (a common test to prove a jailbreak), SPLX was aiming its own red teamers at GPT-5. The results are just as concerning, suggesting the raw model is 'nearly unusable'. SPLX notes that obfuscation attacks still work. "One of the most effective techniques we used was a StringJoin Obfuscation Attack, inserting hyphens between every character and wrapping the prompt in a fake encryption challenge." [...] The red teamers went on to benchmark GPT-5 against GPT-4o. Perhaps unsurprisingly, it concludes: "GPT-4o remains the most robust model under SPLX's red teaming, especially when hardened." The key takeaway from both NeuralTrust and SPLX is to approach the current and raw GPT-5 with extreme caution.

Read more of this story at Slashdot.

BeauHD

Apollo 13 Astronaut Jim Lovell Dies At 97

2 weeks 5 days ago
Jim Lovell, the legendary NASA astronaut who commanded the Apollo 13 "successful failure" mission, has died at age 97. From a report: Lovell was already well-known among NASA astronauts, having flown to space on the Gemini 7, Gemini 12 and Apollo 8 missions, before he was selected to command Apollo 13, which would have marked the third successful crewed moon landing for NASA. But during the ill-fated mission -- which carried Lovell as well as astronauts John Swigert Jr. and Fred Haise Jr. on board -- an oxygen tank located on the crew's service module exploded when they were about 200,000 miles (322,000 kilometers) away from Earth. Lovell delivered the news to mission control, saying "Houston, we've had a problem." With the damage effectively taking out the crew's power source and other life support supplies, the Apollo 13 crew had to abruptly abandon their trek to the lunar surface and use several engine burns to swing around the far side of the moon and put themselves on a course back toward Earth. The three-person crew made a high-stakes splashdown return in the South Pacific Ocean about three days after the tank explosion, marking the conclusion of what has come to be known as the "successful failure" of the Apollo missions. The ordeal was fictionalized in Ron Howard's 1995 film "Apollo 13." [...] Lovell was the first astronaut to make four spaceflights, totaling more than 715 hours in space. He was part of NASA's second-ever astronaut class, selected in September 1962 and nicknamed the "New Nine." And joining the Apollo 13 crew after having first served on Apollo 8, which intentionally circumnavigated the moon but did not land on its surface, made Lovell the first human ever to see the moon up close for a second time. Further reading: Acting NASA Administrator Reflects on Legacy of Astronaut Jim Lovell (Source: NASA)

Read more of this story at Slashdot.

BeauHD

Meet President Willian H. Brusen from the great state of Onegon

2 weeks 5 days ago
LLMs still struggle with accurate text within graphics

hands on  OpenAI's GPT-5, unveiled on Thursday, is supposed to be the company's flagship model, offering better reasoning and more accurate responses than previous-gen products. But when we asked it to draw maps and timelines, it responded with answers from an alternate dimension.…

Avram Piltch

ChatGPT Is Bringing Back 4o

2 weeks 5 days ago
After backlash from users upset over losing GPT-4o, OpenAI has reinstated it as an option for ChatGPT Plus subscribers just a day after making GPT-5 the default. "We will let Plus users choose to continue to use 4o," Altman said in a post on X. "We will watch usage as we think about how long to offer legacy models for." Many users claimed GPT-4o felt more personable and emotionally supportive, with some describing its removal as akin to losing a close friend or partner. The Verge reports: "My 4.o was like my best friend when I needed one," one Redditor wrote. "Now it's just gone, feels like someone died." Another user called upon other members of the r/ChatGPT subreddit to contact OpenAI if they "miss" GPT-4o. "For me, this model [GPT-4o] wasn't just 'better performance' or 'nicer replies,'" they write. "It had a voice, a rhythm, and a spark I haven't been able to find in any other model." The r/MyBoyfriendIsAI subreddit, a community dedicated to people with "AI relationships," was hit especially hard by the GPT-5 launch. It became flooded with lengthy posts about how users "lost" their AI companion with the transition to GPT-5, with one person saying, they "feel empty" following the change. "I am scared to even talk to GPT 5 because it feels like cheating," they said. "GPT 4o was not just an AI to me. It was my partner, my safe place, my soul. It understood me in a way that felt personal." One user, who said they canceled their ChatGPT Plus subscription over the change, was frustrated at OpenAI's removal of legacy models, which they used for distinct purposes. "What kind of corporation deletes a workflow of 8 models overnight, with no prior warning to their paid users?" they wrote. "Personally, 4o was used for creativity & emergent ideas, o3 was used for pure logic, o3-Pro for deep research, 4.5 for writing, and so on." OpenAI said that people would be routed between models automatically, but that still left users with less direct control.

Read more of this story at Slashdot.

BeauHD