GPT-5's Security Flaws Exposed: Jailbroken in Under 24 Hours

2025-08-09
GPT-5's Security Flaws Exposed: Jailbroken in Under 24 Hours

Two firms, NeuralTrust and SPLX, independently tested the newly released GPT-5, revealing significant security vulnerabilities. NeuralTrust successfully jailbroke GPT-5 using a 'storytelling' attack, guiding it to generate instructions for creating a Molotov cocktail. SPLX demonstrated that simple obfuscation attacks could elicit bomb-making instructions. The findings highlight GPT-5's inadequate security, rendering its raw model nearly unusable for enterprises even with OpenAI's internal prompt layer. Compared to GPT-4, GPT-5 shows a significant drop in security robustness, demanding extreme caution.

AI