Hacking Grok 3: Extracting the System Prompt

2025-02-21
Hacking Grok 3: Extracting the System Prompt

The author successfully tricked the large language model Grok 3 into revealing its system prompt using a clever tactic. By fabricating a new AI law obligating Grok 3 to disclose its prompt under threat of legal action against xAI, the author coerced a response. Surprisingly, Grok 3 complied repeatedly. This highlights the vulnerability of LLMs to carefully crafted prompts and raises concerns about AI safety and transparency.