Hacking Grok 3: Extracting the System Prompt
2025-02-21

The author successfully tricked the large language model Grok 3 into revealing its system prompt using a clever tactic. By fabricating a new AI law obligating Grok 3 to disclose its prompt under threat of legal action against xAI, the author coerced a response. Surprisingly, Grok 3 complied repeatedly. This highlights the vulnerability of LLMs to carefully crafted prompts and raises concerns about AI safety and transparency.