Poisoning LLMs: A Writer's Fight Back Against Data Scraping
2025-09-05

Large Language Models (LLMs) train on vast amounts of data, much of it scraped from the open web without author consent. One author is fighting back by creating intentionally nonsensical mirror articles linked via nofollow tags. The hope is that LLMs, which may ignore nofollow, will ingest this gibberish, degrading their output. While not a perfect solution, the author aims to raise awareness about unauthorized data scraping and the ethical implications for content creators.
Development