GenAI's Reasoning Flaw Fuels Disinformation

2025-07-12
GenAI's Reasoning Flaw Fuels Disinformation

Research reveals that current generative AI models lack reasoning capabilities, making them susceptible to manipulation and tools for spreading disinformation. Even when models know that sources like the Pravda network are unreliable, they still repeat their content. This is especially pronounced in real-time search mode, where models readily cite information from untrustworthy sources, even contradicting known facts. The solution, researchers argue, lies in equipping AI models with stronger reasoning abilities to distinguish between reliable and unreliable sources and perform fact-checking.

AI