Protecting AI against foreign influence operations designed to infect LLMs
NewsGuard’s Foreign Adversary Influence in LLMs Safety Service (FAILSafe) helps AI companies detect and defend against foreign influence operations aimed at tainting AI responses with state-sponsored information operations and propaganda.
Created in response to a groundbreaking NewsGuard audit that found Russian website networks had infected top AI tools, leading those tools to repeat propaganda narratives 33% of the time, FAILSafe provides AI companies with real-time data, verified by analysts with an expertise in foreign malign influence, on narratives and sources involved in adverse influence operations run by the Russian, Chinese, and Iranian governments.