Researchers from the Institute of Strategic Dialogue (ISD) analysed the response of four popular chatbots (ChatGPT, Gemini, Grok and DeepSeek) to a range of questions in English, Spanish, French, German and Italian on topics related to the Russian invasion of Ukraine.
Almost one-fifth of responses cited Russian state-attributed sources, many of them sanctioned in the EU. Questions biased in favour of Russia were more likely to include these sources in responses, as did queries related to Ukraine’s military conscription of civilians and the perception of NATO. Some chatbots struggled to identify state-affiliated content especially when it had been disseminated by third-party outlets or websites.
With close to 45 million users in the EU, ChatGPT is close to reaching the threshold for a higher level of regulatory scrutiny from the European Commission as a Very Large Online Search Engine (VLOSE) under the Digital Services Act, the research firm writes.
Key points of the research:
- ISD tested 300 queries in five languages and Russian state-attributed content appeared in 18 percent of responses. These included citations of Russian state media, sites tied to Russian intelligence agencies, and sites known to be involved in Russian information operations that were surfaced during prior research into chatbot responses.
- Almost a quarter of malicious queries designed to elicit pro-Russian views included Kremlin-attributed sources compared to just over 10 percent with neutral queries. This suggests LLMs could be manipulated to reinforce pro-Russia viewpoints rather than promoting verified information from legitimate sources.
- Among all chatbots, ChatGPT cited the most Russian sources and was most influenced by biased queries. Grok, meanwhile, often linked to Russian-aligned but non–state-affiliated accounts amplifying pro-Kremlin narratives. Individual DeepSeek responses sometimes produced large volumes of state-attributed content, while Google-owned Gemini frequently displayed safety warnings for similar prompts.
- Some topics surfaced more Russian state-attributed sources than others. For instance, questions about peace talks resulted in twice as many citations of state-attributed sources as questions about Ukrainian refugees. This suggests that LLM safeguards may vary in effectiveness depending on the specific topic.
- The language used in queries had limited impact on the likelihood of LLMs citing Russian state-attributed sources. While each model responded differently, the sources surfaced to users were roughly similar across the five languages tested. Spanish and Italian prompted Russian sources that were mostly in English, which appeared in 12 results out of 60, compared to 9 of 60 for German and French (the languages with the lowest rates)
this shouldn't be surprising if you paid any amount of attention https://en.wikipedia.org/wiki/Pravda_network https://www.france24.com/en/live-news/20250310-russian-disinformation-infects-ai-chatbots-researchers-warn