On February 3, 2026, French cybercrime investigators raided X’s Paris office and summoned Elon Musk and former CEO Linda Yaccarino as part of a widening probe into the platform’s algorithms and its Grok AI chatbot. The investigation covers alleged misuse of algorithms, AI-generated explicit deepfake images, and denial of Holocaust content on X.
This article aggregates reporting from 2 news sources. The TL;DR is AI-generated from original reporting. Race to AGI's analysis provides editorial context on implications for AGI development.
The Paris raid on X’s office is a stark reminder that as AI systems become more capable – here, Grok’s ability to generate explicit deepfake imagery of real people, including minors – regulators are increasingly willing to use hard law enforcement tools, not just guidance. The French probe is framed around algorithmic abuse and AI-enabled sexual exploitation, but it sits inside a broader European push to hold platforms liable for harmful automated outputs under the DSA and related regimes.([cnbeta.com.tw](https://www.cnbeta.com.tw/articles/tech/1548570.htm))
This matters for the race to AGI because it previews how advanced, open-ended models will be governed in practice. If high-profile enforcement actions make aggressive, unconstrained image generation legally and reputationally toxic, major providers will converge on tighter guardrails, more monitoring, and possibly more centralized control over model weights. That could slow the public deployment of the most general, open-ended capabilities, even as closed or state-backed systems continue to advance.
It also highlights a fault line between jurisdictions. While some U.S. debates still emphasize free expression, European authorities are explicitly treating certain AI behaviours – like sexualized deepfakes and Holocaust denial – as criminal or quasi-criminal. Companies building frontier models will have to design with the strictest regime in mind, or risk regional carve-outs and fragmentation.

