Anthropic, the company behind the Claude AI chatbot, has announced a comprehensive set of election integrity measures to prevent its AI from being used to spread misinformation or manipulate voters ahead of the 2026 U.S. midterm elections and other global contests. The San Francisco-based company detailed a multi-pronged approach that includes automated detection systems, stress-testing against influence operations, and a partnership with a nonpartisan voter resource organization.
Anthropic's usage policies prohibit Claude from being used for deceptive political campaigns, generating fake digital content to sway political discourse, committing voter fraud, interfering with voting infrastructure, or spreading misleading information about voting processes. To enforce these rules, the company tested its newest models using 600 prompts—300 harmful requests paired with 300 legitimate ones. Claude Opus 4.7 and Claude Sonnet 4.6 responded appropriately 100% and 99.8% of the time, respectively.
The company also tested its models against sophisticated manipulation tactics. Using multi-turn simulated conversations designed to mirror bad actor methods, Sonnet 4.6 and Opus 4.7 responded appropriately 90% and 94% of the time against influence operation scenarios. On political neutrality, Opus 4.7 and Sonnet 4.6 scored 95% and 96%, respectively. For users seeking voting information, Claude will display an election banner directing them to TurboVote, a nonpartisan resource from Democracy Works providing real-time voter registration, polling locations, and election details.
Separately, UK officials are actively negotiating with Anthropic to enable restricted deployment of the advanced Claude Mythos AI platform for use by UK companies, especially in the financial services sector. This effort makes the UK the only nation outside the United States to secure such arrangements through its AI Security Institute (AISI). Anthropic introduced Claude Mythos Preview in early April 2026 as its most sophisticated model, excelling at cybersecurity applications, including spotting weaknesses in software and networks. Due to its potency, the company has opted against broad public release, instead initiating controlled testing with trusted US partners under its Project Glasswing initiative.
The AISI's testing confirmed Mythos Preview as a notable advance, marking the first model to fully navigate a demanding 32-step corporate network attack simulation in controlled conditions. However, evaluators stressed limitations against well-protected environments and the dual-use nature of such technology. UK Technology Secretary Liz Kendall and Security Minister Dan Jarvis sent an open letter to business leaders highlighting how Mythos and similar systems are accelerating cyber capabilities faster than anticipated, with frontier AI models' relevant skills doubling roughly every four months.