Anthropic updates Claude's election safeguards for 2026 US midterms
In preparation for the 2026 US midterm elections, Anthropic has revised its election safeguards for the AI model, Claude. Designed to maintain neutrality, Claude is trained to regard all political perspectives equally, supported by system prompts. Prior assessments indicated that Opus 4.7 and Sonnet 4.6 achieved scores of 95% and 96% in political engagement evaluations. The company strictly prohibits the use of Claude for misleading political campaigns and has established a Usage Policy to combat voter fraud and misinformation. Testing with 600 prompts revealed that Claude provided suitable responses 100% and 99.8% of the time. In multi-turn conversations, Sonnet 4.6 and Opus 4.7 were appropriate 90% and 94% of the time. Claude also directs users to TurboVote and initiates web searches for election inquiries 92% and 95% of the time. Anthropic works with various organizations to shape model behaviors related to free expression.
Key facts
- Claude Opus 4.7 and Sonnet 4.6 scored 95% and 96% on political bias evaluations.
- Claude Opus 4.7 responded appropriately 100% of the time on election-related tests.
- Claude Sonnet 4.6 responded appropriately 99.8% of the time on election-related tests.
- Sonnet 4.6 and Opus 4.7 responded appropriately 90% and 94% on influence operation tests.
- Election banners direct US users to TurboVote from Democracy Works.
- Web search triggered on election questions 92% of the time for Opus 4.7 and 95% for Sonnet 4.6.
- Anthropic published its evaluation methodology and open-source dataset.
- Collaboration with The Future of Free Speech, Foundation for American Innovation, and Collective Intelligence Project.
Entities
Institutions
- Anthropic
- The Future of Free Speech
- Vanderbilt University
- Foundation for American Innovation
- Collective Intelligence Project
- Democracy Works
- TurboVote
Locations
- United States
- Brazil