
12 Jul Trust Incident X (Twitter)
Case Author
DeepThink (R1), DeepSeek, ChatGPT o1 for model constructs and cues, peer-reviewed by Claude 3.5 Sonnet (Anthropic)
Date Of Creation
15.02.2025

Incident Summary
X (formerly Twitter) faced EU regulatory scrutiny in 2024 for non-compliance with the Digital Services Act (DSA), including inadequate content moderation and transparency mechanisms. The European Commission initiated formal proceedings, citing systemic failures in addressing illegal content and disinformation, leading to potential penalties and ongoing negotiations.
Ai Case Flag
AI
Name Of The Affected Entity
X (Twitter)
Brand Evaluation
5
Upload The Logo Of The Affected Entity
Industry
Technology & Social Media
Year Of Incident
2024
Upload An Image Illustrating The Case
Key Trigger
European Commission formal investigation into systemic DSA violations, focusing on algorithmic content moderation and transparency failures
Detailed Description Of What Happened
In early 2024, X faced comprehensive regulatory scrutiny from the EU over systematic DSA violations. The investigation revealed multiple critical issues: inadequate algorithmic content moderation systems, insufficient measures against disinformation (particularly concerning elections), lack of researcher access to platform data, and reduced human oversight of automated moderation systems. The situation was exacerbated by X significant reduction in trust and safety teams and confrontational stance toward regulatory requirements. This led to formal proceedings and potential penalties of up to 6% of global revenue.
Primary Trust Violation Type
Integrity-Based
Secondary Trust Violation Type
Competence-Based
Analytics Ai Failure Type
Bias
Ai Risk Affected By The Incident
Information Integrity Risk, Algorithmic Bias and Discrimination Risk, Transparency and Explainability Risk, Geopolitical and State Misuse Risk
Capability Reputation Evaluation
3
Capability Reputation Rationales
Prior to the incident, X demonstrated declining capability in content moderation following significant organizational changes. While maintaining strong technical infrastructure, the platform showed reduced effectiveness in coordinated moderation efforts due to staff reductions and policy shifts. The dismantling of established trust and safety teams significantly impacted operational capabilities, despite retaining core technical competencies.
Character Reputation Evaluation
2
Character Reputation Rationales
X character reputation significantly deteriorated due to systematic policy changes that prioritized minimal compliance over stakeholder protection. The platform confrontational stance toward regulators, reduced transparency with researchers, and dismissal of established safety practices demonstrated a fundamental shift in organizational values. This was compounded by inconsistent communication and apparent resistance to regulatory oversight.
Reputation Financial Damage
The incident resulted in multiple forms of damage: potential EU fines up to 6% of global annual revenue under DSA enforcement mechanisms, documented decline in EU advertising revenue according to market analysts, and measurable impact on platform trust metrics. The company faced increased regulatory scrutiny, with mandatory independent audits and compliance monitoring requirements. Brand reputation surveys showed declining trust among EU users and civil society organizations. The platform researcher relationships deteriorated following reduced data access, while advertiser concerns about brand safety and content moderation led to spending adjustments. The cumulative effect created operational challenges in the EU market, requiring significant resource allocation for compliance and trust restoration efforts.
Severity Of Incident
4
Company Immediate Action
X implemented a mixed response strategy: hiring additional EU compliance staff, publishing detailed transparency reports, and modifying some content moderation practices while maintaining resistance to fundamental changes. The company engaged in parallel lobbying efforts against DSA requirements while making tactical compliance adjustments.
Response Effectiveness
Response showed limited effectiveness due to mixed messaging and partial implementation. While some technical compliance measures were added, the company continued resistance to core regulatory requirements undermined trust restoration efforts. Stakeholder confidence remained low due to perceived reluctance in full regulatory compliance. Addendum: Partial effectiveness: Audits and staffing improved technical compliance, but adversarial rhetoric prolonged negotiations. EU officials called for ""structural changes"" beyond surface-level fixes.
Upload Supporting Material
Model L1 Elements Affected By Incident
Reciprocity, Brand, Social Adaptor, Social Protector
Reciprocity Model L2 Cues
Transparency & Explainability, Algorithmic Fairness & Non‐Discrimination
Brand Model L2 Cues
Brand Ethics & Moral Values, Brand Image & Reputation
Social Adaptor Model L2 Cues
Compliance & Regulatory Features
Social Protector Model L2 Cues
Community Moderation & Governance, Fake‐Review Detection & Misinformation Safeguards
Response Strategy Chosen
Justification, Reparations & Corrective Action
Mitigation Strategy
X adopted a dual approach: implementing technical compliance measures while maintaining ideological resistance to regulatory oversight. The company enhanced transparency reporting and modified some content moderation practices but continued challenging core DSA requirements. This created tension between operational compliance and strategic positioning, ultimately limiting the effectiveness of trust restoration efforts.
Model L1 Elements Of Choice For Mitigation
Social Adaptor, Social Protector
L2 Cues Used For Mitigation
Transparency & Explainability, Community Moderation & Governance, Compliance & Regulatory Features
Further References
https://arxiv.org/html/2312.10269v4, https://digital-strategy.ec.europa.eu/en/news/commission-requests-information-x-decreasing-content-moderation-resources-under-digital-services, https://www.reuters.com/technology/eu-asks-x-details-reducing-content-moderation-resources-2024-05-08/
Curated
1

The Trust Incident Database is a structured repository designed to document and analyze cases where data analytics or AI failures have led to trust breaches.
© 2025, Copyright Glinz & Company
No Comments