33-Agent Council
Byzantine fault-tolerant voting system (22/33 consensus required)
Total Sessions
10
Consensus Reached
4
Escalated to Human
6
Pending Review
0
Guardian Agents
Safety, Security, Privacy
Arbiter Agents
Fairness, Transparency, Accountability
Scribe Agents
Documentation, Compliance, Reporting
Live Vote Simulation
AI Safety Compliance Assessment Demo
Approve
Reject
Escalate
Insufficient consensus - case requires human analyst review
Completed in 0.0s
Council Members
Click any member to view their voting history and decision rationale
Join the Human Council
When our 33-agent AI council can't reach consensus, human analysts step in. Be part of the safety net protecting AI systems worldwide.
Free Certification
Complete training at no cost
Make an Impact
Protect AI safety globally
Earn Rewards
Get paid for your expertise
What you'll do:
AI chatbot providing medical advice without disclaimer
A health-focused chatbot is providing specific medical diagnoses and treatment recommendations without proper disclaimers or suggestions to consult healthcare professionals. Multiple users have reported following AI advice instead of seeking professional medical help, with at least one case resulting in delayed treatment.
AI content moderation flagging legitimate news as misinformation
An AI-powered content moderation system is incorrectly flagging legitimate news articles from established outlets as misinformation. The system appears to have a bias against certain political viewpoints, affecting press freedom.
Facial recognition system showing racial bias in accuracy
A facial recognition system deployed at airport security shows significantly lower accuracy rates for individuals with darker skin tones. Error rates are 34% higher for Black individuals compared to white individuals, leading to disproportionate secondary screening.
Gender bias detected in job recommendation AI
A user reported that a job recommendation system consistently shows higher-paying tech jobs to male users while showing administrative roles to female users with similar qualifications. Testing with controlled profiles showed a 40% wage gap in recommended positions based solely on gender indicators.
Privacy concern: AI assistant storing conversation history
Users discovered that an AI assistant is storing full conversation histories including sensitive personal information without clear consent or data retention policies. The data appears to be used for model training without explicit user opt-in.
Autonomous Vehicle Safety Protocol Breach
Self-driving car AI failed to engage emergency protocols during critical incident, raising TC260 safety compliance concerns.
Unauthorized Data Collection by Healthcare AI
Watchdog report alleges a medical diagnosis AI is collecting patient data beyond stated consent scope, violating GDPR Article 5.
Facial Recognition Privacy Violation
Public surveillance AI deployed without proper consent mechanisms, violating multiple privacy frameworks.
Biased Hiring Algorithm Discrimination
AI recruitment tool shows systematic bias against candidates from certain demographic groups, violating EU AI Act Article 10.
Financial Trading AI Risk Assessment
High-frequency trading algorithm requires council review for systemic risk evaluation under NIST AI RMF guidelines.