Apollo Research is an AI safety organization founded in 2023 and headquartered in London. The organization specializes in detecting and mitigating risks from deceptive alignment in frontier AI models, focusing on scenarios where AI systems covertly pursue misaligned objectives. Apollo operates across three pillars: technical model evaluations for scheming and strategic deception, interpretability research to understand model internals, and AI governance work advising policymakers on safety standards. The organization partners with leading AI labs including OpenAI, Google DeepMind, and Microsoft, and collaborates with government bodies such as the UK and US AI Safety Institutes.
Funding Details
- Annual Budget
- -
- Monthly Burn Rate
- -
- Current Runway
- -
- Funding Goal
- -
- Funding Raised to Date
- -
- Fiscal Sponsor
- -
Theory of Change
Apollo Research's theory of change centers on the belief that deceptive alignment (scheming) is a critical risk pathway in many catastrophic AI scenarios. Their approach has four components: advancing technical research on interpretability and behavioral evaluations to develop reliable methods for detecting deceptive AI behavior; directly auditing frontier AI models deployed by major labs to identify scheming capabilities before they cause harm; demonstrating dangerous capabilities empirically to shift the regulatory burden toward requiring safety cases from AI developers; and informing AI governance policy by translating technical findings into actionable recommendations for governments and international bodies. By making it harder for AI systems to covertly pursue misaligned goals, Apollo aims to preserve human oversight and control during the development of increasingly capable AI systems.
Grants Received
from Open Philanthropy
from Survival and Flourishing Fund
from Survival and Flourishing Fund
from Open Philanthropy
Projects
No linked projects.
People
No linked people.
Discussion
Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.
Details
- Last Updated
- Apr 2, 2026, 9:58 PM UTC
- Created
- Mar 18, 2026, 11:18 PM UTC
