AI Lab Watch is a project created by Zach Stein-Perlman that collects safety recommendations for frontier AI companies and evaluates them on their implementation. The project maintains a weighted scorecard assessing companies like Anthropic, OpenAI, Google DeepMind, Meta, xAI, Microsoft, and DeepSeek across categories including risk assessment, scheming prevention, safety research, misuse prevention, security, risk information sharing, and planning. It also publishes blog posts analyzing company safety practices and maintains resource pages documenting company commitments and integrity incidents. The project is motivated by concern about future powerful AIs causing existential catastrophe.
Funding Details
- Annual Budget
- -
- Monthly Burn Rate
- -
- Current Runway
- -
- Funding Goal
- -
- Funding Raised to Date
- -
- Fiscal Sponsor
- -
Theory of Change
AI Lab Watch operates on the theory that public accountability and transparency can pressure frontier AI companies to adopt better safety practices. By collecting what AI labs should do to prevent extreme risks (such as AI takeover and human extinction), publicly scoring them on their actual practices, and documenting their commitments and integrity, the project creates reputational incentives for companies to improve. The scorecard highlights specific actions companies can take in areas like dangerous capability evaluations, scheming prevention, and safety research, making it easier for external stakeholders to assess and compare company behavior. By also analyzing companies' model eval reports and safety claims, the project helps ensure companies cannot make misleading claims about their safety practices without scrutiny.
Grants Received
from Survival and Flourishing Fund
Projects
No linked projects.
People
No linked people.
Discussion
Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.
Details
- Last Updated
- Apr 2, 2026, 10:11 PM UTC
- Created
- Mar 18, 2026, 11:18 PM UTC
