AI Safety at the Frontier is a free Substack newsletter run by Johannes Gasteiger, a Member of Technical Staff in Anthropic's Alignment Science team. Published monthly since July 2024, it reviews 50+ papers each month to select and summarize the most interesting AI safety research on frontier models. Topics covered include reward hacking, AI scheming, model evaluation, alignment, and AI control. Posts are also cross-posted to LessWrong.
Funding Details
- Annual Budget
- -
- Monthly Burn Rate
- -
- Current Runway
- -
- Funding Goal
- -
- Funding Raised to Date
- -
- Fiscal Sponsor
- -
Theory of Change
By curating and summarizing frontier AI safety research in an accessible monthly format, the newsletter helps researchers, engineers, and policymakers stay informed about the latest technical developments in AI safety, potentially accelerating the spread of important ideas and findings across the field.
Grants Received
No grants recorded.
Projects
No linked projects.
People
No linked people.
Discussion
Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.
Details
- Last Updated
- Apr 2, 2026, 9:50 PM UTC
- Created
- Mar 19, 2026, 10:31 PM UTC