AI Safety Takes is a Substack newsletter written by Daniel Paleka, a PhD student at ETH Zurich researching security and failure modes of artificial intelligence. Launched in November 2022, the newsletter covers AI safety research topics including technical alignment, interpretability, RLHF, adversarial robustness, and related developments. It began as a curated monthly digest of AI safety papers and evolved into broader commentary whenever Paleka has something substantive to say about AI.
Funding Details
- Annual Budget
- -
- Monthly Burn Rate
- -
- Current Runway
- -
- Funding Goal
- -
- Funding Raised to Date
- -
- Fiscal Sponsor
- -
Theory of Change
By synthesizing and commenting on the latest AI safety research, the newsletter aims to help practitioners and researchers stay informed about important developments in alignment, interpretability, and AI robustness. Paleka's implicit theory of change is that better-informed researchers and decision-makers in AI will contribute to safer AI development outcomes.
Grants Received
No grants recorded.
Projects
No linked projects.
People
No linked people.
Discussion
Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.
Details
- Last Updated
- Apr 2, 2026, 9:51 PM UTC
- Created
- Mar 19, 2026, 10:30 PM UTC
