Jeremy Gillen
Bio
Jeremy Gillen is an independent AI alignment researcher based in Berkeley, California, working primarily on agent foundations and the ontology identification problem. He holds an undergraduate degree in Computer Science and Neuroscience with a thesis on statistical learning theory. He participated in the SERI MATS (ML Alignment Theory Scholars Program) cohort 2 under mentor John Wentworth, where he co-authored "Finding Goals in the World Model" — a proposal for aligning model-based RL systems by identifying human values in a world model and using inverse reinforcement learning to guide the policy. Following MATS, he received a Long-Term Future Fund grant to continue independent research on alignment problems in model-based RL. He subsequently joined Vivek Chan's team at MIRI (Machine Intelligence Research Institute) before returning to independent research. His current work focuses on the ontology identification problem and related natural abstractions research, with recent co-authored work on condensation and natural latents. He is an active contributor to LessWrong and the AI Alignment Forum, and has participated in public debates on AI corrigibility.
Links
- Personal Website
- -
- Twitter / X
- LessWrong
- jeremy-gillen
Grants
from Long-Term Future Fund
Discussion
Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.
Details
- Last Updated
- Mar 22, 2026, 4:53 PM UTC
- Created
- Mar 20, 2026, 2:52 AM UTC