Blog
Claude Sonnet 3.7 (often) knows when it’s in alignment evaluations
We evaluate whether Claude Sonnet 3.7 and other frontier models know that they are being evaluated.
Forecasting Frontier Language Model Agent Capabilities
We present a new forecasting technique to predict frontier LM agent capabilities ahead of time.
Demo example - Scheming reasoning evaluations
A brief demonstration video show-casing a representative example of our in-context scheming evaluations.
Apollo 18-month update
Apollo Research is now 18 months old. You can read our latest update here.
Apollo is adopting Inspect
Apollo is adopting Inspect as its evals framework. We will contribute features and potentially example agent evals to Inspect and look forward to work with the Inspect community
The Evals Gap
The quality and quantity of evals required to make rigorous safety statements could outpace available evals. We explain “the evals gap” and what would be required to close it.
An Opinionated Evals Reading List
A long reading list of evals papers with recommendations and comments by the evals team.
The first year of Apollo Research
A summary of what we have achieved in our first year and what we plan to do in the future.
Black-Box Access is Insufficient for Rigorous AI Audits
We were delighted to collaborate on the paper “Black-Box Access is Insufficient for Rigorous AI Audits.”
We need a Science of Evals
In this post, we argue that if AI model evaluations (evals) want to have meaningful real-world impact, we need a “Science of Evals”, i.e. the field needs rigorous scientific processes that provide more confidence in evals methodology and results.