Understand Real Failure Patterns
python main/field_level_findings.py
Produces stratified outcomes by task, model, claim family, lane, and provider, plus top failure gates and policy sensitivity outputs.
AutoMechInterp helps researchers and developers evaluate interpretability claims with a consistent, reproducible verification contract.
When do mechanistic-interpretability claims fail under principled controls, robustness checks, and preregistered statistical policy?
AutoMechInterp is the instrument that lets this question be answered repeatedly across tasks, models, and lanes.
python main/field_level_findings.py
Produces stratified outcomes by task, model, claim family, lane, and provider, plus top failure gates and policy sensitivity outputs.
python main/stress_test_ablation.py --bundle-dir main/output/real_multi_task/ioi_v0_gpt2-small
python main/stress_test_agnostic.py --bundle-dir main/output/real_multi_task/ioi_v0_gpt2-small
python main/stress_test_red_team.py --bundle-dir main/output/real_multi_task/ioi_v0_gpt2-small
Runs complementary stress tests to show how robust your claim-evaluation setup is under different failure pressures.
Start with the Quickstart, then move to Community Submissions and Standards for interoperability details.