This website requires javascript to properly function. Consider activating javascript to get access to all site functionality.
AI ALIGNMENT FORUM
Tags
AF
Login
Anthropic (org)
•
Applied to
Has Anthropic checked if Claude fakes alignment for intended values too?
by
Maloew
12h
ago
•
Applied to
Introducing the Anthropic Fellows Program
by
Miranda Zhang
21d
ago
•
Applied to
Anthropic teams up with Palantir and AWS to sell AI to defense customers
by
Matrice Jacobine
1mo
ago
•
Applied to
Anthropic: Three Sketches of ASL-4 Safety Case Components
by
Vanessa Kosoy
1mo
ago
•
Applied to
Anthropic - The case for targeted regulation
by
anaguma
2mo
ago
•
Applied to
Anthropic rewrote its RSP
by
Ruben Bloom
2mo
ago
•
Applied to
Anthropic's updated Responsible Scaling Policy
by
Zac Hatfield-Dodds
2mo
ago
•
Applied to
Dario Amodei — Machines of Loving Grace
by
Matrice Jacobine
2mo
ago
•
Applied to
John Schulman leaves OpenAI for Anthropic
by
Raymond Arnold
5mo
ago
•
Applied to
Can We Predict Persuasiveness Better Than Anthropic?
by
Lennart Finke
5mo
ago
•
Applied to
On Claude 3.5 Sonnet
by
Tobias D.
6mo
ago
•
Applied to
Anthropic's Certificate of Incorporation
by
Ben Millwood
6mo
ago
•
Applied to
Maybe Anthropic's Long-Term Benefit Trust is powerless
by
Jérémy Perret
7mo
ago
•
Applied to
Quick Thoughts on Scaling Monosemanticity
by
Joel Burget
7mo
ago
•
Applied to
Cicadas, Anthropic, and the bilateral alignment problem
by
kromem
7mo
ago
•
Applied to
EIS XIII: Reflections on Anthropic’s SAE Research Circa May 2024
by
Stephen Casper
7mo
ago
•
Applied to
Anthropic: Reflections on our Responsible Scaling Policy
by
Zac Hatfield-Dodds
7mo
ago
•
Applied to
Anthropic AI made the right call
by
bhauth
8mo
ago
•
Applied to
OMMC Announces RIP
by
Adam Scholl
9mo
ago