This website requires javascript to properly function. Consider activating javascript to get access to all site functionality.
AI ALIGNMENT FORUM
Tags
AF
Login
Anthropic (org)
•
Applied to
Anthropic teams up with Palantir and AWS to sell AI to defense customers
by
Matrice Jacobine
13d
ago
•
Applied to
Anthropic: Three Sketches of ASL-4 Safety Case Components
by
Vanessa Kosoy
15d
ago
•
Applied to
Anthropic - The case for targeted regulation
by
anaguma
18d
ago
•
Applied to
Anthropic rewrote its RSP
by
Ruben Bloom
1mo
ago
•
Applied to
Anthropic's updated Responsible Scaling Policy
by
Zac Hatfield-Dodds
1mo
ago
•
Applied to
Dario Amodei — Machines of Loving Grace
by
Matrice Jacobine
1mo
ago
•
Applied to
John Schulman leaves OpenAI for Anthropic
by
Raymond Arnold
4mo
ago
•
Applied to
Can We Predict Persuasiveness Better Than Anthropic?
by
Lennart Finke
4mo
ago
•
Applied to
On Claude 3.5 Sonnet
by
Tobias D.
5mo
ago
•
Applied to
Anthropic's Certificate of Incorporation
by
Ben Millwood
5mo
ago
•
Applied to
Maybe Anthropic's Long-Term Benefit Trust is powerless
by
Jérémy Perret
6mo
ago
•
Applied to
Quick Thoughts on Scaling Monosemanticity
by
Joel Burget
6mo
ago
•
Applied to
Cicadas, Anthropic, and the bilateral alignment problem
by
kromem
6mo
ago
•
Applied to
EIS XIII: Reflections on Anthropic’s SAE Research Circa May 2024
by
Stephen Casper
6mo
ago
•
Applied to
Anthropic: Reflections on our Responsible Scaling Policy
by
Zac Hatfield-Dodds
6mo
ago
•
Applied to
Anthropic AI made the right call
by
bhauth
7mo
ago
•
Applied to
OMMC Announces RIP
by
Adam Scholl
8mo
ago
•
Applied to
Vaniver's thoughts on Anthropic's RSP
by
Gunnar Zarncke
10mo
ago
•
Applied to
Introducing Alignment Stress-Testing at Anthropic
by
Gunnar Zarncke
10mo
ago