What the people building AI are actually worried about

John O'Neill

Amodei makes the case for light regulation (warning, also, of carefully considering unintended consequences of any legislation) and describes Constitutional AI as one of Anthropic’s core innovations – aspects of which he says have since been adopted by other AI companies.

Dario Amodei, CEO and co-founder of Anthropic
Anthropic CEO Dario Amodei

Amodei describes the approach: “The idea that AI training (specifically the ‘post-training’ stage, in which we steer how the model behaves) can involve a central document of values and principles that the model reads and keeps in mind when completing every training task — to produce a model that almost always follows this constitution.”

Interestingly, he says that training Claude at the level of identity, character, values, and personality – rather than giving it specific instructions or priorities without explaining the reasons behind them – has proven more likely to lead to a “coherent, wholesome, and balanced psychology”.

Sounds almost human?!

Amodei insists that we can navigate the risks of AI, but only if we take them very seriously and do that now. For anyone who wants to think deeply about where AI might take us, and what’s at stake if we get it wrong, this is essential reading, but make sure you set aside plenty of time – the essay is more than 18,000 words in length, akin to a short book – and perfect reading for a long flight.