These are not the words you want to hear when it comes to human extinction, but I was hearing them: โ€œThings are moving uncomfortably fast.โ€ I was sitting in a conference room with Sam Bowman, a safety researcher at Anthropic. Worth $183 billion at the latest estimate, the AI firm has every incentive to speed things up, ship more products, and develop more advanced chatbots to stay competitive with the likes of OpenAI, Google, and the industryโ€™s other giants. But Anthropic is at odds with itselfโ€”thinking deeply, even anxiously, about seemingly every decision.

Anthropic has positioned itself as the AI industryโ€™s superego: the firm that speaks with the most authority about the big questions surrounding the technology, while rival companies develop advertisements and affiliate shopping links (a difference that Anthropicโ€™s CEO, Dario Amodei, was eager to call out during an interview in Davos last week). On Monday, Amodei published a lengthy essay, โ€œThe Adolescence of Technology,โ€ about the โ€œcivilizational concernsโ€ posed by what he calls โ€œpowerful AIโ€โ€”the very technology his firm is developing. The essay has a particular focus on democracy, national security, and the economy. โ€œGiven the horror weโ€™re seeing in Minnesota, its emphasis on the importance of preserving democratic values and rights at home is particularly relevant,โ€ Amodei posted on X, making him one of very few tech leaders to make a public statement against the Trump administrationโ€™s recent actions.

This rhetoric, of course, serves as good brandingโ€”a way for Anthropic to stand out in a competitive industry. But having spent a long time following the company and, recently, speaking with many of its employees and executives, including Amodei, I can say that Anthropic is at least consistent. It messages about the ethical issues surrounding AI constantly, and it appears unusually focused on user safety. Bowmanโ€™s job, for example, is to vet Anthropicโ€™s products before theyโ€™re released into the world, making sure that they will not spew, say, white-supremacist talking points; push users into delusional crises; or generate nonconsensual porn.

So far, the effort seems to be working: Unlike other popular chatbots, including OpenAIโ€™s ChatGPT and Elon Muskโ€™s Grok, Anthropicโ€™s bot, Claude, has not had any major public blowups despite being as advanced as, and by some measures more advanced than, the rest of the field. (That may be in part because its chatbot does not generate images and has a smaller user base than some rival products.) But although Anthropic has so far dodged the various scandals that have plagued other large language models, the company has not inspired much faith that such problems will be avoided forever. When I met Bowman last summer, the company had recently divulged that, in experimental settings, versions of Claude had demonstrated the ability to blackmail users and assist them when they ask about making bioweapons.

๐Ÿ“ฐ

Continue Reading on The Atlantic

This preview shows approximately 15% of the article. Read the full story on the publisher's website to support quality journalism.

Read Full Article โ†’