AI gains “values” with Anthropic’s new Constitutional AI chatbot approach

Anthropic's Constitutional AI logo on a glowing orange background.

Enlarge / Anthropic’s Constitutional AI logo on a glowing orange background. (credit: Anthropic / Benj Edwards)

On Tuesday, AI startup Anthropic detailed the specific principles of its “Constitutional AI” training approach that provides its Claude chatbot with explicit “values.” It aims to address concerns about transparency, safety, and decision-making in AI systems without relying on human feedback to rate responses.

Claude is an AI chatbot similar to OpenAI’s ChatGPT that Anthropic released in March.

“We’ve trained language models to be better at responding to adversarial questions, without becoming obtuse and saying very little,” Anthropic wrote in a tweet announcing the paper. “We do this by conditioning them with a simple set of behavioral principles via a technique called Constitutional AI.”

Read 18 remaining paragraphs | Comments

Source

Leave a Reply

Your email address will not be published. Required fields are marked *