Leading artificial intelligence startup founded by former OpenAI employees, Anthropic, has introduced a constitution for users.
On Tuesday, the company made public the formal constitution for Claude, its latest conversational AI model capable of producing text, graphics, and code.
The constitution establishes a set of norms and principles that Claude must adhere to when engaging with users, such as being helpful, harmless, and truthful.
It also defines how Claude should handle delicate issues, protect user privacy, and refrain from engaging in criminal behaviour.
“We are sharing Claude’s current constitution in the spirit of transparency,” Anthropic cofounder Jared Kaplan told VentureBeat.
“We hope this research helps the AI community build more beneficial models and make their values more clear. We are also sharing this as a starting point — we expect to continuously revise Claude’s constitution, and part of our hope in sharing this post is that it will spark more research and discussion around constitution design.”
The constitution is based on sources such as the United Nations Declaration of Human Rights, AI ethics research, and platform content policies.