Tag

Constitutional Ai

All articles tagged with #constitutional ai

ai2 years ago

Anthropic's AI Constitution: A Moral Framework for Safe and Ethical AI.

AI startup Anthropic has developed a "Constitutional AI" training approach that provides its Claude chatbot with explicit "values" to address concerns about transparency, safety, and decision-making in AI systems without relying on human feedback to rate responses. The principles of the constitution include the United Nations Declaration of Human Rights, portions of Apple's terms of service, several trust and safety "best practices," and Anthropic's AI research lab principles. The model critiques and revises its responses using the set of principles, and reinforcement learning relies on AI-generated feedback to select the more "harmless" output.

ai-ethics2 years ago

Anthropic proposes AI constitution to ensure ethical development.

Anthropic, an AI startup founded by former OpenAI employees, is focusing on "constitutional AI" to make AI systems safe. The company has created a set of principles, inspired by the UN's Universal Declaration of Human Rights, Apple's terms of service, and its own research, to train AI systems to follow certain sets of rules. The principles include guidance to prevent users from anthropomorphizing chatbots, telling the system not to present itself as a human, and to consider non-Western perspectives. The company's intention is to prove the general efficacy of its method and start a public discussion about how AI systems should be trained and what principles they should follow.

ai2 years ago

Anthropic's Claude: The New Competitor to OpenAI's ChatGPT.

Anthropic, a startup co-founded by ex-OpenAI employees, has launched Claude, an AI chatbot that can perform tasks such as searching, summarizing, writing, and coding. Claude is similar to OpenAI's ChatGPT but is "much less likely to produce harmful outputs," "easier to converse with," and "more steerable." Claude was trained using a technique called "constitutional AI," which aligns AI systems with human intentions. However, Claude has limitations, including hallucinations and safety feature bypassing. Anthropic plans to let developers customize Claude's constitutional principles and sees its core users as startups and larger enterprises.