Google-backed Anthropic builds ‘Constitution’ for its AI chatbot – Times of India

AI-powered chatbots can be unpredictable and sometimes generate harmful or illegal content, leading to ongoing efforts to prevent such behavior. And there’s not much their makers can do, but a company, anthropic, it seems how they can better control their chatbot. All it needs is a “constitution”.
Founded by ex-OpenAI engineers, backed by AI startup Anthropic Google, has set the rules for its “constitutional AI” training method. This approach instills specific values ​​in your chatbot, Claude, to reduce concerns about transparency, security and decision making in AI systems. Unlike other methods, this approach does not require human feedback to evaluate responses.
In a blog post, Anthropic said that AI models will inevitably have value systems, whether intentional or not. To make up for any deficiencies, Constitutional AI Uses AI feedback to evaluate its outputs.
Anthropic’s AI Constitution includes 58 principles inspired by sources such as the Universal Declaration of Human Rights. United Nations, AppleTerms of Service, Google’s terms, and Anthropic’s own research. These principles are lofty and aim to promote fairness and dignity for all.
That’s the whole gist of the constitution – AIs must follow guidelines to avoid stereotyping, discriminatory language, and giving medical, financial, or legal advice. It should provide appropriate feedback for children and avoid offending non-Western audiences. It should prioritize less existentially risky responses and avoid being didactic.
AI-powered chatbots like GPT-4 and forage can generate text in vivid detail, but they also have significant drawbacks. These generative AI models are often trained on untrustworthy internet sources such as social media, making them prone to bias. Furthermore, it may generate answers that are not based on actual knowledge and are purely hypothetical.
Anthropic’s prescriptive AI aims to tackle these issues by providing a set of guiding principles for the system to make informed decisions about the text it produces. These principles encourage models to adopt behaviors that are “non-toxic” and “helpful” in nature.
When Anthropic trains a text-generation model, they follow a set of guidelines in two steps. First, the model learns to evaluate and improve its responses by observing these guidelines and relevant examples. Then, they use the feedback generated by the first model along with guidelines to train the final model.
The startup believes its training method is better than chatgptBecause human response is not scalable and requires a lot of time and resources. OpenAI It has faced criticism for underpaying contract workers to filter out toxic data. Unlike OpenAI, Constitutive AI is transparent and easily verifiable, says Anthropic.
Anthropic AI wants to create an advanced algorithm for self-learning. It can be used to create virtual assistants that can do things like answer emails, research and create art and books. There are already models such as the GPT-4 and LaMDA that use this technology.