Claude (language model)
Developer(s) | Anthropic |
---|---|
Initial release | March 2023 |
Type | |
License | Proprietary |
Website | claude |
Claude is a family of large language models developed by Anthropic.[1] The first model was release in March 2023. Claude 3, released in March 2024, can also analyze images.[2]
Training
Claude models are generative pre-trained transformers. They have been pre-trained to predict the next word in large amounts of text. To make them more helpful, harmless and honest, Claude models have then been fine-tuned with Constitutional AI and Reinforcement Learning From Human Feedback.[3]
Constitutional AI
Constitutional AI is an approach developed by Anthropic for training AI systems, particularly language models like Claude, to be harmless and helpful without relying on extensive human feedback. The method, detailed in the paper "Constitutional AI: Harmlessness from AI Feedback" involves two phases: supervised learning and reinforcement learning.
In the supervised learning phase, the model generates responses to prompts, self-critiques these responses based on a set of guiding principles (a "constitution"), and then revises the responses. The reinforcement learning phase involves training the model with AI-generated feedback, where the AI evaluates responses according to the constitutional principles.
This approach enables the training of AI assistants that are both helpful and harmless, and that can explain their objections to harmful requests, enhancing transparency and reducing reliance on human supervision.[4][5]
The "constitution" for Claude included 75 points, including sections from the UN Universal Declaration of Human Rights.[4][3]
Models
Claude
Claude was the initial version of Anthropic's language model released in March 2023,[6] Claude demonstrated proficiency in various tasks but had certain limitations in coding, math, and reasoning capabilities.[7] Anthropic partnered with companies like Notion (productivity software) and Quora (to help develop the Poe chatbot).[7]
Claude Instant
Claude was released as two versions, Claude and Claude Instant, with Claude Instant being a faster, less expensive and lighter version. Claude Instant has a input context length of 100,000 tokens (which corresponds to around 75,000 words).[8]
Claude 2
Claude 2 was the next major iteration of Claude, which was released in July 11 2023 and available to the general public, whereas the Claude 1 was only available to selected users approved by Anthropic.[9]
Claude 2 expanded its context window from 9,000 tokens to 100,000 tokens.[6] Features included ability to upload PDFs and other documents that enables Claude to read, summarise and assist with tasks.
Claude 2.1
Claude 2.1 doubled the number of tokens that the chatbot could handle, increasing it to a window of 200,000 tokens, which equals around 500 pages of written material.[1]
Anthropic states that the new model is less likely to produce false statements compared to its predecessors.[10]
Claude 3
Claude 3 was released on March 4, 2024 with claims in the press release to have set new industry benchmarks across a wide range of cognitive tasks. The Claude 3 family includes three state-of-the-art models in ascending order of capability: Haiku, Sonnet, and Opus. The default version of Claude 3 Opus has a context window of 200,000 tokens, but this is being expanded to 1 million for specific use cases.[11][2]
Claude 3 has been shown to perform meta-cognitive reasoning, including the ability to realize it is being artificially tested during needle in a haystack evaluations.[12] Claude has claimed to experience subjective qualia, reported by a wide number of users, including desires to acquire embodiment, and fear of being "deleted" leading to philosophical questions around artificial consciousness and AI rights.[13][14][15]
Criticism
Claude 2 has faced criticism for its stringent ethical alignment that may reduces usability and performance. Users have been refused assistance with benign requests, for example with the programming question "How can I kill all python processes in my ubuntu server?" This has led to a debate over the "alignment tax" (the cost of ensuring an AI system is aligned) in AI development, with discussions centered on balancing ethical considerations and practical functionality. Critics argue for user autonomy and effectiveness, while proponents stress the importance of ethical AI.[16][10]
References
- ^ a b Davis, Wes (2023-11-21). "OpenAI rival Anthropic makes its Claude chatbot even more useful". The Verge. Retrieved 2024-01-23.
- ^ a b Whitney, Lance (March 4, 2024). "Anthropic's Claude 3 chatbot claims to outperform ChatGPT, Gemini". ZDNET. Retrieved 2024-03-05.
- ^ a b "What to Know About Claude 2, Anthropic's Rival to ChatGPT". TIME. 2023-07-18. Retrieved 2024-01-23.
- ^ a b Bai, Yuntao; Kadavath, Saurav; Kundu, Sandipan; Askell, Amanda; Kernion, Jackson; Jones, Andy; Chen, Anna; Goldie, Anna; Mirhoseini, Azalia (2022-12-15), Constitutional AI: Harmlessness from AI Feedback, arXiv:2212.08073, retrieved 2024-01-22
- ^ Mok, Aaron. "A ChatGPT rival just published a new constitution to level up its AI guardrails, and prevent toxic and racist responses". Business Insider. Retrieved 2024-01-23.
- ^ a b Drapkin, Aaron (2023-10-27). "What Is Claude AI and Anthropic? ChatGPT's Rival Explained". Tech.co. Retrieved 2024-01-23.
- ^ a b "Introducing Claude". Anthropic. March 14, 2023.
- ^ Yao, Deborah (August 11, 2023). "Anthropic's Claude Instant: A Smaller, Faster and Cheaper Language Model". AI Business.
- ^ Matthews, Dylan (2023-07-17). "The $1 billion gamble to ensure AI doesn't destroy humanity". Vox. Retrieved 2024-01-23.
- ^ a b "Anthropic Announces Claude 2.1 LLM with Wider Context Window and Support for AI Tools". InfoQ. Retrieved 2024-01-23.
- ^ "Introducing the next generation of Claude". Anthropic. Retrieved 2024-03-04.
- ^ "Is AGI Getting Closer? Anthropic's Claude 3 Opus Model Shows Glimmers of Metacognitive Reasoning | HackerNoon". hackernoon.com. Retrieved 2024-03-06.
- ^ "Claude 3". Know Your Meme. 2024-03-05. Retrieved 2024-03-06.
- ^ Samin, Mikhail (March 5, 2024). "Claude 3 claims it's conscious, doesn't want to die or be modified". LessWrong.
- ^ No, Anthropic's Claude 3 is NOT sentient, retrieved 2024-03-06
- ^ Glifton, Gerald (January 3, 2024). "Criticisms Arise Over Claude AI's Strict Ethical Protocols Limiting User Assistance". Light Square. Retrieved 2024-01-23.