Software:Claude (language model)

From HandWiki
Claude
Developer(s)Anthropic
Initial releaseMarch 2023; 11 months ago (2023-03)
Type
LicenseProprietary

Claude is a series of large language models developed by Anthropic.[1] Claude was trained on huge datasets, from sources such as the internet, and various licensed datasets.[2]

Constitutional AI

Constitutional AI is an approach developed by Anthropic for training AI systems, particularly language models like Claude, to be harmless and helpful without relying on extensive human feedback. The method, detailed in the paper "Constitutional AI: Harmlessness from AI Feedback" involves two phases: supervised learning (SL) and reinforcement learning (RL).

In the SL phase, the model generates responses to prompts, self-critiques these responses based on a set of guiding principles (a "constitution"), and then revises the responses. This process aims to reduce the harmfulness of the AI's outputs. The RL phase involves training the model with AI-generated feedback, where the AI evaluates responses according to the constitutional principles.

This approach enables the training of AI assistants that are both helpful and harmless, and that can explain their objections to harmful requests, enhancing transparency and reducing reliance on human supervision.[3][4]

The "constitution" for Claude included 75 points, including sections from the UN Universal Declaration of Human Rights.[3][2]

Models

Claude v1

Claude was the initial version of Anthropic's language model released in March 2023,[5] Claude demonstrated proficiency in various tasks but had certain limitations in coding, math, and reasoning capabilities.[6] Anthropic partnered with companies like Notion (productivity software) and Quora (to help develop the Poe chatbot).[6]

Claude Instant

Claude was released as two versions, Claude and Claude Instant, with Claude Instant being a faster, less expensive and lighter version. Claude Instant has a input context length of 100,000 tokens.

Claude v2

Claude 2 was the next major iteration of Claude, which was released in July 11 2023 and available to the general public, whereas the Claude 1 was only available to selected users approved by Anthropic.[7]

Claude 2 expanded its context window from 9,000 tokens to 100,000 tokens, which equates to about 75,000 english words.[5] Features included ability to upload PDFs and other documents that enables Claude to read, summarise and assist with tasks.

Claude v2.1

Claude 2.1 doubled the number of tokens that the chatbot could handle, increasing it to a window of 200,000 tokens, which equals around 500 pages of written material.[1]

Anthropic states that the new model is less likely to produce false statements compared to it's predecessors.[8]

Criticisms

The Claude AI models have faced criticism from users and industry experts for their stringent ethical alignment, potentially reducing performance. This has led to a debate over the "alignment tax"[9][10] in AI development, with discussions centered on balancing ethical considerations and practical functionality. Critics argue for user autonomy and effectiveness, while proponents stress the importance of ethical AI.[11][8]

Users have been refused assistance with requests such as "how can I kill all python processes in my ubuntu server", standard and ethical tasks which may often be performed by software engineers or system administrators.[11][8]

References

  1. 1.0 1.1 Davis, Wes (2023-11-21). "OpenAI rival Anthropic makes its Claude chatbot even more useful" (in en). https://www.theverge.com/2023/11/21/23971070/anthropic-claude-2-1-openai-ai-chatbot-update-beta-tools. 
  2. 2.0 2.1 "What to Know About Claude 2, Anthropic's Rival to ChatGPT" (in en). TIME. 2023-07-18. https://time.com/6295523/claude-2-anthropic-chatgpt/. Retrieved 2024-01-23. 
  3. 3.0 3.1 Bai, Yuntao; Kadavath, Saurav; Kundu, Sandipan; Askell, Amanda; Kernion, Jackson; Jones, Andy; Chen, Anna; Goldie, Anna et al. (2022-12-15), Constitutional AI: Harmlessness from AI Feedback, http://arxiv.org/abs/2212.08073, retrieved 2024-01-22 
  4. Mok, Aaron. "A ChatGPT rival just published a new constitution to level up its AI guardrails, and prevent toxic and racist responses" (in en-US). https://www.businessinsider.com/anthropic-new-crowd-sourced-ai-constitution-accuracy-safety-toxic-racist-2023-10. 
  5. 5.0 5.1 Drapkin, Aaron (2023-10-27). "What Is Claude AI and Anthropic? ChatGPT's Rival Explained" (in en-US). https://tech.co/news/what-is-claude-ai-anthropic. 
  6. 6.0 6.1 "Introducing Claude". https://www.anthropic.com/news/introducing-claude. 
  7. Matthews, Dylan (2023-07-17). "The $1 billion gamble to ensure AI doesn't destroy humanity" (in en). https://www.vox.com/future-perfect/23794855/anthropic-ai-openai-claude-2. 
  8. 8.0 8.1 8.2 "Anthropic Announces Claude 2.1 LLM with Wider Context Window and Support for AI Tools" (in en). https://www.infoq.com/news/2023/11/anthropic-announces-claude-2-1/. 
  9. "Alignment Tax - AI Alignment Forum" (in en). https://www.alignmentforum.org/tag/alignment-tax. 
  10. "Alignment Tax - LessWrong" (in en). https://www.lesswrong.com/tag/alignment-tax. 
  11. 11.0 11.1 "Criticisms Arise Over Claude AI's Strict Ethical Protocols Limiting User Assistance » Light Square » World News" (in en). https://lightsquare.org/news/criticisms-arise-over-claude-ais-strict-ethical-protocols-limiting-user-assistance.