Anthropic releases Claude 2.1 with 200k context window

  • Anthropic released Claude 2.1, the latest version of the model that powers its AI chatbot.
  • Claude 2.1 has a context window of 200,000 tokens, equivalent to 500 pages of text.
  • Anthropic has improved the model’s accuracy and reduced instances of hallucinating by half.

Claude.ai is Anthropic’s answer to ChatGPT and the AI model that powers it just got a major upgrade.

Claude 2.1 delivers a significant improvement in performance compared to Claude 2.0. Probably the biggest upgrade is the doubling of its context window, but there are some other impressive features too.

Anthropic’s blog post goes into a lot more detail but here is a simplified summary of the upgrades.

Context window

The context window of a model determines how many tokens it can keep in its memory. Claude 2.1 now has a 200,000 token context window, double that of Claude 2.0 and the largest in the industry. For comparison, GPT-4 Turbo has a 128k context window.

This means you could drop around 150,000 words, or 500 pages into a chat with Claude and ask questions related to the material. Well, in theory.

Greg Kamradt did a stress test of Claude 2.1’s recall capability and it struggled with accuracy as the context grew longer and when the fact to recall was somewhere in the middle of the document.

It’s still pretty impressive despite some recall accuracy issues.

More accuracy and honesty

Claude 2.1 makes fewer mistakes and will lie to you less often. It’s 30% more likely to give a correct answer than before.

Like other AI models, it will still hallucinate but it does so around half as much as Claude 2.0 does.

Claude 2.1 got an upgrade in its intellectual humility too. It is also almost twice as likely to decline to answer a question when it doesn’t know the answer rather than make something up.

Claude 2.1 accuracy when answering hard questions. Source: Anthropic

API tool use

Claude can now interact with a user’s databases, search over web resources for an answer, or interact with other tools via APIs.

Anthropic says a user can now define a set of tools, ask a question, and then Claude will decide which tools to use to answer the question.

Claude already integrates with Zapier so this added ability to translate natural language into API or function calls could be huge. Could we see Anthropic’s version of OpenAI’s GPTs soon?

The tool use feature is in beta so we’ll have to wait to see what it’s capable of.

System prompts

This feature allows an API call to give Claude context and instructions on how to respond before a human prompt is entered.

This means you can have Claude assume a certain character or voice and tell it things it should or shouldn’t do when interacting with the user. Subsequent interactions with Claude will then have the chat responses stay in character for longer.

If you want to experience Claude 2.1 you can head over to the chat interface at claude.ai. The 200k context window is only available to Claude Pro users though.

© 2023 Intelliquence Ltd. All Rights Reserved.

Privacy Policy | Terms and Conditions

×
 
 

FREE PDF EXCLUSIVE
Stay Ahead with DailyAI


 

Sign up for our weekly newsletter and receive exclusive access to DailyAI's Latest eBook: 'Mastering AI Tools: Your 2024 Guide to Enhanced Productivity'.



 
 

*By subscribing to our newsletter you accept our Privacy Policy and our Terms and Conditions