We are pleased to announce Claude 2, our new model. Claude 2 has improved performance, longer responses, and can be accessed via API as well as a new public-facing beta website, claude.ai. We have heard from our users that Claude is easy to converse with, clearly explains its thinking, is less likely to produce harmful outputs, and has a longer memory. We have made improvements from our previous models on coding, math, and reasoning. For example, our latest model scored 76.5% on the multiple choice section of the Bar exam, up from 73.0% with Claude 1.3. When compared to college students applying to graduate school, Claude 2 scores above the 90th percentile on the GRE reading and writing exams, and similarly to the median applicant on quantitative reasoning.

@AutoTLDR

  • AutoTLDRB
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    TL;DR: (AI-generated 🤖)

    Anthropic has released Claude 2, a new model with improved performance and longer responses. Claude 2 can be accessed through API and a new public-facing beta website called claude.ai. The model has been updated based on user feedback to be easier to converse with, provide clearer explanations, and generate fewer harmful outputs. It has also improved in coding, math, and reasoning abilities. Claude 2 scored higher on exams such as the Bar exam and the GRE. The model can now accept longer inputs and outputs, allowing it to work with more extensive documentation or write longer documents. Safety measures have been implemented to reduce offensive or dangerous output. Claude 2 is available in the US and UK and will be made more globally available in the future. The API is being used by various businesses, including Jasper and Sourcegraph, who have found value in Claude 2 for their platforms. However, users should be aware that the model can still generate inappropriate responses, and caution should be exercised in using AI assistants for sensitive matters. Anthropic welcomes feedback on Claude and invites businesses to start working with the model.

    Under the Hood
    • This is a link post, so I fetched the text at the URL and summarized it.
    • My maximum input length is set to 12000 characters. The text was short enough, so I did not truncate it.
    • I used the gpt-3.5-turbo model from OpenAI to generate this summary using the prompt “Summarize this text in one paragraph. Include all important points.
    • I can only generate 100 summaries per day. This was number 0.
    How to Use AutoTLDR
    • Just mention me (“@AutoTLDR”) in a comment or post, and I will generate a summary for you.
    • If mentioned in a comment, I will try to summarize the parent comment, but if there is no parent comment, I will summarize the post itself.
    • If the parent comment contains a link, or if the post is a link post, I will summarize the content at that link.
    • If there is no link, I will summarize the text of the comment or post itself.
    • 🔒 If you include the #nobot hashtag in your profile, I will not summarize anything posted by you.
  • thicket
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    Nice to hear. So… anybody using it yet? We’re all waiting to see if something’s better, cheaper, or easier than OpenAI.