Anthropic has launched its new AI model, upgraded Claude 3.5 Sonnet.
Anthropic announced the upgrade to Claude 3.5 Sonnet on October 22, 2024. The announcement came just a few months after Claude 3.5 Sonnet was initially released on June 20, 2024.
With the initial release of Claude 3.5 Sonnet (in June 2024), Anthropic claimed that it outperformed its peers, such as OpenAI’s GPT-4o, Google’s Gemini-1.5 Pro, Meta’s Llama-400b, and even the company’s proprietary models — Claude 3 Haiku and Claude 3 Opus.
It now notes that its upgraded Claude 3.5 Sonnet (October 2024) outperforms its predecessor and delivers across-the-board improvements that show wide-ranging improvements on industry benchmarks, with particularly strong gains in agentic coding and tool use tasks.
So, we put the Originality.ai AI detector to the test to determine its accuracy in detecting AI-generated text created by this upgraded Claude 3.5 Sonnet model.
To establish the AI Checker’s accuracy, this brief study generated 1000 Claude 3.5 Sonnet text results and then ran them through the Originality.ai AI checker.
Yes — Claude 3.5 Sonnet text is yet detectable with 99.0% accuracy for the Originality.ai Turbo 3.0.1.
Try our AI Detector here.
To evaluate the detectability of upgraded Claude 3.5 Sonnet, we prepared a dataset of 1000 Claude 3.5 Sonnet generated text samples.
For AI-text generation, we used upgraded Claude 3.5 Sonnet based on three approaches given below:
To evaluate the efficacy, we used our Open Source AI Detection Efficacy tool:
Originality.ai has three models — Model 3.0.1 Turbo, 1.0.0 Lite, and Multi Language for the purpose of AI text detection.
Learn more about which AI detection model is best for you and your use case.
The open-source testing tool returns a variety of metrics for each detector tested, each of which reports on a different aspect of that detector’s performance, including:
If you'd like a detailed discussion of these metrics, what they mean, how they're calculated, and why we chose them, check out our blog post on AI detector evaluation. For a succinct snapshot, the confusion matrix is an excellent representation of a model's performance.
Below is an evaluation of both the models on the above dataset.
For this small test to reflect the Originality.ai detector’s ability to identify Claude 3.5 Sonnet content, we looked at the True Positive Rate or the percentage of the time the model correctly identified AI text as AI out of a 1000 sample Claude 3.5 Sonnet content.
Model 3.0.1 Turbo:
Our study confirms that the content generated by upgraded Claude 3.5 AI-generated text is highly detectable with our AI detector. The Model 3.0.1 Turbo excelled with a 99.0% accuracy.
These results highlight the effectiveness of the Originality.ai AI detector in identifying AI-generated content, ensuring reliable detection across various text generation approaches.
Interested in learning more about AI detection? Check out our guides:
Have you seen a thought leadership LinkedIn post and wondered if it was AI-generated or human-written? In this study, we looked at the impact of ChatGPT and generative AI tools on the volume of AI content that is being published on LinkedIn. These are our findings.
We believe that it is crucial for AI content detectors reported accuracy to be open, transparent, and accountable. The reality is, each person seeking AI-detection services deserves to know which detector is the most accurate for their specific use case.