Heaptalk, Jakarta — Google and Amazon-backed AI startup Anthropic unleashed its recent AI model family Claude 3 to defeat ChatGPT and Gemini in the AI race (03/04).
The release consists of three models in ascending order of capability, namely Haiku, Sonnet, and Opus. The startup said that each model delivers powerful performance, enabling users to select the optimal balance of intelligence, speed, and cost for their specific application.
Anthropic stated that all Claude 3 models exhibit improved capabilities in analysis and forecasting, nuanced content creation, code generation, and conversing in non-English languages such as Spanish, Japanese, and French.
Opus is the smartest model as it outperforms its peers in most common evaluation benchmarks for AI systems, including undergraduate-level expert knowledge (MMLU), graduate-level expert reasoning (GPQA), and basic mathematics (GSM8K). This model demonstrates a near-human level of understanding and fluency in complex tasks.
Support live customer chat and data extraction
In addition, the Claude 3 models can support live customer chat, auto-completions, and data extraction tasks where responses must be immediate and in real-time. Haiku is claimed as the fastest and most cost-effective model on the market. This model can read an information and data-dense research paper on arXiv (~10k tokens) with charts and graphs in less than three seconds.
Meanwhile, for the vast majority of workloads, Sonnet can perform 2x faster than Claude 2 and 2.1 with higher levels of intelligence. This model outperforms others on tasks that demand rapid responses, such as knowledge retrieval or sales automation. Opus delivers similar speeds to Claude 2 and 2.1 but with much higher levels of intelligence.
Further, Anthropic plans to release more updates on this AI model family shortly. “We do not believe that model intelligence is anywhere near its limits, and we plan to release frequent updates to the Claude 3 model family over the next few months. We’re also excited to release a series of features to enhance our models’ capabilities, particularly for enterprise use cases and large-scale deployments. These new features will include Tool Use (aka function calling), interactive coding (aka REPL), and more advanced agentic capabilities,” the company concluded.