Trader consensus overwhelmingly favors Anthropic at 98.6% implied probability for having the best AI model by end of March 2024, driven by the March 4 release of its Claude 3 family—particularly Opus—which surged to the top of the LMSYS Chatbot Arena leaderboard, a blind user-voted benchmark of real-world large language model capabilities. Claude 3 outperformed rivals like OpenAI's GPT-4 and Google's Gemini on key metrics including reasoning (GPQA), multilingual tasks (MMLU), and vision understanding, maintaining the lead through month's end amid no major competing releases. This skin-in-the-game positioning reflects demonstrated AI superiority, though slim odds on others acknowledge potential leaderboard disputes or alternative resolution criteria like official benchmarks. Upcoming resolution will hinge on precise end-of-March snapshot data.
Experimental AI-generated summary referencing Polymarket data · UpdatedAnthropic 98.6%
Google <1%
xAI <1%
OpenAI <1%
$11,732,975 Vol.
$11,732,975 Vol.

Anthropic
99%

<1%

xAI
<1%

OpenAI
<1%

DeepSeek
<1%

Baidu
<1%

Moonshot
<1%

Meituan
<1%

Alibaba
<1%

Z.ai
<1%

Mistral
<1%
Anthropic 98.6%
Google <1%
xAI <1%
OpenAI <1%
$11,732,975 Vol.
$11,732,975 Vol.

Anthropic
99%

<1%

xAI
<1%

OpenAI
<1%

DeepSeek
<1%

Baidu
<1%

Moonshot
<1%

Meituan
<1%

Alibaba
<1%

Z.ai
<1%

Mistral
<1%
Results from the "Arena Score" section on the Leaderboard tab of https://lmarena.ai/leaderboard/text with the style control off will be used to resolve this market.
If two models are tied for the highest arena score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order (e.g., if both were tied, "Google" would resolve to "Yes", and "xAI" would resolve to "No")
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Market Opened: Dec 2, 2025, 6:02 PM ET
Resolver
0x2F5e3684c...Resolver
0x2F5e3684c...Trader consensus overwhelmingly favors Anthropic at 98.6% implied probability for having the best AI model by end of March 2024, driven by the March 4 release of its Claude 3 family—particularly Opus—which surged to the top of the LMSYS Chatbot Arena leaderboard, a blind user-voted benchmark of real-world large language model capabilities. Claude 3 outperformed rivals like OpenAI's GPT-4 and Google's Gemini on key metrics including reasoning (GPQA), multilingual tasks (MMLU), and vision understanding, maintaining the lead through month's end amid no major competing releases. This skin-in-the-game positioning reflects demonstrated AI superiority, though slim odds on others acknowledge potential leaderboard disputes or alternative resolution criteria like official benchmarks. Upcoming resolution will hinge on precise end-of-March snapshot data.
Experimental AI-generated summary referencing Polymarket data · Updated
Beware of external links.
Beware of external links.
Frequently Asked Questions