Anthropic's Claude 3 Opus has driven its 93.5% implied probability as the second-best AI model by March 31, propelled by its March 4 launch that vaulted it to #2 on the LMSYS Chatbot Arena leaderboard, trailing only OpenAI's GPT-4 series, with superior scores in graduate-level reasoning (GPQA) and coding (HumanEval). Official benchmarks and user-voted Elo ratings solidified trader consensus amid hype from demos showcasing multimodal vision capabilities. Google's Gemini 1.5 trails at ~4.5% odds due to experimental status and lower Arena rankings; xAI's Grok-1 lags further. Challenges include a late Gemini Pro upgrade or unexpected benchmark volatility, though tight timelines make upsets unlikely before resolution.
Experimental AI-generated summary referencing Polymarket data · UpdatedAnthropic 94%
Google 3.9%
xAI <1%
OpenAI <1%
$408,763 Vol.
$408,763 Vol.

Anthropic
94%

4%

xAI
1%

OpenAI
1%

DeepSeek
<1%

Z.ai
<1%

Alibaba
<1%

Baidu
<1%

Moonshot
<1%

Mistral
<1%

Meituan
<1%
Anthropic 94%
Google 3.9%
xAI <1%
OpenAI <1%
$408,763 Vol.
$408,763 Vol.

Anthropic
94%

4%

xAI
1%

OpenAI
1%

DeepSeek
<1%

Z.ai
<1%

Alibaba
<1%

Baidu
<1%

Moonshot
<1%

Mistral
<1%

Meituan
<1%
Results from the "Arena Score" section on the Leaderboard tab of https://lmarena.ai/leaderboard/text with the style control off will be used to resolve this market.
If two models are tied for the second best arena score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order (e.g. if both were tied, "Google" would resolve to "Yes", and "xAI" would resolve to "No")
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Market Opened: Dec 2, 2025, 6:02 PM ET
Resolver
0x2F5e3684c...Resolver
0x2F5e3684c...Anthropic's Claude 3 Opus has driven its 93.5% implied probability as the second-best AI model by March 31, propelled by its March 4 launch that vaulted it to #2 on the LMSYS Chatbot Arena leaderboard, trailing only OpenAI's GPT-4 series, with superior scores in graduate-level reasoning (GPQA) and coding (HumanEval). Official benchmarks and user-voted Elo ratings solidified trader consensus amid hype from demos showcasing multimodal vision capabilities. Google's Gemini 1.5 trails at ~4.5% odds due to experimental status and lower Arena rankings; xAI's Grok-1 lags further. Challenges include a late Gemini Pro upgrade or unexpected benchmark volatility, though tight timelines make upsets unlikely before resolution.
Experimental AI-generated summary referencing Polymarket data · Updated


Beware of external links.
Beware of external links.
Frequently Asked Questions