Anthropic leads trader consensus at 50.5% implied probability for having the second-best AI model by end of June, propelled by the June 20 release of Claude 3.5 Sonnet, which rocketed to the top of the LMSYS Chatbot Arena leaderboard with superior coding, math, and reasoning benchmarks, surpassing OpenAI's GPT-4o and pressuring Google's Gemini 1.5 series. Google's 29.5% reflects its edge in long-context processing and multimodal capabilities, keeping it competitive amid recent leaderboard volatility. OpenAI's drop to 9.5% stems from GPT-4o's displacement, while xAI's 8.5% bets on Grok-1.5V improvements; smaller players like DeepSeek trail on open-source coding strength. Watch for late-June benchmark updates or surprise model drops that could swing the closely contested race.
Experimental AI-generated summary referencing Polymarket data · UpdatedAnthropic 51%
Google 30%
xAI 10%
OpenAI 10%
$69,154 Vol.
$69,154 Vol.

Anthropic
51%

30%

xAI
10%

OpenAI
10%

DeepSeek
4%

Mistral
2%

Meituan
2%

Alibaba
1%

Z.ai
<1%

Moonshot
<1%
Anthropic 51%
Google 30%
xAI 10%
OpenAI 10%
$69,154 Vol.
$69,154 Vol.

Anthropic
51%

30%

xAI
10%

OpenAI
10%

DeepSeek
4%

Mistral
2%

Meituan
2%

Alibaba
1%

Z.ai
<1%

Moonshot
<1%
Results from the "Arena Score" section on the Leaderboard tab of https://lmarena.ai/leaderboard/text with the style control off will be used to resolve this market.
If two models are tied for the second best arena score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order (e.g. if both were tied, "Google" would resolve to "Yes", and "xAI" would resolve to "No")
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Market Opened: Oct 10, 2025, 5:27 PM ET
Resolver
0x2F5e3684c...Results from the "Arena Score" section on the Leaderboard tab of https://lmarena.ai/leaderboard/text with the style control off will be used to resolve this market.
If two models are tied for the second best arena score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order (e.g. if both were tied, "Google" would resolve to "Yes", and "xAI" would resolve to "No")
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x2F5e3684c...Anthropic leads trader consensus at 50.5% implied probability for having the second-best AI model by end of June, propelled by the June 20 release of Claude 3.5 Sonnet, which rocketed to the top of the LMSYS Chatbot Arena leaderboard with superior coding, math, and reasoning benchmarks, surpassing OpenAI's GPT-4o and pressuring Google's Gemini 1.5 series. Google's 29.5% reflects its edge in long-context processing and multimodal capabilities, keeping it competitive amid recent leaderboard volatility. OpenAI's drop to 9.5% stems from GPT-4o's displacement, while xAI's 8.5% bets on Grok-1.5V improvements; smaller players like DeepSeek trail on open-source coding strength. Watch for late-June benchmark updates or surprise model drops that could swing the closely contested race.
Experimental AI-generated summary referencing Polymarket data · Updated


Beware of external links.
Beware of external links.
Frequently Asked Questions