OpenAI's o1 reasoning models have driven overwhelming trader consensus at 95.5% implied probability for dominating the best AI model for coding by March 31, 2025, thanks to their state-of-the-art performance on key benchmarks like SWE-bench (48.9% for o1-preview) and superior chain-of-thought reasoning that excels in complex programming tasks. This edges out Anthropic's Claude 3.5 Sonnet (49.0% on SWE-bench), bolstered by OpenAI's massive compute resources, rapid iteration cycle, and anticipation for the full o1 release alongside potential GPT-5 (codenamed Orion) in early 2025. While the lead reflects skin-in-the-game confidence in OpenAI's competitive moat, realistic challenges include an unexpected Claude 4 launch from Anthropic, breakthroughs from Google DeepMind's Gemini 2.0, or regulatory hurdles delaying OpenAI's roadmap.
Experimental AI-generated summary referencing Polymarket data · UpdatedOpenAI 95%
Anthropic 4.0%
Google <1%
DeepSeek <1%
$959,197 Vol.
$959,197 Vol.

OpenAI
95%

Anthropic
4%

<1%

DeepSeek
<1%

Mistral
<1%

Alibaba
<1%

xAI
<1%

Z.ai
<1%

Moonshot
<1%
OpenAI 95%
Anthropic 4.0%
Google <1%
DeepSeek <1%
$959,197 Vol.
$959,197 Vol.

OpenAI
95%

Anthropic
4%

<1%

DeepSeek
<1%

Mistral
<1%

Alibaba
<1%

xAI
<1%

Z.ai
<1%

Moonshot
<1%
If two models are tied for the top LiveBench coding average score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order.
The primary source of resolution for this market will be LiveBench’s AI leaderboard, specifically the “coding average” category, found at livebench.ai. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Market Opened: Dec 12, 2025, 1:29 PM ET
Resolver
0x2F5e3684c...Resolver
0x2F5e3684c...OpenAI's o1 reasoning models have driven overwhelming trader consensus at 95.5% implied probability for dominating the best AI model for coding by March 31, 2025, thanks to their state-of-the-art performance on key benchmarks like SWE-bench (48.9% for o1-preview) and superior chain-of-thought reasoning that excels in complex programming tasks. This edges out Anthropic's Claude 3.5 Sonnet (49.0% on SWE-bench), bolstered by OpenAI's massive compute resources, rapid iteration cycle, and anticipation for the full o1 release alongside potential GPT-5 (codenamed Orion) in early 2025. While the lead reflects skin-in-the-game confidence in OpenAI's competitive moat, realistic challenges include an unexpected Claude 4 launch from Anthropic, breakthroughs from Google DeepMind's Gemini 2.0, or regulatory hurdles delaying OpenAI's roadmap.
Experimental AI-generated summary referencing Polymarket data · Updated



Beware of external links.
Beware of external links.
Frequently Asked Questions