Trader consensus assigns OpenAI a commanding 90.5% implied probability of delivering the top AI coding model by March 31, propelled by expectations of imminent releases like an expanded o1 series or GPT-5, fueled by its vast compute scale via Microsoft Azure and history of benchmark dominance. The o1-preview already leads LiveCodeBench (82%) and excels in reasoning-intensive coding via chain-of-thought, underscoring its edge in real-world software tasks. Anthropic's Claude 3.5 Sonnet, topping SWE-Bench Verified at 49%, poses the chief threat through rapid iteration or Claude 4 launch, while laggards like Google or xAI need breakthrough announcements to shift odds realistically.
Polymarketデータを参照したAI生成の実験的な要約 · 更新日OpenAI 91%
Anthropic 7.4%
Google 1.3%
DeepSeek <1%
$963,310 Vol.
$963,310 Vol.

OpenAI
91%

Anthropic
7%

1%

DeepSeek
<1%

xAI
<1%

Z.ai
<1%

Mistral
<1%

アリババ
<1%

ムーンショット
<1%
OpenAI 91%
Anthropic 7.4%
Google 1.3%
DeepSeek <1%
$963,310 Vol.
$963,310 Vol.

OpenAI
91%

Anthropic
7%

1%

DeepSeek
<1%

xAI
<1%

Z.ai
<1%

Mistral
<1%

アリババ
<1%

ムーンショット
<1%
If two models are tied for the top LiveBench coding average score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order.
The primary source of resolution for this market will be LiveBench’s AI leaderboard, specifically the “coding average” category, found at livebench.ai. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
マーケット開始日: Dec 12, 2025, 1:29 PM ET
Resolver
0x2F5e3684c...Resolver
0x2F5e3684c...Trader consensus assigns OpenAI a commanding 90.5% implied probability of delivering the top AI coding model by March 31, propelled by expectations of imminent releases like an expanded o1 series or GPT-5, fueled by its vast compute scale via Microsoft Azure and history of benchmark dominance. The o1-preview already leads LiveCodeBench (82%) and excels in reasoning-intensive coding via chain-of-thought, underscoring its edge in real-world software tasks. Anthropic's Claude 3.5 Sonnet, topping SWE-Bench Verified at 49%, poses the chief threat through rapid iteration or Claude 4 launch, while laggards like Google or xAI need breakthrough announcements to shift odds realistically.
Polymarketデータを参照したAI生成の実験的な要約 · 更新日
外部リンクに注意してください。
外部リンクに注意してください。
よくある質問