OpenAI's o1 reasoning models have established unchallenged dominance in artificial intelligence math capabilities, scoring 83.3% on the rigorous MATH benchmark—more than double rivals like Anthropic's Claude 3.5 Sonnet (49%) and Google's Gemini 1.5 Pro (42%)—fueling 99.3% market-implied odds for OpenAI on March 31. This trader consensus, backed by real capital, stems from o1's late-2024 release showcasing superior chain-of-thought reasoning on GSM8K and other math evaluations, with no competitive model releases or benchmark upsets in recent weeks. While delays in rival announcements reinforce the lead, a surprise pre-deadline launch from Google DeepMind or xAI with verified superior scores remains a slim but realistic challenge.
基于Polymarket数据的AI实验性摘要 · 更新于OpenAI 99.3%
谷歌 <1%
DeepSeek <1%
Anthropic <1%
$457,920 交易量
$457,920 交易量

OpenAI
99%

谷歌
<1%

DeepSeek
<1%

Anthropic
<1%

xAI
<1%

Moonshot
<1%

Z.ai
<1%

Mistral
<1%

阿里巴巴
<1%
OpenAI 99.3%
谷歌 <1%
DeepSeek <1%
Anthropic <1%
$457,920 交易量
$457,920 交易量

OpenAI
99%

谷歌
<1%

DeepSeek
<1%

Anthropic
<1%

xAI
<1%

Moonshot
<1%

Z.ai
<1%

Mistral
<1%

阿里巴巴
<1%
If two models are tied for the highest LiveBench Mathematics Average score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order.
The primary source of resolution for this market will be LiveBench’s AI leaderboard, specifically the “Mathematics Average” category, found at livebench.ai. If this resolution source is unavailable at check time, this market will remain open until the LiveBench AI leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
市场开放时间: Dec 12, 2025, 1:25 PM ET
Resolver
0x2F5e3684c...If two models are tied for the highest LiveBench Mathematics Average score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order.
The primary source of resolution for this market will be LiveBench’s AI leaderboard, specifically the “Mathematics Average” category, found at livebench.ai. If this resolution source is unavailable at check time, this market will remain open until the LiveBench AI leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x2F5e3684c...OpenAI's o1 reasoning models have established unchallenged dominance in artificial intelligence math capabilities, scoring 83.3% on the rigorous MATH benchmark—more than double rivals like Anthropic's Claude 3.5 Sonnet (49%) and Google's Gemini 1.5 Pro (42%)—fueling 99.3% market-implied odds for OpenAI on March 31. This trader consensus, backed by real capital, stems from o1's late-2024 release showcasing superior chain-of-thought reasoning on GSM8K and other math evaluations, with no competitive model releases or benchmark upsets in recent weeks. While delays in rival announcements reinforce the lead, a surprise pre-deadline launch from Google DeepMind or xAI with verified superior scores remains a slim but realistic challenge.
基于Polymarket数据的AI实验性摘要 · 更新于
警惕外部链接哦。
警惕外部链接哦。
常见问题