OpenAI's o1-preview and o1-mini models command a 93.5% implied probability of having the top math-performing AI by March 31, driven by their unmatched scores on benchmarks like MATH (94.8%) and AIME, leveraging advanced chain-of-thought reasoning that outpaces rivals. Trader consensus reflects no confirmed rival launches—such as Anthropic's Claude 4, xAI's Grok-3, or DeepSeek-V3 iterations—poised to challenge this lead before the deadline, with o1's September 2024 release still the gold standard amid stagnant leaderboards. Realistic threats include surprise previews from Google Gemini 2.0 or accelerated open-source advances from Mistral, but historical delays in frontier model rollouts bolster OpenAI's positioning.
基於Polymarket數據的AI實驗性摘要 · 更新於OpenAI 92%
DeepSeek 1.3%
Anthropic 1.0%
xAI 1.0%
$277,740 交易量
$277,740 交易量

OpenAI
92%

DeepSeek
1%

Anthropic
1%

xAI
1%

Moonshot
1%

谷歌
1%

Z.ai
<1%

阿里巴巴
<1%

Mistral
<1%
OpenAI 92%
DeepSeek 1.3%
Anthropic 1.0%
xAI 1.0%
$277,740 交易量
$277,740 交易量

OpenAI
92%

DeepSeek
1%

Anthropic
1%

xAI
1%

Moonshot
1%

谷歌
1%

Z.ai
<1%

阿里巴巴
<1%

Mistral
<1%
If two models are tied for the highest LiveBench Mathematics Average score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order.
The primary source of resolution for this market will be LiveBench’s AI leaderboard, specifically the “Mathematics Average” category, found at livebench.ai. If this resolution source is unavailable at check time, this market will remain open until the LiveBench AI leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
市場開放時間: Dec 12, 2025, 1:25 PM ET
Resolver
0x2F5e3684c...Resolver
0x2F5e3684c...OpenAI's o1-preview and o1-mini models command a 93.5% implied probability of having the top math-performing AI by March 31, driven by their unmatched scores on benchmarks like MATH (94.8%) and AIME, leveraging advanced chain-of-thought reasoning that outpaces rivals. Trader consensus reflects no confirmed rival launches—such as Anthropic's Claude 4, xAI's Grok-3, or DeepSeek-V3 iterations—poised to challenge this lead before the deadline, with o1's September 2024 release still the gold standard amid stagnant leaderboards. Realistic threats include surprise previews from Google Gemini 2.0 or accelerated open-source advances from Mistral, but historical delays in frontier model rollouts bolster OpenAI's positioning.
基於Polymarket數據的AI實驗性摘要 · 更新於
警惕外部連結哦。
警惕外部連結哦。
Frequently Asked Questions