OpenAI's o1 reasoning models command a dominant position on math benchmarks like the challenging MATH dataset, achieving over 94% accuracy—well ahead of rivals such as Google's Gemini or Anthropic's Claude 3.5 Sonnet—driving the 98.8% trader consensus for leadership through March 31. Released in September 2024, o1's chain-of-thought capabilities have set new standards in mathematical problem-solving, with no competitive releases or demonstrations surpassing it in recent weeks. This skin-in-the-game sentiment anticipates OpenAI's iterative improvements amid a quiet period for peers. Realistic challenges include surprise launches like xAI's Grok-3, Google's next Gemini iteration, or breakthroughs from DeepSeek, potentially shifting leaderboards before the deadline.
Polymarketデータを参照したAI生成の実験的な要約 · 更新日OpenAI 98.8%
Google <1%
DeepSeek <1%
アンソロピック <1%
$442,827 Vol.
$442,827 Vol.

OpenAI
99%

1%

DeepSeek
<1%

アンソロピック
<1%

xAI
<1%

ムーンショット
<1%

Z.ai
<1%

ミストラル
<1%

アリババ
<1%
OpenAI 98.8%
Google <1%
DeepSeek <1%
アンソロピック <1%
$442,827 Vol.
$442,827 Vol.

OpenAI
99%

1%

DeepSeek
<1%

アンソロピック
<1%

xAI
<1%

ムーンショット
<1%

Z.ai
<1%

ミストラル
<1%

アリババ
<1%
If two models are tied for the highest LiveBench Mathematics Average score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order.
The primary source of resolution for this market will be LiveBench’s AI leaderboard, specifically the “Mathematics Average” category, found at livebench.ai. If this resolution source is unavailable at check time, this market will remain open until the LiveBench AI leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
マーケット開始日: Dec 12, 2025, 1:25 PM ET
Resolver
0x2F5e3684c...If two models are tied for the highest LiveBench Mathematics Average score at this market's check time, resolution will be based on whichever company's name, as it is described in this market group, comes first in alphabetical order.
The primary source of resolution for this market will be LiveBench’s AI leaderboard, specifically the “Mathematics Average” category, found at livebench.ai. If this resolution source is unavailable at check time, this market will remain open until the LiveBench AI leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x2F5e3684c...OpenAI's o1 reasoning models command a dominant position on math benchmarks like the challenging MATH dataset, achieving over 94% accuracy—well ahead of rivals such as Google's Gemini or Anthropic's Claude 3.5 Sonnet—driving the 98.8% trader consensus for leadership through March 31. Released in September 2024, o1's chain-of-thought capabilities have set new standards in mathematical problem-solving, with no competitive releases or demonstrations surpassing it in recent weeks. This skin-in-the-game sentiment anticipates OpenAI's iterative improvements amid a quiet period for peers. Realistic challenges include surprise launches like xAI's Grok-3, Google's next Gemini iteration, or breakthroughs from DeepSeek, potentially shifting leaderboards before the deadline.
Polymarketデータを参照したAI生成の実験的な要約 · 更新日
外部リンクに注意してください。
外部リンクに注意してください。
よくある質問