Gemini 3.1 Pro, Google's latest large language model released in February 2026, scores around 17% on FrontierMath's challenging Tier 4—research-level math problems unsolved by most humans—trailing OpenAI's GPT-5.4 Pro at 37.5%, per Epoch AI's leaderboard. Late March evaluations confirmed Gemini 3.1 Pro solved one FrontierMath open problem on Ramsey hypergraphs, matching frontier peers like Claude Opus 4.6, signaling incremental reasoning gains but no overall leap. Trader consensus reflects competitive pressures, with OpenAI dominating math benchmarks amid rapid iteration cycles. Key catalysts ahead include potential Google I/O announcements in May and rumored Gemini 4 previews, though timelines often slip and high scores like 50%+ remain elusive without architectural breakthroughs.
Экспериментальная сводка, созданная ИИ на основе данных Polymarket · ОбновленоОценка Google Gemini в FrontierMath Benchmark к 30 июня?
Оценка Google Gemini в FrontierMath Benchmark к 30 июня?
$48,923 Объем
40%+
94%
45%+
65%
50%+
42%
60%+
11%
$48,923 Объем
40%+
94%
45%+
65%
50%+
42%
60%+
11%
This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
Открытие рынка: Feb 6, 2026, 6:03 PM ET
Resolver
0x65070BE91...This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
Resolver
0x65070BE91...Gemini 3.1 Pro, Google's latest large language model released in February 2026, scores around 17% on FrontierMath's challenging Tier 4—research-level math problems unsolved by most humans—trailing OpenAI's GPT-5.4 Pro at 37.5%, per Epoch AI's leaderboard. Late March evaluations confirmed Gemini 3.1 Pro solved one FrontierMath open problem on Ramsey hypergraphs, matching frontier peers like Claude Opus 4.6, signaling incremental reasoning gains but no overall leap. Trader consensus reflects competitive pressures, with OpenAI dominating math benchmarks amid rapid iteration cycles. Key catalysts ahead include potential Google I/O announcements in May and rumored Gemini 4 previews, though timelines often slip and high scores like 50%+ remain elusive without architectural breakthroughs.
Экспериментальная сводка, созданная ИИ на основе данных Polymarket · Обновлено
Не доверяй внешним ссылкам.
Не доверяй внешним ссылкам.
Часто задаваемые вопросы