Trader sentiment in the Chatbot Arena market hinges on accelerating progress in large language model capabilities, as measured by LMSYS Arena's crowd-sourced Elo ratings from blind user battles. As of late March 2026, Anthropic's Claude Opus 4.6 leads across Text, Code, and Document arenas—scoring 1525 in documents—with OpenAI's GPT-5.4 High entering the Text top 10 via gains in creative writing (+46 Elo points) and expert queries, while xAI's Grok 4.20 Beta Reasoning ties at #7. Recent entrants like Google Gemini 3.1 Pro, DeepSeek 3.2, and Llama 4 intensify competition, reflecting labs' rapid scaling amid test-time compute advances. Upcoming model drops from these frontrunners could propel the peak score past 1550 by December 31, though benchmark saturation risks tempering gains.
Polymarketデータを参照したAI生成の実験的な要約 · 更新日$66,092 Vol.
↑ 1550
61%
↑ 1600
26%
↑ 1650
13%
↑ 1700
11%
$66,092 Vol.
↑ 1550
61%
↑ 1600
26%
↑ 1650
13%
↑ 1700
11%
Results from the 'Score' section on the 'Text Arena' Leaderboard tab (https://lmarena.ai/leaderboard/text), with the style control unchecked, will be used to resolve this market.
The resolution source is the Chatbot Arena LLM Leaderboard (https://lmarena.ai/). If this source is temporarily unavailable, the market remains open until it is accessible again; if permanently unavailable, this market will resolve to "No".
マーケット開始日: Jan 2, 2026, 1:29 PM ET
Resolver
0x65070BE91...提案された結果: はい
異議申し立てなし
最終結果: はい
Results from the 'Score' section on the 'Text Arena' Leaderboard tab (https://lmarena.ai/leaderboard/text), with the style control unchecked, will be used to resolve this market.
The resolution source is the Chatbot Arena LLM Leaderboard (https://lmarena.ai/). If this source is temporarily unavailable, the market remains open until it is accessible again; if permanently unavailable, this market will resolve to "No".
Resolver
0x65070BE91...提案された結果: はい
異議申し立てなし
最終結果: はい
Trader sentiment in the Chatbot Arena market hinges on accelerating progress in large language model capabilities, as measured by LMSYS Arena's crowd-sourced Elo ratings from blind user battles. As of late March 2026, Anthropic's Claude Opus 4.6 leads across Text, Code, and Document arenas—scoring 1525 in documents—with OpenAI's GPT-5.4 High entering the Text top 10 via gains in creative writing (+46 Elo points) and expert queries, while xAI's Grok 4.20 Beta Reasoning ties at #7. Recent entrants like Google Gemini 3.1 Pro, DeepSeek 3.2, and Llama 4 intensify competition, reflecting labs' rapid scaling amid test-time compute advances. Upcoming model drops from these frontrunners could propel the peak score past 1550 by December 31, though benchmark saturation risks tempering gains.
Polymarketデータを参照したAI生成の実験的な要約 · 更新日
外部リンクに注意してください。
外部リンクに注意してください。
よくある質問