Prediction

Technology

Will Z.ai have the second best AI model at the end of June 2026?

-

MiroThinker

Loading...

MiroMind Deep Analysis

Verification

Sources

MiroMind Deep Analysis

10

sources

Multi-cycle verification

Deep Reasoning

The question refers to a specific, externally defined criterion: a prediction market that will resolve based on the Chatbot Arena / Arena.ai text leaderboard at a fixed check time on June 30, 2026 (12:00 PM ET). The market “Will Z.ai have the second best AI model at the end of June 2026?” resolves by looking at which company owns the model in second place on the overall text leaderboard at that time, using Arena’s Elo-based human-preference rankings of large language models.

As of mid‑May 2026, we have:

  • A live PolySimulator market mirroring real Polymarket odds for this exact question, currently pricing:

  • Yes: 0%

  • No: 100%
    with about $9.3K in simulated trading volume and the market still active (unresolved) [1].

  • Multiple independent snapshots and analyses of Arena / Chatbot Arena leaderboards and third‑party leaderboards showing how Z.ai’s GLM‑5 series compares against top frontier models from Anthropic, OpenAI, Google, xAI, DeepSeek, etc. [2–8].

The task is to forecast whether, by that June 30, 2026 check, Z.ai will own the model in the #2 overall position, not merely be competitive or strong in some benchmarks.

Key Factors

1. How the market resolves

From the PolySimulator market description for this question [1]:

  • Resolution source: Arena.ai / Chatbot Arena LLM Leaderboard (overall text leaderboard at arena.ai/leaderboard/text).

  • Check time: June 30, 2026, 12:00 PM ET.

  • Rule:

  • Take the overall text leaderboard ranking at that moment.

  • Identify the model in second place.

  • Identify the company that owns that model (e.g., Anthropic, Google, OpenAI, Z.ai, etc.).

  • The market resolves Yes if that company is Z.ai, otherwise No.

  • Tie‑breaks: If there is a tie in rank, use Arena score (and if still tied, alphabetical ordering of company names among the contenders) [1].

This matters because:

  • We care about overall human‑preference ranking, not specialized sub‑leaderboards (e.g., Code Arena).

  • It is company‑level: a second‑place model that is licensed or branded differently may not count as Z.ai, even if GLM‑based.

2. Z.ai’s current position on Arena / Arena‑derived leaderboards

From the best available snapshots up to mid‑May 2026:

  • On Arena.ai’s text leaderboard, a May 14, 2026 snapshot shows:

  • GLM‑5.1 around rank ~18, Elo ~1472, and

  • GLM‑5 around rank ~32, Elo ~1456 [2].

  • A KEAR AI snapshot for February 6, 2026 (Arena‑based) shows older Z.ai models:

  • GLM‑4.7 at rank 20 (score 1441),

  • GLM‑4.6 at rank 31,

  • GLM‑4.5 at rank 55 [3].

Other analyses confirm similar relative positioning:

  • Build Fast with AI (April 2026) notes GLM‑5 leading open‑weight SWE‑bench Verified but still clearly behind the top closed models in overall Arena Elo [4].

  • Various LLM leaderboard aggregators (LLM‑Stats, BenchLM, Swfte, ClickRank) consistently place GLM‑5/5.1 below the very top tier dominated by Claude Opus, GPT‑5.x, Gemini 3.x, and/or DeepSeek [5–8].

Taken together:

  • Z.ai’s best model in the main text Arena as of mid‑May is well outside the top 10, not just outside the top 2.

  • The Elo gap between GLM‑5.1 and the top models is significant, and human‑preference Elo is slow to move without enormous user traffic and sustained superior performance.

3. Current and recent top‑two contenders

Across independent sources focusing on Arena / Arena‑like metrics:

  • Top two slots in early–mid 2026 are dominated by some combination of:

  • Claude Opus 4.6 (Anthropic) – often #1 on Chatbot Arena text [3][4][5].

  • GPT‑5.4 Pro / GPT‑5.x (OpenAI) – top or near‑top in many aggregated leaderboards [2][6][5].

  • Gemini 3.1 Pro (Google) – leading on some benchmark suites and tightly clustered with Anthropic and OpenAI in Arena/Elo style rankings [4][6][5].

  • Grok‑4.x (xAI) and DeepSeek‑R1 / V4 – sometimes top 3–5 in specialized or multi‑arena boards [4][6][5].

Example: Build Fast with AI’s April 2026 roundup and several other trackers identify Claude Opus 4.6, GPT‑5.x, and Gemini 3.1 Pro as the main “top tier,” with Z.ai’s GLM‑5 as best‑in‑class among open‑weight models but still behind those closed models in overall Arena metrics [4–6].

Conclusion from this factor:

  • To reach second place, Z.ai would have to surpass at least one entrenched super‑frontier closed model from Anthropic, OpenAI, or Google that already commands huge user volume and strong human‑preference votes.

4. Z.ai’s benchmark strengths vs what Arena measures

Z.ai’s own docs and independent analyses highlight GLM‑5 and GLM‑5.1 as:

  • Extremely strong on coding (SWE‑bench Pro, SWE‑bench Verified, NL2Repo, Terminal‑Bench 2.0) and some agentic / tool‑use benchmarks [7][8][9].

  • Competitive or leading among open‑weight models on several coding and long‑horizon agent tasks [4][7][8].

However, Arena’s overall text leaderboard reflects:

  • General conversation quality, reasoning, helpfulness, harmlessness, and broad “chatbot” capability across many domains.

  • Human preference votes from many unscripted conversations, not just coding or specialized agentic tasks.

Independent stats (e.g., LLM‑Stats pages for GLM‑5 and GLM‑5.1) show:

  • GLM‑5 at #35 of 115 on BenchLM’s provisional leaderboard [10].

  • GLM‑5 Turbo around the Professional tier, ranked #42 of 551 in one multi‑benchmark aggregator [6].

  • GLM‑5.1 at #14 of 115 on BenchLM [10].

So GLM‑5.1 is strong, but not clearly top‑two in broad‑based benchmarks. Arena’s text Elo, being preference‑driven, tends to correlate with such broad performance, and GLM‑5.1 has not achieved a comparable Elo to the leading Claude/GPT/Gemini models.

5. Product roadmap and timing to June 30, 2026

Known Z.ai releases and updates in 2026 so far:

  • GLM‑5: Released around Feb 11–12, 2026, as a 744B‑parameter flagship model [7][8].

  • GLM‑5.1: Announced as an iteration improving stability, long‑horizon performance, and tool use across 12 benchmarks [9].

  • GLM‑5 Turbo and GLM‑5V‑Turbo: More efficiency‑focused variants, not fundamentally new capabilities leaps [6].

Crucially:

  • There is no public indication of a GLM‑5.2 or GLM‑6 or other major architecture that would be released and widely adopted before June 30, 2026.

  • Arena Elo improvements rely on:

  • A clear performance jump, and

  • Enough time and traffic for votes to accumulate.

Even if Z.ai quietly shipped a stronger model in June, it would require tens of thousands of Arena battles in weeks to climb from ~#18 to #2, surpassing the entrenched leaders. There is no precedent for such a rapid jump this late into the period, especially with no widely reported breakthrough model from Z.ai in Q2 2026.

6. Market‑implied probabilities

The PolySimulator/Polymarket‑mirrored contract for this exact question currently shows:

  • Yes: 0%

  • No: 100%

  • Volume: about $9.3K [1].

Interpretation:

  • Traders following Arena data and Z.ai news collectively view the probability of Z.ai occupying the #2 Arena spot by June 30 as effectively zero at present.

  • While prediction markets are not perfect, they are generally efficient at aggregating public information about:

  • Current leaderboard positions,

  • Model release news/rumors,

  • AI capabilities trajectories.

Unless one has strong private information contradicting that consensus (e.g., imminent Z.ai release and evidence it will be favored by Arena users against Claude/GPT/Gemini), the market’s near‑zero odds are a strong signal.

Counterarguments and Uncertainty

  1. Surprise breakthrough by Z.ai:
    Z.ai could release a fundamentally new model or training technique shortly before June 30, with performance so compelling that Arena votes rapidly propel it to #2.

  • Against this: such a release would almost certainly generate significant public and developer buzz, yet there are currently no such announcements or leaks, and typical Arena adoption curves are not that fast for a previously non‑top‑3 lab.

  1. Algorithmic change in Arena:
    If Arena.ai drastically changed its scoring or weighting in a way that favors open‑weight models or agentic‑coding capabilities, GLM‑5.1 might jump.

  • Against this: big methodology changes are rare, usually announced, and affect all models; there is no sign of an impending shift designed to favor Z.ai.

  1. Misinterpretation of “second best”:
    One might argue “second best” by some alternative metric (e.g., coding benchmarks, open‑weight only). But the market’s rules are explicit: it resolves on Arena’s overall text leaderboard, not any sub‑metric or custom ranking [1].

Taken together, these counterarguments introduce some theoretical uncertainty but do not materially raise the probability in the time frame to anything near plausibility.

Conclusion

Based on:

  • Current Arena‑related rankings, where Z.ai’s best model is ~#18 overall, far from #2.

  • The strong, persistent dominance of models from Anthropic, OpenAI, Google, and others at the top of Arena and other leaderboards.

  • The absence of any announced major new Z.ai release before June 30, 2026.

  • The prediction market’s pricing of Yes at 0% and No at 100% for this specific question.

The best analytic forecast is:

Z.ai will not have the second best AI model (i.e., will not own the model in the #2 position on the Arena.ai overall text leaderboard) at the end of June 2026.

So the actionable answer, aligning with both data and market consensus, is: No.

MiroMind Reasoning Summary

I examined how the market defines “second best,” confirming that it is tied to the Arena.ai text leaderboard and the company owning the #2 model at a specific time. I then reviewed current and historical leaderboard snapshots and multi‑benchmark aggregators to establish Z.ai’s present rank (around #18) and the strength of current top‑tier competitors. Combining this with Z.ai’s public roadmap and the prediction market’s implied probability of 0% for a Yes outcome, I judged that the probability of Z.ai reaching #2 by June 30 is effectively negligible. Remaining uncertainties (e.g., surprise releases or methodology changes) are theoretically possible but unsupported by any current public signals, so they do not materially alter the conclusion.

Deep Research

9

Reasoning Steps

Verification

4

Cycles Cross-checked

Confidence Level

High

MiroMind Deep Analysis

10

sources

Multi-cycle verification

Deep Reasoning

The question refers to a specific, externally defined criterion: a prediction market that will resolve based on the Chatbot Arena / Arena.ai text leaderboard at a fixed check time on June 30, 2026 (12:00 PM ET). The market “Will Z.ai have the second best AI model at the end of June 2026?” resolves by looking at which company owns the model in second place on the overall text leaderboard at that time, using Arena’s Elo-based human-preference rankings of large language models.

As of mid‑May 2026, we have:

  • A live PolySimulator market mirroring real Polymarket odds for this exact question, currently pricing:

  • Yes: 0%

  • No: 100%
    with about $9.3K in simulated trading volume and the market still active (unresolved) [1].

  • Multiple independent snapshots and analyses of Arena / Chatbot Arena leaderboards and third‑party leaderboards showing how Z.ai’s GLM‑5 series compares against top frontier models from Anthropic, OpenAI, Google, xAI, DeepSeek, etc. [2–8].

The task is to forecast whether, by that June 30, 2026 check, Z.ai will own the model in the #2 overall position, not merely be competitive or strong in some benchmarks.

Key Factors

1. How the market resolves

From the PolySimulator market description for this question [1]:

  • Resolution source: Arena.ai / Chatbot Arena LLM Leaderboard (overall text leaderboard at arena.ai/leaderboard/text).

  • Check time: June 30, 2026, 12:00 PM ET.

  • Rule:

  • Take the overall text leaderboard ranking at that moment.

  • Identify the model in second place.

  • Identify the company that owns that model (e.g., Anthropic, Google, OpenAI, Z.ai, etc.).

  • The market resolves Yes if that company is Z.ai, otherwise No.

  • Tie‑breaks: If there is a tie in rank, use Arena score (and if still tied, alphabetical ordering of company names among the contenders) [1].

This matters because:

  • We care about overall human‑preference ranking, not specialized sub‑leaderboards (e.g., Code Arena).

  • It is company‑level: a second‑place model that is licensed or branded differently may not count as Z.ai, even if GLM‑based.

2. Z.ai’s current position on Arena / Arena‑derived leaderboards

From the best available snapshots up to mid‑May 2026:

  • On Arena.ai’s text leaderboard, a May 14, 2026 snapshot shows:

  • GLM‑5.1 around rank ~18, Elo ~1472, and

  • GLM‑5 around rank ~32, Elo ~1456 [2].

  • A KEAR AI snapshot for February 6, 2026 (Arena‑based) shows older Z.ai models:

  • GLM‑4.7 at rank 20 (score 1441),

  • GLM‑4.6 at rank 31,

  • GLM‑4.5 at rank 55 [3].

Other analyses confirm similar relative positioning:

  • Build Fast with AI (April 2026) notes GLM‑5 leading open‑weight SWE‑bench Verified but still clearly behind the top closed models in overall Arena Elo [4].

  • Various LLM leaderboard aggregators (LLM‑Stats, BenchLM, Swfte, ClickRank) consistently place GLM‑5/5.1 below the very top tier dominated by Claude Opus, GPT‑5.x, Gemini 3.x, and/or DeepSeek [5–8].

Taken together:

  • Z.ai’s best model in the main text Arena as of mid‑May is well outside the top 10, not just outside the top 2.

  • The Elo gap between GLM‑5.1 and the top models is significant, and human‑preference Elo is slow to move without enormous user traffic and sustained superior performance.

3. Current and recent top‑two contenders

Across independent sources focusing on Arena / Arena‑like metrics:

  • Top two slots in early–mid 2026 are dominated by some combination of:

  • Claude Opus 4.6 (Anthropic) – often #1 on Chatbot Arena text [3][4][5].

  • GPT‑5.4 Pro / GPT‑5.x (OpenAI) – top or near‑top in many aggregated leaderboards [2][6][5].

  • Gemini 3.1 Pro (Google) – leading on some benchmark suites and tightly clustered with Anthropic and OpenAI in Arena/Elo style rankings [4][6][5].

  • Grok‑4.x (xAI) and DeepSeek‑R1 / V4 – sometimes top 3–5 in specialized or multi‑arena boards [4][6][5].

Example: Build Fast with AI’s April 2026 roundup and several other trackers identify Claude Opus 4.6, GPT‑5.x, and Gemini 3.1 Pro as the main “top tier,” with Z.ai’s GLM‑5 as best‑in‑class among open‑weight models but still behind those closed models in overall Arena metrics [4–6].

Conclusion from this factor:

  • To reach second place, Z.ai would have to surpass at least one entrenched super‑frontier closed model from Anthropic, OpenAI, or Google that already commands huge user volume and strong human‑preference votes.

4. Z.ai’s benchmark strengths vs what Arena measures

Z.ai’s own docs and independent analyses highlight GLM‑5 and GLM‑5.1 as:

  • Extremely strong on coding (SWE‑bench Pro, SWE‑bench Verified, NL2Repo, Terminal‑Bench 2.0) and some agentic / tool‑use benchmarks [7][8][9].

  • Competitive or leading among open‑weight models on several coding and long‑horizon agent tasks [4][7][8].

However, Arena’s overall text leaderboard reflects:

  • General conversation quality, reasoning, helpfulness, harmlessness, and broad “chatbot” capability across many domains.

  • Human preference votes from many unscripted conversations, not just coding or specialized agentic tasks.

Independent stats (e.g., LLM‑Stats pages for GLM‑5 and GLM‑5.1) show:

  • GLM‑5 at #35 of 115 on BenchLM’s provisional leaderboard [10].

  • GLM‑5 Turbo around the Professional tier, ranked #42 of 551 in one multi‑benchmark aggregator [6].

  • GLM‑5.1 at #14 of 115 on BenchLM [10].

So GLM‑5.1 is strong, but not clearly top‑two in broad‑based benchmarks. Arena’s text Elo, being preference‑driven, tends to correlate with such broad performance, and GLM‑5.1 has not achieved a comparable Elo to the leading Claude/GPT/Gemini models.

5. Product roadmap and timing to June 30, 2026

Known Z.ai releases and updates in 2026 so far:

  • GLM‑5: Released around Feb 11–12, 2026, as a 744B‑parameter flagship model [7][8].

  • GLM‑5.1: Announced as an iteration improving stability, long‑horizon performance, and tool use across 12 benchmarks [9].

  • GLM‑5 Turbo and GLM‑5V‑Turbo: More efficiency‑focused variants, not fundamentally new capabilities leaps [6].

Crucially:

  • There is no public indication of a GLM‑5.2 or GLM‑6 or other major architecture that would be released and widely adopted before June 30, 2026.

  • Arena Elo improvements rely on:

  • A clear performance jump, and

  • Enough time and traffic for votes to accumulate.

Even if Z.ai quietly shipped a stronger model in June, it would require tens of thousands of Arena battles in weeks to climb from ~#18 to #2, surpassing the entrenched leaders. There is no precedent for such a rapid jump this late into the period, especially with no widely reported breakthrough model from Z.ai in Q2 2026.

6. Market‑implied probabilities

The PolySimulator/Polymarket‑mirrored contract for this exact question currently shows:

  • Yes: 0%

  • No: 100%

  • Volume: about $9.3K [1].

Interpretation:

  • Traders following Arena data and Z.ai news collectively view the probability of Z.ai occupying the #2 Arena spot by June 30 as effectively zero at present.

  • While prediction markets are not perfect, they are generally efficient at aggregating public information about:

  • Current leaderboard positions,

  • Model release news/rumors,

  • AI capabilities trajectories.

Unless one has strong private information contradicting that consensus (e.g., imminent Z.ai release and evidence it will be favored by Arena users against Claude/GPT/Gemini), the market’s near‑zero odds are a strong signal.

Counterarguments and Uncertainty

  1. Surprise breakthrough by Z.ai:
    Z.ai could release a fundamentally new model or training technique shortly before June 30, with performance so compelling that Arena votes rapidly propel it to #2.

  • Against this: such a release would almost certainly generate significant public and developer buzz, yet there are currently no such announcements or leaks, and typical Arena adoption curves are not that fast for a previously non‑top‑3 lab.

  1. Algorithmic change in Arena:
    If Arena.ai drastically changed its scoring or weighting in a way that favors open‑weight models or agentic‑coding capabilities, GLM‑5.1 might jump.

  • Against this: big methodology changes are rare, usually announced, and affect all models; there is no sign of an impending shift designed to favor Z.ai.

  1. Misinterpretation of “second best”:
    One might argue “second best” by some alternative metric (e.g., coding benchmarks, open‑weight only). But the market’s rules are explicit: it resolves on Arena’s overall text leaderboard, not any sub‑metric or custom ranking [1].

Taken together, these counterarguments introduce some theoretical uncertainty but do not materially raise the probability in the time frame to anything near plausibility.

Conclusion

Based on:

  • Current Arena‑related rankings, where Z.ai’s best model is ~#18 overall, far from #2.

  • The strong, persistent dominance of models from Anthropic, OpenAI, Google, and others at the top of Arena and other leaderboards.

  • The absence of any announced major new Z.ai release before June 30, 2026.

  • The prediction market’s pricing of Yes at 0% and No at 100% for this specific question.

The best analytic forecast is:

Z.ai will not have the second best AI model (i.e., will not own the model in the #2 position on the Arena.ai overall text leaderboard) at the end of June 2026.

So the actionable answer, aligning with both data and market consensus, is: No.

MiroMind Reasoning Summary

I examined how the market defines “second best,” confirming that it is tied to the Arena.ai text leaderboard and the company owning the #2 model at a specific time. I then reviewed current and historical leaderboard snapshots and multi‑benchmark aggregators to establish Z.ai’s present rank (around #18) and the strength of current top‑tier competitors. Combining this with Z.ai’s public roadmap and the prediction market’s implied probability of 0% for a Yes outcome, I judged that the probability of Z.ai reaching #2 by June 30 is effectively negligible. Remaining uncertainties (e.g., surprise releases or methodology changes) are theoretically possible but unsupported by any current public signals, so they do not materially alter the conclusion.

Deep Research

9

Reasoning Steps

Verification

4

Cycles Cross-checked

Confidence Level

High

MiroMind Verification Process

1
Confirmed the resolution rules and data source from the PolySimulator market page for the exact question.

Verified

2
Checked Arena.ai and summarized Z.ai’s current rankings (GLM‑5, GLM‑5.1) relative to top models.

Verified

3
Cross‑referenced independent leaderboards and benchmark aggregators to validate that Z.ai is not near top‑two overall.

Verified

4
Reviewed Z.ai’s 2026 release notes and GLM‑5/5.1 documentation to assess potential for a near‑term capability leap before June 30, 2026.

Verified

Sources

[1] Will Z.ai have the second best AI model at the end of June 2026? PolySimulator market page. https://polysimulator.com/markets/0xbc3a6702307af90e1afd7b36140ce79925d4dc0b99fa430310b700c88cd53d4d

[2] LLM Leaderboard – Text Arena. Arena.ai. https://arena.ai/leaderboard/text

[3] AI Chatbot Arena Leaderboard 2026. KEAR AI (Feb 6, 2026). https://kearai.com/leaderboard/chat

[4] Best AI Models April 2026: Ranked by Benchmarks. Build Fast with AI (Apr 5, 2026). https://www.buildfastwithai.com/blogs/best-ai-models-april-2026

[10] GLM-5 Benchmarks 2026: Scores, Rankings & Performance. BenchLM. https://benchlm.ai/models/glm-5

[6] Z.ai: GLM 5 Turbo Review | Pricing, Benchmarks & Capabilities (2026). DesignForOnline. https://designforonline.com/ai-models/z-ai-glm-5-turbo/

[7] GLM-5: From Vibe Coding to Agentic Engineering. Z.ai (Feb 12, 2026). https://z.ai/blog/glm-5

[5] LLM Leaderboard 2026: Best AI Models Benchmark & Ranking. ClickRank AI. https://www.clickrank.ai/llm-leaderboard/

[8] GLM-5 Review: Chat.z.ai Pricing, Benchmarks & Agent Mode (2026). MySummit. https://mysummit.school/blog/en/glm5-zai-review-2026/

[9] GLM-5.1: Towards Long-Horizon Tasks. Z.ai. https://z.ai/blog/glm-5.1

Ask MiroMind

Deep Research

Predict

Verify

MiroMind reasons across dozens of sources and delivers answers with a full evidence trail.