Zhipu AI Temporarily Caps GLM Coding Plan Sales After GLM‑4.7 Triggers Surge in Demand

Zhipu AI will temporarily cap daily sales of its GLM Coding Plan to 20% of current levels starting January 23 after the GLM‑4.7 release caused spikes in usage and intermittent slowdowns during weekday peak hours. The move protects existing subscribers and highlights persistent infrastructure and cost challenges for providers of code‑focused large language models.

Young professional woman working on a laptop in an office setting, concentrating on her task.

Key Takeaways

  • 1Zhipu will limit new daily sales of GLM Coding Plan to 20% starting Jan 23, with daily quota refreshes at 10:00.
  • 2The restriction follows a surge in users after the launch of GLM‑4.7, which caused concurrency errors and slower model responses during 15:00–18:00 weekdays.
  • 3Existing automatic renewals are unaffected; the end date for the sales cap has not been announced.
  • 4The move signals compute and capacity constraints that could prompt infrastructure scaling, tiered access or pricing changes.

Editor's
Desk

Strategic Analysis

This temporary cap is a tactical response to a strategic problem: demand for capable coding models is outstripping supply of inference capacity. For Zhipu, the choice to ration access preserves service quality for paying customers but risks alienating potential new adopters and opening a window for rivals. In the medium term, expect Zhipu to pursue a mix of measures—securing more cloud or on‑prem capacity, segmenting products into priority tiers, or tightening commercial terms—to reconcile demand with finite compute budgets. The episode also illustrates a broader industry reality in China and globally: model innovation alone is insufficient without scalable, cost‑efficient infrastructure and clear access policies.

China Daily Brief Editorial
Strategic Insight
China Daily Brief

Zhipu AI has announced a temporary, steep reduction in daily sales of its GLM Coding Plan after the rollout of GLM‑4.7 produced a sharp increase in user activity. Beginning January 23 at 10:00, the company will limit new daily sales to 20% of current levels and refresh that limited quota each day at 10:00, while leaving existing automatic renewals untouched. Zhipu cited intermittent concurrency errors and slow responses during weekday peak hours (15:00–18:00) as the immediate reason for the move, and said the suspension will remain until further notice.

The decision underscores the heavy compute demands of modern, code‑focused large language models and the operational strain they can place on vendors both large and small. GLM, Zhipu’s family of models, has been positioned as a domestic alternative to Western offerings, and GLM‑4.7 appears to have attracted a wave of new users that outpaced the firm’s capacity planning. Concurrency throttling and slower inference at peak times are common symptoms when model adoption grows faster than infrastructure upgrades can match.

For customers and developers the cap is a mixed signal: it protects service quality for existing subscribers but limits market access for new users and could slow onboarding. By prioritizing renewals and established customers, Zhipu is choosing short‑term stability over open growth, a pragmatic step that also reveals near‑term limits in its AI infrastructure or cloud provisioning strategy. Competitors and cloud providers will watch closely for signs of whether this is a temporary squeeze or a longer‑term capacity gap.

The episode is instructive about the economics of AI in China. Running large inference fleets remains expensive and technically complex, whether hosted on domestic clouds or self‑managed hardware. How Zhipu responds—by expanding compute, introducing tiered access, raising prices, or partnering for capacity—will shape developer sentiment and commercial traction for GLM‑branded models in a crowded Chinese AI market. Observers should track the duration of the cap, any changes to pricing or SLAs, and whether rivals seek to poach frustrated users.

Share Article

Related Articles

📰
No related articles found