Chinese AI firm Zhipu AI (Z.ai) announced it trained its new GLM-Image model exclusively using Huawei's Ascend and Kunpeng hardware, marking a potential milestone in domestic AI capability. The company's claim, however, lacks details on the scale and efficiency of the training process, leaving questions about its practical impact on the global chip market.
Chinese artificial intelligence company Zhipu AI, which operates under the brand Z.ai, has announced a significant technical achievement: the training of its new GLM-Image model was conducted entirely on hardware from Huawei. This claim, if verified, represents the first time a company has built an advanced AI model using a fully domestic Chinese hardware stack, bypassing the need for Nvidia or AMD GPUs.
The model in question, GLM-Image, is described by Z.ai as employing a novel "autoregressive + diffusion decoder" hybrid architecture. This design enables the joint generation of images and language, representing an evolution beyond previous image-generation models. The company has made the model open source, with details available on its Hugging Face page.

The hardware at the center of this claim is Huawei's Ascend Atlas 800T A2 server. This system is built around Huawei's Kunpeng 920 processors, which use Arm-based cores of Huawei's own design, and is equipped with Huawei's Ascend 910 AI accelerators. The most recent iteration, the Ascend 910C, is reported by Huawei to deliver approximately 800 teraflops of computing power at FP16 precision, which they position as roughly 80% of the performance of Nvidia's H100 chip from 2022.
Z.ai's technical breakdown of GLM-Image outlines a two-part architecture. The first component is an autoregressive generator, a 9-billion-parameter model initialized from their existing GLM-4-9B-0414 model but with an expanded vocabulary to handle visual tokens. This generator first creates a compact encoding of about 256 tokens, which is then expanded to a range of 1,000 to 4,000 tokens to produce high-resolution images. The second component is a diffusion decoder, a 7-billion-parameter model based on a single-stream Diffusion Transformer (DiT) architecture for decoding images in latent space. This decoder includes a Glyph Encoder text module, specifically designed to improve the accurate rendering of text within generated images.
According to Z.ai, "the entire process from data preprocessing to large-scale training" was completed using the Atlas server platform. The company frames this as proof of "the feasibility of training cutting-edge models on a domestically produced full-stack computing platform."
However, a critical piece of information is missing: Z.ai has not disclosed the scale of the hardware deployment. The company has not revealed how many Ascend Atlas servers or how many Ascend 910 accelerators were used to train GLM-Image, nor has it provided any metrics on the training time or cost. This lack of transparency makes it difficult to assess the true performance and economic viability of Huawei's hardware for large-scale model training compared to established industry leaders like Nvidia.
Without these details, the announcement can be viewed as a strategic proof-of-concept. It demonstrates that a capable model can be built using Chinese hardware, but it does not yet provide evidence that this method can compete on speed, efficiency, or cost at the scale required for the most advanced AI systems. The contribution of Arm's architecture to the Kunpeng processors also complicates the narrative of a purely domestic technology stack.
Despite these caveats, the announcement is significant within the current geopolitical and technological landscape. It aligns with predictions that future AI development may trend toward smaller, specialized models for niche domains. If China can reliably produce such models without relying on Western hardware, it represents a potential long-term threat to the revenue streams of Nvidia and AMD, particularly as the U.S. government tightens export controls on advanced semiconductors to China. Recent regulations require U.S. authorities to assess every application to sell certain GPUs to Chinese buyers, further incentivizing domestic alternatives.
The development also intersects with broader strategic concerns. Organizations like the Australian Strategic Policy Institute (ASPI) have argued that China uses AI to export its culture and governance norms, urging nations to "prevent China's AI models, governance norms and industrial policies from shaping global technology ecosystems and entrenching digital authoritarianism." As open-source models like GLM-Image become more accessible, the competition over technological standards and influence intensifies.
In summary, Z.ai's achievement is a notable milestone in China's push for technological self-sufficiency in AI. It showcases the potential of Huawei's Ascend platform and signals a growing capability within China's domestic tech ecosystem. However, the absence of performance benchmarks and deployment scale means that while the claim is technically interesting, its immediate impact on the global hardware market remains uncertain. The industry will be watching closely for more concrete data on training efficiency and cost to determine whether this represents a genuine shift in the AI hardware landscape.

Comments
Please log in or register to join the discussion