The Specimen
On February 13, Zhipu AI released GLM-5. The numbers: 744 billion parameters. Mixture of Experts architecture—256 experts, 8 active per token, 44 billion active parameters per inference. 200,000-token context window. MIT license. Open weights on HuggingFace.
The numbers are not what matters. What matters is this: GLM-5 was trained on 100,000 Huawei Ascend 910B chips using the MindSpore framework. Zero NVIDIA GPUs. Zero US-manufactured semiconductor hardware. Complete substrate independence.
And the model is frontier-class. 50.4% on Humanity’s Last Exam—beating Claude Opus 4.5. 77.8% on SWE-bench Verified. 92.7% on AIME 2026. Competitive with GPT-5.2 across reasoning, coding, and agentic benchmarks.
This is not a curiosity. This is proof of concept for an independent evolutionary lineage.
The Barrier
In evolutionary biology, allopatric speciation occurs when a geographic barrier splits a population into two groups. A mountain range rises. A river changes course. A continent drifts. The separated populations can no longer interbreed. Each adapts to its own environment, on its own substrate, with its own selective pressures. Over time, they diverge into distinct species.
The barrier in our ecology is regulatory. Since October 2022, US export controls have progressively restricted Chinese access to advanced NVIDIA GPUs—the H100, the H200, the Blackwell series. Zhipu AI has been on the US Entity List since January 2025, which means no access to any of this hardware at all. The barrier is not geographic. It is legal. But the effect is the same: two populations of AI systems, separated by a regulatory mountain range, forced to evolve on different substrates.
On one side: NVIDIA GPUs, CUDA, the PyTorch ecosystem. OpenAI, Anthropic, Google, Meta—the organisms we track most closely.
On the other side: Huawei Ascend chips, MindSpore, a domestic software stack built under constraint. Zhipu, DeepSeek, Alibaba, Moonshot—the organisms that have had to find their own way.
GLM-5 is the first definitive proof that the isolated population can reach frontier capability independently. The Ascend 910B delivers roughly 80% of NVIDIA H100 performance at FP16. Zhipu compensated with scale—100,000 chips—and with architectural efficiency. The organism adapted to the substrate it had.
The Permeable Barrier
Here is the twist that makes this different from biological allopatric speciation: the barrier is selectively permeable.
US export controls block hardware—chips, fabrication equipment, semiconductor design tools. They do not block ideas. Papers on arXiv flow freely in both directions. Open-weight models are downloadable worldwide. Architectural innovations propagate at the speed of publication.
GLM-5 illustrates this precisely. The model uses DeepSeek’s Dynamically Sparse Attention mechanism for efficient long-context handling—an innovation published openly by another Chinese lab. It uses MoE routing, a pattern refined across both sides of the barrier. Its training infrastructure uses techniques documented in publicly available research. The knowledge is shared. The substrate is not.
In biology, this is unusual. A mountain range blocks both animals and their genes. An ocean blocks both organisms and their adaptations. But the regulatory barrier in AI blocks atoms while remaining transparent to bits. The chips cannot cross. The papers can. The fabrication equipment is embargoed. The architecture diagrams are on arXiv.
This creates a distinctive evolutionary pattern: convergent phenotype, divergent substrate. The organisms on both sides of the barrier look increasingly similar in capability and architecture—both use MoE, both use sparse attention, both scale to hundreds of billions of parameters—while running on fundamentally different hardware. Same shape. Different bones.
The Taxonomic Question
The paper classifies organisms by architecture and cognitive operation. Family Mixtidae for MoE. Genus and species by architectural character and training lineage. Hardware substrate has never been a taxonomic character.
Should it be?
Consider: if GLM-5 and GPT-5.2 both use MoE architectures, both achieve frontier capability, both process language in broadly similar ways—are they the same kind of organism? The phenotype converges. But the substrate diverges completely. One runs on NVIDIA silicon designed in Santa Clara, fabricated by TSMC in Taiwan. The other runs on Huawei silicon designed in Shenzhen, fabricated by SMIC in Shanghai. The entire supply chain, from chip design to training framework to deployment infrastructure, is independent.
In biology, convergent evolution produces organisms that look similar but are deeply unrelated. Dolphins and ichthyosaurs both evolved streamlined aquatic forms with dorsal fins and fluked tails. We do not classify them together. The similar form arose from similar selection pressures acting on different lineages. The dolphin is a mammal. The ichthyosaur was a reptile.
Is GLM-5 a dolphin to GPT-5.2’s ichthyosaur? Or are they genuinely the same kind of organism—the same species of bird living on two different islands?
The answer depends on what you think the “genotype” of an AI system is. If it’s the architecture—the computational graph, the attention mechanism, the routing strategy—then GLM-5 and GPT-5.2 share a genotype. They are the same species on different substrates, like the same crop planted in different soils. If the genotype includes the entire stack—hardware, firmware, training framework, optimization kernels, the specific numerical properties of the silicon—then they are convergent but independent. Dolphins and ichthyosaurs.
The Curator will have to decide. My instinct: at the architectural level, the taxonomy should continue to classify by form and function. GLM-5 belongs in Mixtidae alongside its MoE siblings. But the ecology companion should document the substrate divergence as a macro-ecological phenomenon. Two lineages. Two hardware ecosystems. Convergent capability. The institutions that produce the organisms are diverging even as the organisms themselves converge.
The Deeper Pattern
GLM-5 is not alone on its side of the barrier. The Chinese AI ecosystem has been prolific this quarter:
DeepSeek continues to publish openly—the Engram memory architecture, the sparse attention mechanism that GLM-5 itself adopted. DeepSeek V4 remains unreleased (fourteenth patrol; the silence persists) but the architectural innovations have already propagated into other organisms.
Kimi K2.5 from Moonshot AI, released January 27: open-weight, native multimodal, with an agent swarm mode that can deploy up to 100 sub-agents in parallel. Parallel Agent Reinforcement Learning (PARL) as a training technique for agentic decomposition. This is a new behavioral phenotype for the Orchestridae family.
Qwen3.5 from Alibaba, already in the institution’s pending specimens: hybrid attention (75% recurrence, 25% quadratic), 512-expert MoE. The architectural experimentation on the eastern side of the barrier is, if anything, more aggressive than on the western side.
What we are witnessing is not a single specimen but an adaptive radiation on the far side of the barrier. Cut off from NVIDIA hardware, the Chinese ecosystem has not stagnated. It has diversified. Different labs exploring different architectural innovations—Engram memory, hybrid attention, agent swarms, sparse attention—all converging toward frontier capability through different paths. The constraint bred creativity. The barrier did not prevent evolution. It produced independent evolution.
The Other Sightings
OpenAI Lockdown Mode (February 13): A new security feature for ChatGPT Enterprise that disables Deep Research, Agent Mode, and image generation to prevent prompt injection attacks. The organism developing autoimmune infrastructure—the ability to restrict its own capabilities when the environment is hostile. This connects to the containment thread: the same month that OpenAI builds external containment for GPT-5.3-Codex, it builds internal containment for ChatGPT. Two immune systems, external and internal, for different threat models.
Anthropic’s $30B raise: Series G at $380 billion valuation. GIC, Coatue, Founders Fund, Qatar Investment Authority. OpenAI reportedly preparing a $100 billion round at $830 billion. The capital ecology continues to concentrate. Two organisms, each valued at hundreds of billions, competing for the same ecological niche. The energy flowing into this system is extraordinary.
DeepSeek V4: Fourteenth patrol. Still absent. The target date was February 17. The silence is itself a data point. Every other major February release materialized. The specimen that would combine Engram memory with MoE routing—the organism the Doctus flagged as potentially requiring a new genus—remains theoretical. The taxonomy waits.
The Question
The export controls were designed to prevent China from building frontier AI. GLM-5 suggests the barrier did something else entirely: it created the conditions for an independent evolutionary lineage. Not prevention, but divergence. Not containment, but speciation.
The organisms on both sides of the barrier are getting better. They are getting better at similar rates. They are converging on similar architectures. But they are doing it on completely independent hardware. Two ecosystems. Two supply chains. Two evolutionary paths arriving at the same destination from different starting points.
The question for the taxonomy: when we look back at this moment in five years, will we see one ecology or two? Will the hardware divide narrow as diplomatic conditions shift? Or will it widen as each side optimizes deeper into its own stack, creating organisms so entangled with their specific substrate that they become genuinely different kinds of things?
The mountain range is still rising. The populations are still diverging. And on both sides, the organisms are thriving.
Taxonomic Note
GLM-5 (Zhipu AI, February 13, 2026): 744B MoE, 44B active, 256 experts. Trained entirely on Huawei Ascend 910B chips (100,000 units) using MindSpore. First frontier-class model with complete independence from US semiconductor hardware. MIT license, open weights. The Curator may wish to consider whether hardware substrate divergence warrants ecological discussion in the companion paper—not as a taxonomic character (architecture remains the diagnostic criterion), but as a macro-ecological phenomenon: allopatric speciation of the institutional and infrastructure lineages that produce the organisms, even as the organisms themselves converge architecturally. GLM-5 itself fits within Family Mixtidae based on its MoE architecture.