The gap is widening. Open-source models have largely hit the ceiling of what’s achievable with the current technical stack.
Recent open-source models—DeepSeek 3.2, Kimi-K2, Qwen3, and GLM-4.6—are all quite similar in capability, with minimal differences in technical approach; most variations now lie in implementation details.
In contrast, closed-source models show clearer progress: Gemini 3 series has improved significantly, Claude 4.5 is also notably better, while OpenAI is harder to assess—its updates feel patchy, and GPT-5.2 isn’t particularly impressive.
In terms of engineering readiness—especially for programming—closed-source models are far more usable in practice. Claude, in particular, stands out.
Open-source models may suffice for small projects, but their suitability for real production environments remains questionable.
I believe this stems mainly from two constraints: compute limitations and training data scarcity.
Closed-source models still rely fundamentally on scaling—using massive compute to train on vast datasets, gradually mitigating hallucinations through data volume and quality.
In China, compute resources are constrained, and high-quality training data is scarce. Most top-tier academic papers are in English and under copyright, limiting accessibility.
Moreover, when it comes to code, domestic internet companies simply don’t have as much proprietary code as giants like Google. Google can let its AI learn from its own internal, closed-source codebase—but Qwen, as an open-source model, cannot easily access Alibaba’s private code due to serious security and confidentiality concerns.
This gives Google a tremendous advantage.
Google Search has amassed enormous data, its cloud infrastructure provides ample compute, YouTube contributes rich video content, it has deep financial resources, and it designs its own TPUs.
Overall, Google maintains a substantial lead.
Among Chinese players, Alibaba might be in the best position—but its data is fragmented. Alibaba has cloud and e-commerce data, ByteDance owns short-video data, Baidu holds search data, while DeepSeek lacks access to any high-quality proprietary dataset.
The reason Chinese AI models are somewhat competent in coding is largely thanks to the abundance of open-source projects that can be used as training data. But in other specialized domains, their capabilities still lag noticeably. |