Published: 9:00pm, 11 Mar 2025Updated: 9:22pm, 11 Mar 2025
TikTok owner ByteDance said it has achieved a 1.71 times efficiency improvement in large language model (LLM) training, the latest Chinese tech company to achieve a breakthrough that could potentially reduce demand for Nvidia’s high-end graphics processing units (GPUs).
Advertisement
The company’s Doubao development team said they managed to “speed up” LLM training efficiency by “1.71 times” through COMET, an optimised Mixture-of-Experts (MoE) system, according to a recent paper published on arXiv, an online forum for professionals in the scientific community.
MoE is a machine learning technique where multiple expert networks are used to divide a problem space into homogeneous sections.
The technique has been extensively adopted to scale LLMs to trillion-plus parameters, while maintaining fixed computing cost. It is widely adopted by leading artificial intelligence (AI) models such as Grok and DeepSeek.

The new system has already been adopted in the company’s production environment of clusters using over 10,000 GPUs, achieving “savings of millions of GPU hours”, according to the Doubao team.
Advertisement