Tencent’s Hunyuan T1 AI reasoning model rivals DeepSeek in performance and price

Tencent Holdings has unveiled a new artificial intelligence (AI) reasoning model, Hunyuan T1, that rivals DeepSeek’s R1 in both performance and pricing.

Advertisement

The Chinese technology giant’s latest offering, launched on Friday, leverages large-scale reinforcement learning, a technique also employed by DeepSeek in its R1 reasoning model, which launched in January.

The release is an official version of the model after a beta run of T1-preview on its chatbot Yuanbao. It scored 87.2 points on the Massive Multitask Language Understanding (MMLU) Pro benchmark, a test that gauges a model’s knowledge. That bested DeepSeek-R1’s 84 points but trailed the 89.3 points achieved by OpenAI’s o1, the reasoning model that the ChatGPT maker launched in December.

T1 has achieved high scores in other benchmarks as well, including 78.2 in the American Invitational Mathematics Examination (AIME) 2024, slightly behind R1’s 79.8 and o1’s 79.2. In terms of Chinese language capabilities, T1 excelled with 91.8 points in the C-Eval suite evaluation, the same score as R1 and better than o1’s 87.8, according to Tencent.

It also rivals DeepSeek on pricing, which has been a primary advantage for the popular Chinese start-up. T1 charges 1 yuan (US$0.14) per 1 million tokens of input, while output costs 4 yuan per million tokens. The input rate is in line with R1, which charges 1 yuan per million tokens during daytime hours, and just 0.25 yuan overnight. The output pricing is also comparable, given R1’s daytime rate of 16 yuan per million tokens, which drops to 4 yuan overnight.

Advertisement

Tencent claimed to be the first in the industry to adopt a hybrid architecture combining Google’s Transformer and Mamba, developed by Carnegie Mellon University and Princeton University. Compared with pure Transformer architecture, the hybrid approach “significantly reduces training and inference costs” by cutting memory usage, according to the Chinese tech giant.

  

Read More

Leave a Reply