Tencent Unveils Hunyuan-MT-7B and Chimera-7B: Open-Source Breakthrough in Multilingual Translation

New open models from Tencent Hunyuan

Tencent’s Hunyuan team released two open-source translation systems: Hunyuan-MT-7B, a compact 7B-parameter translation model, and Hunyuan-MT-Chimera-7B, an ensemble-style weak-to-strong fusion model. Both are targeted at multilingual machine translation and were introduced alongside Tencent’s WMT2025 submission, where Hunyuan-MT-7B ranked first in 30 of 31 language pairs.

Model designs and capabilities

Hunyuan-MT-7B is a 7 billion parameter model engineered for mutual translation among 33 languages. The supported set includes major languages as well as Chinese minority languages such as Tibetan, Mongolian, Uyghur, and Kazakh. The model is optimized for both high-resource and low-resource scenarios and claims state-of-the-art results for models of comparable size.

Hunyuan-MT-Chimera-7B is an integrated weak-to-strong fusion model. At inference time it combines multiple candidate outputs and uses reward-driven aggregation and reinforcement learning to produce a refined final translation. According to the team, Chimera-7B is the first open-source model of this class and yields quality improvements beyond single-system outputs.

Training methodology

Tencent describes a five-stage framework used to train these models:

Benchmark and human evaluations

Automatic benchmarks report strong results across multiple test suites:

Comparative highlights include outperformance of Google Translate by 15–65% across different evaluation categories. Despite its smaller parameter count, Hunyuan-MT-7B also beats some specialized translation models such as Tower-Plus-9B and Seed-X-PPO-7B. Chimera-7B contributes an additional approximate 2.3% improvement on FLORES-200, especially in Chinese⇔Other and non-English⇔non-Chinese directions.

Human evaluation used a custom multi-domain set (social, medical, legal, internet). Scores were: Hunyuan-MT-7B average 3.189, Gemini-2.5-Pro 3.223, DeepSeek-V3 3.219, and Google Translate 2.344. These results show that a 7B model can approach the quality of much larger proprietary systems.

Real-world strengths and case studies

The technical report highlights practical translation examples:

Implications for research and deployment

By open-sourcing Hunyuan-MT-7B and Hunyuan-MT-Chimera-7B, Tencent provides high-performance, accessible tools for multilingual translation research and real-world deployment. The combination of targeted pre-training, careful data curation, and RL-based refinement demonstrates a practical path for improving translation quality in both high- and low-resource languages. Researchers and engineers can inspect the GitHub repo and technical report for details, data processing recipes, and evaluation protocols to reproduce or extend these results.

For full technical details, see the team’s report and repository:

https://github.com/Tencent-Hunyuan/Hunyuan-MT/blob/main/Hunyuan_MT_Technical_Report.pdf https://github.com/Tencent-Hunyuan/Hunyuan-MT