Alibaba Open-Sources Tongyi DeepResearch: 30B MoE LLM Built for Long-Horizon Web Research

What Tongyi DeepResearch is

Alibaba Tongyi Lab has released Tongyi-DeepResearch-30B-A3B, an open-source agent-specialized large language model designed for long-horizon, tool-augmented information-seeking. The model uses a mixture-of-experts (MoE) architecture with about 30.5 billion total parameters and roughly 3.0–3.3 billion active parameters per token. That design aims to keep inference cost close to a small dense model while preserving specialist capacity needed for complex reasoning during multi-turn research workflows.

Benchmarks and reported performance

Tongyi DeepResearch reports leading results on several agentic search and deep-research benchmarks. Notable scores include:

The team also reports strong performance across WebWalkerQA, GAIA, FRAMES, and SimpleQA. According to the release, the model matches or outperforms many proprietary and open-source agents on these agentic, tool-mediated tasks.

Architecture and inference profile

Key architectural and inference details:

This combination targets a practical balance between throughput and specialist capability for deep research tasks.

Training pipeline: synthetic data and on-policy RL

Tongyi DeepResearch is trained end-to-end as an agent rather than only a chat model. The release highlights a fully automated and scalable data engine powering the model:

These elements are designed to make the model robust to multi-turn tool interactions and to reduce hallucination and drift during long sessions.

Role in document and web research workflows

Deep-research tasks typically require four capabilities: long-horizon planning, iterative retrieval and verification across multiple sources, evidence tracking with low hallucination, and synthesis under large contexts. Tongyi DeepResearch addresses these with:

The reported benchmark gains suggest improved robustness on multi-hop, tool-mediated queries where previous agents often overfit to prompt patterns or saturate at shallow depths.

Key features summarized

Practical implications and where to find the release

For teams building long-horizon research agents, Tongyi DeepResearch represents a reproducible, open-source stack that trades off inference cost and capability in a practical way. The release includes model weights under an Apache-2.0 license, inference scripts, and evaluation utilities. Models and code are available on platforms like Hugging Face and GitHub with technical details, tutorials, and notebooks to help practitioners reproduce and extend the work.

For researchers focused on multi-turn web exploration, evidence-driven synthesis, and tool-mediated agent workflows, this release is worth investigating as both a benchmark and a baseline for further development.