WebThinker
[NeurIPS 2025] π WebThinker: Empowering Large Reasoning Models with Deep Research Capability
Install / Use
/learn @RUC-NLPIR/WebThinkerREADME
π£ Latest News
- [Sep 18, 2025]: π Our paper WebThinker: Empowering Large Reasoning Models with Deep Research Capability has been accepted at NeurIPS 2025!
- [May 30, 2025]: π WebThinker now supports Google Serper API for web search! Important: Bing Search API will be retired in August 2025.
- [May 9, 2025]: The brief introduction of WebThinker can be found on platforms like X, Zhihu, and WeChat.
- [May 1, 2025]: π€ WebThinker Model Collection is now available on Hugging Face. You can deploy our optimized models for your deep research tasks.
- [May 1, 2025]: π Our paper is now available on arXiv and Hugging Face.
- [March 31, 2025]: π WebThinker Notion Page launched with comprehensive project details.
- [March 31, 2025]: π Full codebase released. WebThinker now supports deep research with open-source reasoning models like QwQ-32B.
π₯ Deep Research Agent Family
<details open><summary>Welcome to try our deep research agent series: </summary><p>DeepAgent: A General Reasoning Agent with Scalable Toolsets (New!) <br> Authors: Xiaoxi Li, Wenxiang Jiao, Jiarui Jin, Guanting Dong, Jiajie Jin, Yinuo Wang, Hao Wang, Yutao Zhu, Ji-Rong Wen, Yuan Lu, Zhicheng Dou <br> TLDR: An end-to-end deep reasoning agent that performs autonomous thinking, tool discovery, and action execution with brain-inspired memory folding mechanism. <br>
![]()
![]()
![]()
WebThinker: Empowering Large Reasoning Models with Deep Research Capability (NeurIPS 2025) <br> Authors: Xiaoxi Li*, Jiajie Jin*, Guanting Dong*, Hongjin Qian, Yutao Zhu, Yongkang Wu, Ji-Rong Wen, Zhicheng Dou <br> TLDR: A deep research agent that empowers large reasoning models with autonomous search, web browsing, and research report drafting capabilities. <br>
![]()
![]()
![]()
</p></details>Search-o1: Agentic Search-Enhanced Large Reasoning Models (EMNLP 2025) <br> Authors: Xiaoxi Li, Guanting Dong, Jiajie Jin, Yuyao Zhang, Yujia Zhou, Yutao Zhu, Peitian Zhang, Zhicheng Dou <br> TLDR: An agentic search-enhanced framework that integrates autonomous knowledge retrieval with large reasoning models through Agentic RAG and reasoning-in-documents modules. <br>
![]()
![]()
![]()
![]()
π¬ Demo
<div align="center"> <video src="https://github.com/user-attachments/assets/a38e82ec-5aed-4efe-a8b8-e9ee2d97e9b9" /> </div>π‘ Overview
WebThinker is a deep research framework fully powered by large reasoning models (LRMs). WebThinker enables LRMs to autonomously search, deeply explore web pages, and draft research reports, all within their thinking process.
Unlike existing open-source deep search agents that typically employ retrieval-augmented generation (RAG) with predefined workflows, WebThinker allows the reasoning model itself to perform actions during thinking, achieving end-to-end task execution in a single generation.
π Overall Performance
<p align="center"> <img src="figures/performance.png" width="100%" /> </p>As shown above, WebThinker consistently outperforms competing approaches on both knowledge-intensive complex reasoning benchmarks (GPQA, GAIA, WebWalkerQA, HLE) and open-ended reasoning tasks for report generation. Our WebThinker-32B with QwQ-32B as backbone reasoning model achieves superior performance across all tasks.
β¨ The WebThinker Framework

WebThinker enables reasoning models to autonomously conduct web searches and web page navigations to acquire external knowledge during their reasoning process. This approach significantly reduces the time and costs associated with information gathering for researchers in knowledge-intensive fields. Furthermore, WebThinker allows LRMs to draft section content while thinking and searching, producing comprehensive, customized reports that directly address users' research questions.
Key Features:
- We introduce a Deep Web Explorer that empowers LRMs to search, navigate pages by clicking interactive elements (like links or buttons), and extract relevant information. Based on initial search results, the LRM can initiate follow-up searches and traverse deeper links until it collects all relevant information.
- For scientific reporting, our Autonomous Think-Search-and-Draft strategy integrates real-time knowledge seeking with report creation. We equip LRMs with three specialized tools: (1) drafting content for specific chapters, (2) checking the current report, and (3) editing the reportβensuring reports remain comprehensive, coherent, and adaptive to new insights.
- We're developing RL-based training strategies to optimize end-to-end task performance by leveraging large-scale reasoning trajectories from complex tasks. Using accuracy of reasoning, tool usage, and final outputs, we construct preference pairs for online DPO training, enabling the model to progressively improve its research capabilities.
π§ Installation
Environment Setup
# Create conda environment
conda create -n webthinker python=3.9
conda activate webthinker
# Install requirements
cd WebThinker-main
pip install -r requirements.txt
π Quick Start
Pre-preparation
Model Serving
Before running WebThinker, ensure your reasoning model and auxiliary model are served using vLLM. In our experiments, we use QwQ-32B as the reasoning model and Qwen-32B-Instruct as the auxiliary model. You can also explore other instruction-tuned models as your auxiliary model, which will be used in webpage reading, report writting/editting, evaluation, etc. For detailed instructions on model serving, see here.
Web Parser Client
For better web crawling performance, we recommend setting up a web parser client in scripts/search/bing_search.py using Crawl4AI. This will help handle JavaScript-rendered content and provide more reliable webpage extraction.
Now you can run different inference modes using the provided scripts. Below are examples of how to execute each mode:
Problem Solv
Related Skills
YC-Killer
2.7kA library of enterprise-grade AI agents designed to democratize artificial intelligence and provide free, open-source alternatives to overvalued Y Combinator startups. If you are excited about democratizing AI access & AI agents, please star βοΈ this repository and use the link in the readme to join our open source AI research team.
groundhog
398Groundhog's primary purpose is to teach people how Cursor and all these other coding agents work under the hood. If you understand how these coding assistants work from first principles, then you can drive these tools harder (or perhaps make your own!).
last30days-skill
18.3kAI agent skill that researches any topic across Reddit, X, YouTube, HN, Polymarket, and the web - then synthesizes a grounded summary
sec-edgar-agentkit
10AI agent toolkit for accessing and analyzing SEC EDGAR filing data. Build intelligent agents with LangChain, MCP-use, Gradio, Dify, and smolagents to analyze financial statements, insider trading, and company filings.
