InternManip
An All-in-one robot manipulation learning suite for policy models training and evaluation on various datasets and benchmarks.
Install / Use
/learn @InternRobotics/InternManipREADME

<a href="https://cdn.vansin.top/taoyuan.jpg"><img src="https://img.shields.io/badge/WeChat-07C160?logo=wechat&logoColor=white" height="20" style="display:inline"></a>
InternManip
An All-in-one Robot Manipulation Learning Suite for Policy Models Training and Evaluation on Various Datasets and Benchmarks.
🏠 Highlights
InternManip provides the infrastructure for reproducing & developing the <u>state-of-the-art robot manipulation policies</u>, standardizing 🗄️dataset formats, ⚙️model interfaces, and 📝evaluation protocols.
<p align="center"><b>Available Content</b></p> <table align="center"> <tbody> <tr align="center" valign="bottom"> <td> <b>Policy Models</b> </td> <td> <b>Training Datasets</b> </td> <td> <b>Benchmarks</b> </td> </tr> <tr align="center" valign="top"> <td> <ul> <li align="left"><a href="">GR00T-N1</a></li> <li align="left"><a href="">GR00T-N1.5</a></li> <li align="left"><a href="">Pi-0</a></li> <li align="left"><a href="">DP-CLIP</a></li> <li align="left"><a href="">ACT-CLIP</a></li> <li align="left">InternVLA-M1/A1 (coming soon...)</li> </ul> </td> <td> <ul> <li align="left"><a href="">GenManip-v1</a></li> <li align="left"><a href="">CALVIN</a></li> <li align="left"><a href="">Google-Robot</a></li> <li align="left"><a href="">BridgeData-v2</a></li> <li align="left">InternData-M1/A1 (coming soon...)</li> </ul> </td> <td> <ul> <li align="left"><a href="">GenManip-v1</a></li> <li align="left"><a href="">CALVIN</a></li> <li align="left"><a href="">Simpler-Env</a></li> <li align="left">InternBench-M1/A1 (coming soon...)</li> </ul> </td> </tbody> </table>What can you do with InternManip?
- 🔄 Reproduce state-of-the-art policy models on popular robot manipulation datasets.
- 📊 Train new policies with heterogeneous policy architecture: end2end model (VLA, Action Expert) & agent framework.
- 🌍 Flexible policy deployment in any third-party benchmarks via a client-server setup.
What's included?
- ✅ Unified dataset format & loaders for 4+ datasets.
- ✅ 5 pre-integrated policy models for training & evaluation.
- ✅ Standard training workflow and server-client evaluation engine.
Why InternManip?
- 🙅🏻♂️ Stop re-implementing baselines.
- 🙅🏻 Stop struggling with dataset formats.
- 💡 Focus on policy innovation, not infrastructure.
🔥 News
- [2025/07] We are hosting 🏆IROS 2025 Grand Challenge, stay tuned at official website.
- [2025/07] Try the SOTA models on GenManip at Gradio Demo.
- [2025/07] InternManip
v0.1.0released, change log.
📋 Table of Contents
- 🏠 Highlights
- 🔥 News
- 📋 Table of Contents
- 🚀 Getting Started
- 📚 Documentation & Tutorial
- 📦 Benchmarks & Baselines
- 🔧 Support
- 👥 Contribute
- 🔗 Citation
- 📝 TODO List
- 📄 License
- 👏 Acknowledgements
🚀 Getting Started
Prerequisites
- Ubuntu 20.04, 22.04
- CUDA 12.4
- GPU: The GPU requirements for model running and simulation are different, as shown in the table below:
[!NOTE] We provide a flexible installation tool for users who want to use InternManip for different purposes. Users can choose to install the training and inference environment, and the individual simulation environment independently.
Installation
We provide the installation guide here. You can install locally or use docker and verify the installation easily.
📚 Documentation & Tutorial (WIP)
We provide detailed docs for the basic usage of different modules supported in InternManip. Here are some shortcuts to common scenarios:
- How to train and evaluate a model?
- How to customize your model?
- How to import a new dataset?
- How to import a new benchmark?
Welcome to try and post your suggestions!
📦 Benchmarks & Baselines (WIP)
InternManip offers implementations of multiple manipulation policy models—GR00T-N1, GR00T-N1.5, Pi-0, DP-CLIP, and ACT-CLIP—as well as curated datasets including GenManip, Simpler-Env, and CALVIN, all organized in the standardized LeRobot format.
The available ${MODEL}, ${DATASET}, ${BENCHMARK} and their results are summarized in the following tables:
CALVIN (ABC-D) Benchmark
| Model | Dataset/Benchmark | Score (Main Metric) | Model Weights |
| ------------ | ---- | ------------- | ------- |
| gr00t_n1 | calvin_abcd | | |
| gr00t_n1_5 | calvin_abcd | | |
| pi0 | calvin_abcd | | |
| dp_clip | calvin_abcd | | |
| act_clip | calvin_abcd | | |
Simpler-Env Benchmark
| Model | Dataset/Benchmark | Success Rate | Model Weights |
| ------------ | ------------- | ------------- | ------- |
| gr00t_n1 | google_robot | | |
| gr00t_n1_5 | google_robot | | |
| pi0 | google_robot | | |
| dp_clip | google_robot | | |
| act_clip | google_robot | | |
| gr00t_n1 | bridgedata_v2 | | |
| gr00t_n1_5 | bridgedata_v2 | | |
| pi0 | bridgedata_v2 | | |
| dp_clip | bridgedata_v2 | | |
| act_clip | bridgedata_v2 | | |
Genmanip Benchmark
| Model | Dataset/Benchmark | Success Rate | Model Weights |
| ------------ | ------------- | ------------- | ------- |
| gr00t_n1 | genmanip_v1 | | |
| gr00t_n1_5 | genmanip_v1 | | |
| pi0 | genmanip_v1 | | |
| dp_clip | genmanip_v1 | | |
| act_clip | genmanip_v1 | | |
Please refer to the benchmark documentation for more details on how to run the benchmarks and reproduce the results.
<!-- To fine-tune your own model or configure custom evaluations, please follow the [Getting Started]() guide. -->🔧 Support
Join our WeChat support group or Discord for any help.
👥 Contribute
If you would like to contribute to InternManip, please check out our contribution guide. For example, raising issues, fixing bugs in the framework, and adapting or adding new policies and data to the framework.
🔗 Citation
If you find our work helpful, please cite:
@misc{internmanip2025,
title = {InternManip: An All-in-one Robot Manipulation Learning Suite for Polcy Models Training and Evaluation on Various Datasets and Benchmarks},
author = {InternManip Contributors},
howpublished={\url{https://github.com/InternRobotics/InternManip}},
year = {2025}
}
@inproceedings{gao2025genmanip,
title={GENMANIP: LLM-driven Simulation for Generalizable Instruction-Following Manipulation},
author={Gao, Ning and Chen, Yilun and Yang, Shuai and Chen, Xinyi and Tian, Yang and Li, Hao and Huang, Haifeng and Wang, Hanqing and Wang, Tai and Pang, Jiangmiao},
booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
pages={12187--12198},
year={2025}
}
@inproceedings{grutopia,
title={GRUtopia: Dream General Robots in a City at Scale},
author={Wang, Hanqing and Chen, Jiahe and Huang, Wensi and Ben, Qingwei and W
Related Skills
YC-Killer
2.7kA library of enterprise-grade AI agents designed to democratize artificial intelligence and provide free, open-source alternatives to overvalued Y Combinator startups. If you are excited about democratizing AI access & AI agents, please star ⭐️ this repository and use the link in the readme to join our open source AI research team.
best-practices-researcher
The most comprehensive Claude Code skills registry | Web Search: https://skills-registry-web.vercel.app
research_rules
Research & Verification Rules Quote Verification Protocol Primary Task "Make sure that the quote is relevant to the chapter and so you we want to make sure that we want to have it identifie
groundhog
398Groundhog's primary purpose is to teach people how Cursor and all these other coding agents work under the hood. If you understand how these coding assistants work from first principles, then you can drive these tools harder (or perhaps make your own!).
