# RoboOS
**Repository Path**: nilbody_0/RoboOS
## Basic Information
- **Project Name**: RoboOS
- **Description**: No description available
- **Primary Language**: Unknown
- **License**: Apache-2.0
- **Default Branch**: main
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 0
- **Forks**: 0
- **Created**: 2026-03-18
- **Last Updated**: 2026-03-18
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
# RoboOS
RoboOS: A Hierarchical Embodied Framework for Cross-Embodiment and Multi-Agent Collaboration
  ⭐️ Project   │   🌎 Demo   │   📑 Technical Report  
  🤖 RoboBrain 2.0: Advanced version of RoboBrain. See Better. Think Harder. Do Smarter.
  🤖 RoboBrain 1.0: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete.
Please visit the **[stone-alone](https://github.com/FlagOpen/RoboOS/tree/stand-alone)** branch for **RoboOS version 2.0**, which supports **lightweight single-machine deployment!**
💬 If you have any questions, feel free to contact us via WeChat.
## 🔥 Overview
The rise of embodied intelligence has intensified the need for robust multi-agent collaboration in industrial automation, service robotics, and smart manufacturing. However, current robotic systems struggle with critical limitations, including poor cross-embodiment adaptability, inefficient task scheduling, and inadequate dynamic error correction. While end-to-end vision-language-action (VLA) models (e.g., OpenVLA, RDT, Pi-0) exhibit weak long-horizon planning and task generalization, hierarchical VLA models (e.g., Helix, Gemini-Robotics, GR00T-N1) lack cross-embodiment compatibility and multi-agent coordination capabilities.
To address these challenges, we present **RoboOS**, the first open-source embodied operating system based on a *Brain-Cerebellum* hierarchical architecture, facilitating a paradigm shift from single-agent to swarm intelligence. Specifically, RoboOS comprises three key components: **(1) the Embodied Cloud Model**, a multimodal large language model (MLLM) for global perception and high-level decision-making; **(2) the Cerebellum Skill Library**, a modular, plug-and-play toolkit for seamless multi-skill execution; and **(3) Real-Time Shared Memory**, a spatiotemporal synchronization mechanism for multi-agent state coordination. By integrating hierarchical information flow, RoboOS bridges the Embodied Brain and Cerebellum Skill Library, enabling robust planning, scheduling, and error correction for long-horizon tasks while ensuring efficient multi-agent collaboration by Real-Time Shared Memory. Moreover, we optimize edge-cloud communication and cloud-based distributed inference to support high-frequency interactions and scalable deployment.
Extensive real-world experiments across diverse scenarios (e.g., restaurant, household, supermarket) demonstrate RoboOS’s versatility, supporting heterogeneous embodiments (single-arm, dual-arm, humanoid, wheeled), which provides a scalable and practical solution for cross-embodiment collaboration, pushing the boundaries of embodied intelligence.
### Structure for RoboOS 2.0 (SaaS + MCP)
### Structure for RoboOS 1.0
## 🎯 RoadMap
- [x] Release **RoboOS-1.0** version
- [x] Release **[Technical Report](https://arxiv.org/abs/2505.03673)** of RoboOS.
- [ ] Release **RoboOS-2.0** version *(by the end of this month)*
- [ ] Release friendly and detailed **User Guide Manual**.
- [ ] Release more comprehensive multi-agent collaboration **DEMOs** based on RoboOS.
## ⭐️ Full Guide Manual (RoboOS-2.0)
*Coming soon ...*
Due to the substantial code refactoring and engineering efforts required, we'll need to postpone the release by *several days*. We appreciate your patience as we ensure the highest quality standards.
## ⭐️ Simple Guide Manual (Only for RoboOS-1.0)
### 1. Prerequisites
- Python 3.8+
- Redis server
- pip package manager
### 2. Installation
```bash
# Clone the repository
git clone https://github.com/FlagOpen/RoboOS.git
cd RoboOS
# Install dependencies
pip install -r requirements.txt
```
### 3. Quick Start
```bash
# 1. Start Redis
redis-server
# 2. Start Master
python master/run.py
# 3. Start Slaver (for multi-agent, your should run at different robots respectively)
python slaver/run.py
# 4. Launch Web Interface
python gradio_ui.py
# Then, access the web interface at: http://localhost:7861
```
## ✨ Example Demo
### 🔍 Master Console
### 🤖 Slaver Console
#### Subtask_1 for Realman Single-ARM Robot
#### Subtask_2 for Agilex Dual-ARM Robot
#### Subtask_3 for Realman Single-ARM Robot
## 📑 Citation
If you find this project useful, welcome to cite us.
```bib
@article{tan2025roboos,
title={RoboOS: A Hierarchical Embodied Framework for Cross-Embodiment and Multi-Agent Collaboration},
author={Tan, Huajie and Hao, Xiaoshuai and Lin, Minglan and Wang, Pengwei and Lyu, Yaoxu and Cao, Mingyu and Wang, Zhongyuan and Zhang, Shanghang},
journal={arXiv preprint arXiv:2505.03673},
year={2025}
}
@article{ji2025robobrain,
title={RoboBrain: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete},
author={Ji, Yuheng and Tan, Huajie and Shi, Jiayu and Hao, Xiaoshuai and Zhang, Yuan and Zhang, Hengyuan and Wang, Pengwei and Zhao, Mengdi and Mu, Yao and An, Pengju and others},
journal={arXiv preprint arXiv:2502.21257},
year={2025}
}
```