Physical Agent Operation System (PhyAgentOS) is a self-evolving embodied AI framework based on Agentic workflows. Moving away from the "black-box" model of traditional "large models directly controlling hardware," PhyAgentOS pioneers a "Cognitive-Physical Decoupling" architectural paradigm. By constructing a Language-Action Interface, it completely decouples action representation from embodiment morphology, enabling standardized mapping from high-reasoning cloud models to edge physical execution layers.
PhyAgentOS utilizes a "State-as-a-File" protocol matrix, natively supporting zero-code migration across hardware platforms, sandbox-driven tool self-generation, and safety correction mechanisms based on Multi-Agent Critic verification.
- ๐ State-as-a-File: Software and hardware communicate by reading/writing local Markdown files (e.g.,
ENVIRONMENT.md,ACTION.md), ensuring complete decoupling and extreme transparency. - ๐ง Dual-Track Multi-Agent System:
- Track A (Cognitive Core): Includes Planner and Critic mechanisms. Large models do not issue commands directly; they must be verified by the Critic against the current robot's runtime
EMBODIED.md(copied from profiles) before being committed. - Track B (Physical Execution): An independent hardware watchdog (
hal_watchdog.py) monitors and executes commands. Supports both single-instance mode and Fleet mode for multi-robot coordination.
- Track A (Cognitive Core): Includes Planner and Critic mechanisms. Large models do not issue commands directly; they must be verified by the Critic against the current robot's runtime
- ๐ Dynamic Plugin Mechanism: Supports dynamic loading of external hardware drivers via
hal/drivers/, allowing for new hardware support without modifying core code. - ๐ก๏ธ Safety Correction Mechanism: Strict action verification and
LESSONS.mdexperience library prevent Agent workflows from going out of control. - ๐ฎ Simulation Loop: Built-in lightweight simulation support allows verification of the full chain from natural language instructions to physical state changes without real hardware.
- ๐บ๏ธ Semantic Navigation & Perception: Built-in
SemanticNavigationToolandPerceptionServicesupport resolving high-level semantic goals into physical coordinates and constructing scene graphs by fusing geometric and semantic information.
PhyAgentOS quickly connects to XLeRobot, automatically checks the robot's status, and after confirming safety, performs basic chassis movement and dual-arm motion.
PhyAgentOS achieves realtime dialog and natural language-driven pick&up task through ReKep (Franka Research 3).
PhyAgentOS's core is a local workspace where software and hardware operate as independent daemons reading/writing files:
git clone https://github.com/PhyAgentOS/PhyAgentOS.git
cd PhyAgentOS
pip install -e .
# Install simulation dependencies (e.g., watchdog)
pip install watchdog
# Optional: Install external ReKep real-world plugin
python scripts/deploy_rekep_real_plugin.py \
--repo-url https://github.com/baiyu858/PhyAgentOS-rekep-real-plugin.gitpaos onboardThis generates core Markdown protocol files in the current workspace. Single-instance mode defaults to ~/.PhyAgentOS/workspace/; Fleet mode uses a shared workspace and multiple robot workspaces under ~/.PhyAgentOS/workspaces/.
Open two terminals:
Terminal 1: Start Hardware Watchdog & Simulation (Track B)
python hal/hal_watchdog.pyTo pass driver-specific runtime configuration without specializing the watchdog CLI, use:
python hal/hal_watchdog.py --driver <driver_name> --driver-config path/to/driver.jsonThe config file must be a JSON object and its keys are passed through to the selected driver constructor unchanged.
To use real-world ReKep instead of simulation, install the plugin and run:
python hal/hal_watchdog.py --driver rekep_realTerminal 2: Start Brain Agent (Track A)
paos agentIn the paos agent CLI, input:
"Look at what is on the table, then grasp that apple for me."
You will see the action execution in the simulation logs in Terminal 1, and receive completion confirmation from the Agent in Terminal 2.
- More Information
User Manual: Operations guide for end users, integrators, and demo operators.
Development Guide: Layered guide for secondary developers, hardware integrators, plugin authors, and maintainers.
Physical Agent Operating System/
โโโ PhyAgentOS/ # Track A: Software Brain Core
โ โโโ agent/ # Agent Logic (Planner, Critic)
โ โโโ templates/ # Workspace Markdown Templates
โ โโโ ...
โโโ hal/ # Track B: Hardware HAL & Simulation
โ โโโ hal_watchdog.py # Hardware Watchdog Daemon
โ โโโ simulation/ # Simulation Environment Code
โโโ scripts/ # External HAL Plugin Deployment
โ โโโ deploy_rekep_real_plugin.py
โโโ workspace/ # Single-instance Runtime Workspace
โ โโโ EMBODIED.md # Runtime Robot Profile
โ โโโ ENVIRONMENT.md # Current Scene-Graph
โ โโโ ACTION.md # Pending Action Commands
โ โโโ LESSONS.md # Failure Experience Records
โ โโโ SKILL.md # Successful Workflow SOP
โโโ workspaces/ # Fleet Topology
โ โโโ shared/ # Agent Workspace & Global ENVIRONMENT.md
โ โโโ go2_edu_001/ # Robot-local ACTION.md / EMBODIED.md
โ โโโ ...
โโโ docs/ # Project Documentation
โ โโโ PLAN.md # Detailed Implementation Plan
โ โโโ PROJ.md # Project Whitepaper & Architecture
โโโ README.md # English Documentation
โโโ README_zh.md # Chinese Documentation
- Phase 1: Desktop Loop & Markdown Protocol Establishment.
- v0.0.1: Framework Design & Initialization
- v0.0.2: Embodied Skill Plugin Deployment & Invocation Design
- v0.0.3: Visual Decoupling + Grasping Pipeline (SAM3 & ReKep)
- v0.0.4: Atomic Action-based VLN Pipeline (SAM3)
- v0.0.5: Multi-Agent Protocol Design
- v0.0.6: Long-horizon Task Decomposition, Orchestration & Execution
- v0.0.7: IoT Device Integration (e.g., XiaoZhi)
- Phase 2: Multi-Embodiment Coordination & Multi-modal Memory.
- Phase 3: Constraint Solving & High-level Heterogeneous Coordination.
PhyAgentOS supports various embodiment types through the HAL (Hardware Abstraction Layer) protocol.
| Embodiment Type | Robot | Status | Remarks |
|---|---|---|---|
| Desktop Robot Arm | AgileX PIPER | ๐ข Verified | Full-chain verified with ReKep & SAM3 |
| Composite Robot | AgileX PIPER + Unitree Go2 | ๐ก Partial | locomotion adaptation in progress |
| Desktop Robot Arm | Dobot Nova 2 | ๐ข Verified | ReKep deployment verified |
| Quadruped Robot | Unitree Go2 | ๐ก Partial | Currently supports mobility and semantic navigation |
| Dual-Arm Control | XLeRobot | ๐ข Verified | Deployment and dual-arm capture have now been achieved |
| IoT Device | XiaoZhi (ESP32) | ๐ก Partial | Currently supports voice dialogue interaction |
| Industrial Robot | Franka Research 3 | ๐ข Verified | Visual Reasoning and capture have been achieved |
| Edu Robot | Hiwonder Series | ๐ด Unsupported | Awaiting driver plugin development |
| General Environment | Built-in Simulator | ๐ข Verified | Lightweight simulation based on disk mapping |
Note: PhyAgentOS is designed with a plugin architecture. Any hardware that supports a Python control interface can be quickly integrated via
hal/drivers/. A community plugin template is available atdocs/user_development_guide/PLUGIN_DEVELOPMENT_GUIDE.md, with the Chinese version atdocs/user_development_guide/PLUGIN_DEVELOPMENT_GUIDE_zh.md.
PRs and Issues are welcome! Please refer to docs/user_development_guide/README.md for detailed architecture design and development guidelines.
Special Thanks: This project is developed based on nanobot, thanks for providing the lightweight Agent framework. Everyone is welcome to go to the nanobot repository and give it a star!
We welcome any individual or team to join as a contributor๏ผ








