Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2025 Dec 23:8:1047.
doi: 10.34133/research.1047. eCollection 2025.

Orchestrating Embodied Systems through the Embodied Context Protocol: Motivation, Progress, and Directions

Affiliations
Review

Orchestrating Embodied Systems through the Embodied Context Protocol: Motivation, Progress, and Directions

Fuyu Ma et al. Research (Wash D C). .

Abstract

The emergence of embodied intelligence has brought a fundamental shift to the robotics field, emphasizing the integration of perception, cognition, and control in dynamic physical environments. Although substantial progress has been made in artificial intelligence models, control middleware, and industrial communication protocols within their respective domains, the fragmentation in semantic interaction and task-level coordination still limits the scalability and deployment of embodied intelligence systems. This review synthesizes the current research on system deployment and coordination in embodied intelligence, particularly focusing on the challenges in achieving task-level coordination and semantic interoperability across heterogeneous components. We examine key coordination requirements, such as context semantics, capability declaration, and workflow composition, and highlight the existing gaps in addressing these issues within current systems. In response, we propose the Embodied Context Protocol (ECP) as an emerging solution, designed to bridge these gaps and enhance interoperability across various subsystems. It then presents the design philosophy, interface specification, and execution workflow of ECP, followed by its current implementation progress validated through practical deployments, while also highlighting the future directions and unresolved challenges that will shape its standardization and large-scale adoption. As an interface protocol, ECP aims to evolve into a standardized interoperability ecosystem for embodied intelligence and industrial automation. Realizing this vision will require collaboration across academic and industrial communities to jointly advance the development, adoption, and standardization of ECP.

PubMed Disclaimer

Conflict of interest statement

Competing interests: The authors declare that they have no competing interests.

Figures

Fig. 1.
Fig. 1.
Coordination limitations occur when robot operating systems, industrial protocols, and simulation platforms are applied to embodied tasks. OPC UA, Open Platform Communications Unified Architecture.
Fig. 2.
Fig. 2.
Task workflow of a typical embodied intelligence system, illustrating staged interactions across task definition, data acquisition (physical and simulated), policy model training, and robotic task execution.
Fig. 3.
Fig. 3.
This architecture of an embodied intelligence system illustrates how the Embodied Context Protocol (ECP) is embedded within practical task workflows by mediating coordination among functional modules. The architecture is derived from an actual deployment scenario developed in our laboratory. IDE, integrated development environment.
Fig. 4.
Fig. 4.
Real-world deployment scenarios of embodied task workflows, including (A and B) a humanoid robotic system performing multiview perception and picking operations (type: AgiBot G1), (C) a dual-arm platform (type: AgileX Robotics COBOT Magic), (D) local fog servers hosting the Action Chunking Transformer (ACT)-based policy model, (E) an industrial control cabinet integrating programmable logic controllers (PLCs) and sensor interfaces to decompose high-level tasks into atomic operations, (F) the data acquisition interface in the simulation environment, (G) the model training interface developed in our lab, and (H) model inference executed on the humanoid robot.

References

    1. Tong Y, Liu H, Zhang Z. Advancements in humanoid robots: A comprehensive review and future prospects. IEEE/CAA J Autom Sin. 2024;11(2):301–328.
    1. Shen T, Sun J, Kong S, Wang Y, Li J, Li X, Wang FY. The journey/DAO/TAO of embodied intelligence: From large models to foundation intelligence and parallel intelligence. IEEE/CAA J Autom Sin. 2024;11(6):1313–1316.
    1. Dai N, Zhang K, Zhang F, Li J, Zhong J, Huang Y, Ding H. AI-assisted flexible electronics in humanoid robot heads for natural and authentic facial expressions. Int J Hydrogen Energ. 2025;6(2): Article 100752. - PMC - PubMed
    1. Cheng N, Xu J, Guan C, Gao J, Wang W, Li Y, Meng F, Zhou J, Fang B, Han W. Touch100k: A large-scale touch-language-vision dataset for touch-centric multimodal representation. Inf Fusion. 2025;124: Article 103305.
    1. Zhao W, Yuan Y. Development of intelligent robots in the wave of embodied intelligence. Natl Sci Rev. 2025;12(7): Article nwaf159. - PMC - PubMed

LinkOut - more resources