Current humanoid robot control paradigms place the burden of feasibility assessment on human operators, who must carefully design commands within perceived robot limitations. This constraint significantly hinders practical deployment and limits the expressiveness of robot behaviors. This study proposed an inverting paradigm: rather than constraining operator inputs, robots should autonomously evaluate their capacity to execute commanded motions and intelligently adapt references to align with their physical constraints and learned skills. This study introduced the Performance Prediction Network (PPN), a transformer-based architecture that forecasts execution quality for arbitrary reference trajectories by analyzing both the commanded motion sequence and current robot state. Given a high-level task specification, our framework synthesizes multiple viable motion candidates and employs PPN to rank them across six dimensions: collision avoidance, kinematic feasibility, dynamic stability, trajectory smoothness, and goal satisfaction. This ranking enables autonomous selection of the most suitable reference motion before execution begins. Our complete system integrates motion generation, kinematic retargeting, and learned control policies with PPN-guided adaptation, creating a closed-loop framework where robots reason about their own limitations. Validated on 100,000 diverse human motions span walking, running, jumping, and acrobatic maneuvers, PPN achieves 99.14% accuracy in predicting imminent failures while maintaining low prediction error across all performance metrics. In deployment, our system successfully prevents 62% of anticipated falls by autonomously modifying commanded references, demonstrating that explicit capability modeling enables safer and more reliable humanoid control without sacrificing behavioral diversity.
Dautenhahn K, 2007, Socially Intelligent Robots: Dimensions of Human–Robot Interaction. Philosophical Transactions of the Royal Society B, 362: 679–704.
Thomaz A, Cakmak M, 2016, Learning About Objects with Human Teachers. Human–Robot Interaction, 5: 1–42.
Khatib O, Sentis L, Park J, et al., 2008, Whole-Body Dynamic Behavior and Control of Human-Like Robots. International Journal of Humanoid Robotics, 5: 29–43.
Mainprice J, Hayne R, Berenson D, 2015, Predicting Human Reaching Motion in Collaborative Tasks Using Inverse Optimal Control and Iterative Replanning. IEEE International Conference on Robotics and Automation: 885–892.
Lasota P, Fong T, Shah J, 2017, A Survey of Methods for Safe Human–Robot Interaction. Foundations and Trends in Robotics, 5: 261–349.
Javdani S, Admoni H, Pellegrinelli S, et al., 2018, Shared Autonomy via Hindsight Optimization for Teleoperation and Teaming. International Journal of Robotics Research, 37: 717–742.
Dragan A, Lee K, Srinivasa S, 2013, Legibility and Predictability of Robot Motion. IEEE International Conference on Human–Robot Interaction: 301–308.
Peng X, Abbeel P, Levine S, et al., 2018, DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills. ACM Transactions on Graphics, 37: 1–14.
Atkeson C, Schaal S, 1997, Robot Learning from Demonstration. International Conference on Machine Learning: 12–20.
Hwangbo J, Lee J, Dosovitskiy A, et al., 2019, Learning Agile and Dynamic Motor Skills for Legged Robots. Science Robotics, 4: eaau5872.
Mordatch I, Todorov E, Popović Z, 2012, Discovery of Complex Behaviors through Contact-Invariant Optimization. ACM Transactions on Graphics, 31: 1–8.
Pratt J, Carff J, Drakunov S, et al., 2006, Capture Point: A Step Toward Humanoid Push Recovery. IEEE-RAS International Conference on Humanoid Robots: 200–207.
Stephens B, Atkeson C, 2010, Push Recovery by Stepping for Humanoid Robots with Force-Controlled Joints. IEEE-RAS International Conference on Humanoid Robots: 52–59.
Koolen T, de Boer T, Rebula J, et al., 2012, Capturability-Based Analysis and Control of Legged Locomotion, Part 1: Theory and Application. International Journal of Robotics Research, 31: 1094–1113.
Bretl T, Lall S, 2008, Testing Static Equilibrium for Legged Robots. IEEE Transactions on Robotics, 24: 794–807.
Hauser K, Bretl T, Latombe J, et al., 2008, Motion Planning for Legged Robots on Varied Terrain. International Journal of Robotics Research, 27: 1325–1349.
Dai H, Valenzuela A, Tedrake R, 2014, Whole-Body Motion Planning with Centroidal Dynamics and Full Kinematics. IEEE-RAS International Conference on Humanoid Robots: 295–302.
Zucker M, Ratliff N, Dragan A, et al., 2013, CHOMP: Covariant Hamiltonian Optimization for Motion Planning. International Journal of Robotics Research, 32: 1164–1193.
Kalakrishnan M, Chitta S, Theodorou E, et al., 2011, STOMP: Stochastic Trajectory Optimization for Motion Planning. IEEE International Conference on Robotics and Automation: 4569–4574.
Tedrake R, Manchester I, Tobenkin M, et al., 2010, LQR-Trees: Feedback Motion Planning via Sums-of-Squares Verification. International Journal of Robotics Research, 29: 1038–1052.
Ott C, Roa M, Hirzinger G, 2008, Posture and Balance Control for Biped Robots Based on Contact Force Optimization. IEEE-RAS International Conference on Humanoid Robots: 26–33.
Sentis L, Khatib O, 2005, Synthesis of Whole-Body Behaviors through Hierarchical Control of Behavioral Primitives. International Journal of Humanoid Robotics, 2: 505–518.
Peng X, Guo Y, Halper L, et al., 2022, ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters. ACM Transactions on Graphics, 41: 1–17.
Rempe D, Birdal T, Zhao Y, et al., 2021, HuMoR: 3D Human Motion Model for Robust Pose Estimation. IEEE International Conference on Computer Vision: 11488–11499.
Schaal S, 1999, Is Imitation Learning the Route to Humanoid Robots? Trends in Cognitive Sciences, 3: 233–242.
Argall B, Chernova S, Veloso M, et al., 2009, A Survey of Robot Learning from Demonstration. Robotics and Autonomous Systems, 57: 469–483.
Abbeel P, Ng A, 2004, Apprenticeship Learning via Inverse Reinforcement Learning. International Conference on Machine Learning: 1–8.
Pomerleau D, 1988, ALVINN: An Autonomous Land Vehicle in a Neural Network. Advances in Neural Information Processing Systems, 1: 305–313.
Ho J, Ermon S, 2016, Generative Adversarial Imitation Learning. Advances in Neural Information Processing Systems, 29: 4565–4573.
Ross S, Gordon G, Bagnell J, 2011, A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning. International Conference on Artificial Intelligence and Statistics, 15: 627–635.
Mehta S, Ciftci Y, Ramachandran B, et al., 2024, Stable-BC: Controlling Covariate Shift with Stable Behavior Cloning. arXiv preprint arXiv:2408.06246.
Park J, Kim Y, Song K, et al., 2024, Mitigating Covariate Shift in Behavioral Cloning via Robust Stationary Distribution Correction. Advances in Neural Information Processing Systems, 37.
Ng A, Russell S, 2000, Algorithms for Inverse Reinforcement Learning. International Conference on Machine Learning: 663–670.
Ziebart B, Maas A, Bagnell J, et al., 2008, Maximum Entropy Inverse Reinforcement Learning. AAAI Conference on Artificial Intelligence: 1433–1438.
Wulfmeier M, Ondruska P, Posner I, 2015, Maximum Entropy Deep Inverse Reinforcement Learning. arXiv preprint arXiv:1507.04888.
Finn C, Levine S, Abbeel P, 2016, Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization. International Conference on Machine Learning, 48: 49–58.
Kajita S, Kanehiro F, Kaneko K, et al., 2003, Biped Walking Pattern Generation Using Preview Control of Zero-Moment Point. IEEE International Conference on Robotics and Automation, 3: 1620–1626.
Vukobratovic M, Borovac B, 2004, Zero-Moment Point—Thirty-Five Years of Its Life. International Journal of Humanoid Robotics, 1: 157–173.
Kuffner J, Kagami S, Nishiwaki K, et al., 2002, Dynamically Stable Motion Planning for Humanoid Robots. Autonomous Robots, 12: 105–118.
Kuffner J, LaValle S, 2000, RRT-Connect: An Efficient Approach to Single-Query Path Planning. IEEE International Conference on Robotics and Automation: 995–1001.
Li Z, Cheng X, Peng X, et al., 2021, Reinforcement Learning for Robust Parameterized Locomotion Control of Bipedal Robots. IEEE International Conference on Robotics and Automation: 2811–2817.
Lee J, Hwangbo J, Wellhausen L, et al., 2020, Learning Quadrupedal Locomotion over Challenging Terrain. Science Robotics, 5: eabc5986.
Xie Z, Clary P, Dao J, et al., 2020, Learning Locomotion Skills for Cassie: Iterative Design and Sim-to-Real. Conference on Robot Learning, 100: 317–329.
Radosavovic I, Xiao T, Zhang B, et al., 2024, Real-World Humanoid Locomotion with Reinforcement Learning. Science Robotics, 9: edi9579.
Renner R, Behnke S, 2006, Instability Detection and Fall Avoidance for a Humanoid Using Attitude Sensors and Reflexes. IEEE/RSJ International Conference on Intelligent Robots and Systems: 2967–2973.
Zhong S, Gao J, Li M, et al., 2025, Fall Analysis and Prediction for Humanoids. Robotics and Autonomous Systems, 185: 104995.
Yang T, Zhang W, Yu Z, et al., 2018, Falling Prediction and Recovery Control for a Humanoid Robot. IEEE-RAS International Conference on Humanoid Robots: 481–487.
Mastalli C, Merkt W, Xin G, et al., 2024, Know Your Limits! Optimize the Robot’s Behavior through Self-Awareness. arXiv preprint arXiv:2409.10308.
Cheng X, Shi K, Agarwal A, et al., 2024, Extreme Parkour with Legged Robots. arXiv preprint arXiv:2309.14341.
Khamassi M, Lallée S, Enel P, et al., 2018, Toward Self-Aware Robots. Frontiers in Robotics and AI, 5: 88.