Photo taken by my wife! Bookmark this to keep an eye on my project updates!
IEEE Transactions on Robotics, 2025 · vol. 42, pp. 152–169
Yang, X., Wei M., Ji Z., Lai Y.-K.
Manipulating granular materials such as sand and soil with precision is extremely difficult — their physical properties are unknown in advance and they behave in complex, unpredictable ways. DDBot addresses this by combining a GPU-accelerated differentiable physics simulator tailored for granular materials with gradient-based optimisation. The robot first identifies unknown material properties (system identification) and then optimises its digging strategy — all without prior knowledge of the material — achieving high-precision results when deployed directly in the real world (zero-shot transfer).
Key contributions:
📄 Preprint · 📰 Published Paper · ▶ Video Demo
International Journal of Robotics Research (IJRR), 2025
Yang, X., Ji Z., Lai Y.-K.
Materials like clay and dough deform permanently when manipulated — making them notoriously hard to simulate or control. This work introduces DPSI (Differentiable Physics-based System Identification), a framework that lets a robot arm estimate the physical properties of such materials (Young’s modulus, Poisson’s ratio, yield stress, friction) from just a single real-world interaction and incomplete 3D point clouds. The estimated parameters are physically interpretable — unlike neural network “black-box” approaches — and enable accurate simulation of long-horizon deformation behaviours.
Key contributions:
📄 Preprint · 📰 Published Paper · ▶ Video Demo · 🌐 Project Page & Code
IEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2025
Wei M., Yang X., Lai Y.-K., Tafrishi S.A., Ji Z.
Standard robot learning methods struggle with granular materials (e.g., grains, powder) because collecting training data is expensive and their physical properties are complex. AutomaChef bypasses costly data collection by using a differentiable physics-based simulator to automatically generate expert demonstrations via gradient-based optimisation, which then guide policy learning. The resulting policies successfully transfer to real-world granular transport tasks.
Key contributions:
📄 Preprint · 📰 Published Paper
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2025
Wei M., Yang X., Yan J., Lai Y.-K., Ji Z.
Optimising robot skills for granular material manipulation (e.g., soil excavation and levelling) using differentiable simulation suffers from instability due to the highly nonlinear dynamics involved. Celebi addresses this by incorporating causal inference into the optimisation loop: it models causal relationships between task-relevant features (extracted from point cloud observations) and skill parameters, and uses these to adaptively adjust gradient update step sizes — improving both stability and convergence efficiency.
Key contributions:
Neurocomputing, 2024 · vol. 578, 127386
Yang, X., Wu J., Lai Y.-K., Ji Z.
Disentangling objects that are in contact with each other — such as separating tangled cables or nested containers — requires a robot to understand fine-grained spatial affordances (where and how to grasp) and execute contact-rich motions. GAM proposes a general affordance-based manipulation framework that learns to predict these affordances and execute disentangling strategies for diverse object configurations.
Key contributions:
27th International Conference on Automation and Computing (ICAC), 2022
Yang, X., Ji Z., Wu J., Lai Y.-K.
Long-horizon robotic manipulation tasks with sparse rewards (e.g., stacking several blocks where only final success is rewarded) are extremely hard for standard deep RL. This paper proposes A²E, a technique that combines two human-inspired ideas: decomposing complex tasks into ordered sub-tasks and providing abstract demonstrations of the correct sub-task sequence, while training the agent to explore more freely on poorly learnt sub-tasks and more deterministically on well-mastered ones.
Key contributions:
IEEE Transactions on Automation Science and Engineering (TASE), 2023 · vol. 21, no. 4, pp. 5457–5472 · 21 citations
Gao Y., Wu J., Yang X., Ji Z.
Mapless robot navigation from raw high-dimensional sensor data (LiDAR, cameras) is expensive to learn end-to-end. This work proposes an efficient hierarchical RL framework using a novel sub-goal scoring model — Predictive Neighbouring Space Scoring (PNSS) — which estimates how explorable nearby positions are from the current observation. PNSS generates compact, meaningful sub-goals that guide a low-level navigation policy, dramatically reducing learning complexity.
Key contributions:
IEEE Transactions on Cognitive and Developmental Systems (TCDS), 2023 · vol. 15, no. 3, pp. 1139–1149
Yang, X., Ji Z., Wu J., Lai Y.-K.
A survey and analysis of deep robotic affordance learning — the study of how robots learn where and how to act on objects. This review classifies methods from a reinforcement learning perspective, discusses technical details and limitations, and proposes a promising future direction: RL-based affordance definitions that predict the consequences of arbitrary actions.
Key contributions:
📄 Preprint · 📰 Published Paper
IEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2021 · vol. 33, no. 9, pp. 4727–4741 · 133 citations
Yang, X., Ji Z., Wu J., Lai Y.-K., Wei C., Liu G., Setchi R.
Teaching robots to perform long, multi-step manipulation tasks (e.g., pick, stack, insert in sequence) with sparse rewards is a fundamental challenge. This work proposes the Universal Option Framework (UOF), a hierarchical RL architecture where a high-level policy decomposes complex tasks into sub-goals and low-level universal policies are reused across sub-tasks, enabling efficient learning of long-horizon manipulation.
Key contributions:
📰 Published Paper · ▶ Video Demo · 💻 GitHub
Towards Autonomous Robotic Systems (TAROS), 2021
Yang, X., Ji Z., Wu J., Lai Y.-K.
A re-implementation of OpenAI Gym’s robotic manipulation benchmarks on the free, open-source PyBullet physics engine (replacing the commercial MuJoCo engine). Includes new APIs for joint control, image observations, and a built-in on-hand camera, plus a suite of novel multi-step, long-horizon, sparse-reward tasks designed to challenge goal-conditioned RL algorithms.
Key contributions:
For a full publication list, visit my Google Scholar profile. Last updated: March 2026.