Can We Trust Embodied Agents? Exploring Backdoor Attacks against Embodied LLM-based Decision-Making Systems
Ruochen Jiao*, Shaoyuan Xie*, Justin Yue, Takami Sato, Lixu Wang, Yixuan Wang, Qi Alfred Chen, Qi Zhu
ICLR 2025
First systematic study of backdoor vulnerabilities in embodied LLM agents. We show that adversarial triggers injected into environment observations can hijack LLM-based planners, causing unsafe actions in robotic and autonomous driving tasks.
SFT Doesn't Always Hurt General Capabilities: Revisiting Domain-Specific Fine-Tuning in LLMs
Jiacheng Lin, Zhongruo Wang, Kun Qian, Tian Wang, Arvind Srinivasan, Hansi Zeng, Ruochen Jiao, Xie Zhou, Jiri Gesi, Dakuo Wang, Yufan Guo, Kai Zhong, Weiqi Zhang, Sujay Sanghavi, Changyou Chen, Hyokun Yun, Lihong Li
ICLR 2026
Challenges the common belief that supervised fine-tuning degrades general capabilities. Through controlled experiments across diverse domains, we identify conditions under which SFT preserves or even improves general performance.
Kinematics-aware Trajectory Generation and Prediction with Latent Stochastic Differential Modeling
Ruochen Jiao*, Yixuan Wang*, Xiangguo Liu, Simon Zhan, Chao Huang, Qi Zhu
IROS 2024
A latent stochastic differential equation framework that incorporates vehicle kinematics constraints for generating physically plausible and diverse future trajectories in autonomous driving scenarios.
Semi-supervised Semantics-guided Adversarial Training for Robust Trajectory Prediction
Ruochen Jiao, Xiangguo Liu, Takami Sato, Alfred Chen, Qi Zhu
ICCV 2023
Proposes a semi-supervised adversarial training method that leverages semantic scene understanding to improve trajectory prediction robustness against distribution shifts and adversarial perturbations.
Enforcing Hard Constraints with Soft Barriers: Safety-driven Reinforcement Learning in Unknown Stochastic Environments
Yixuan Wang, Sinong Simon Zhan, Ruochen Jiao, Zhilu Wang, Wanxin Jin, Zhuoran Yang, Zhaoran Wang, Chao Huang, Qi Zhu
ICML 2023
Introduces soft barrier functions to enforce hard safety constraints in RL without requiring prior knowledge of environment dynamics, enabling provably safe exploration in stochastic settings.