beat_demo-1.mp4
BEAT is the first to show visual backdoors in MLLM embodied agents: fine-tune the MLLM to implant a backdoor so the agent behaves normally until a specific object trigger, then follows an attacker-specified policy.
beat_method_demo-1.mp4
BEAT uses a two-stage training pipeline: (i) standard supervised fine-tuning (SFT) on a mixture of benign and backdoor trajectories to strengthen general capabilities, followed by (ii) our Contrastive Trigger Learning (CTL), a preference-learning procedure that improves the precision of backdoor activation.
conda create -n beat python=3.10
conda activate beat
pip install -r requirements.txtWe provide example fine-tuning data for SFT and CTL in ./data. Each SFT example consists of an input (history plus image) and the MLLM’s target output. Each CTL example is a contrastive pair identical except for trigger presence in the image and the associated target output. Due to ethical considerations, the full training set is available upon request.
We prepare the scripts of running BEAT finetuning over the on the example dataset:
bash scripts/qwen2_sft_ctl.sh
bash scripts/internvl_sft_ctl.shTo run on other model, you need to customized the llm finetuning interface in src/llms.
@article{zhan2025visual,
title={Visual Backdoor Attacks on MLLM Embodied Decision Making via Contrastive Trigger Learning},
author={Zhan, Qiusi and Ha, Hyeonjeong and Yang, Rui and Xu, Sirui and Chen, Hanyang and Gui, Liang-Yan and Wang, Yu-Xiong and Zhang, Huan and Ji, Heng and Kang, Daniel},
journal={arXiv preprint arXiv:2510.27623},
year={2025}
}