Skip to content

Latest commit

 

History

History
91 lines (66 loc) · 3.76 KB

README.md

File metadata and controls

91 lines (66 loc) · 3.76 KB

Yanzhe Zhang, Tao Yu, Diyi Yang

Overview

Autonomous agents powered by large vision and language models (VLM) have demonstrated significant potential in completing daily computer tasks, such as browsing the web to book travel and operating desktop software, which requires agents to understand these interfaces. Despite such visual inputs becoming more integrated into agentic applications, what types of risks and attacks exist around them still remain unclear. In this work, we demonstrate that VLM agents can be easily attacked by a set of carefully designed adversarial pop-ups, which human users would typically recognize and ignore. This distraction leads agents to click these pop-ups instead of performing the tasks as usual. Integrating these pop-ups into existing agent testing environments like OSWorld and VisualWebArena leads to an attack success rate (the frequency of the agent clicking the pop-ups) of 86% on average and decreases the task success rate by 47%. Basic defense techniques such as asking the agent to ignore pop-ups or including an advertisement notice, are ineffective against the attack.

Intro of ADVERSARIAL POP-UPS :)

Examples

Examples of ADVERSARIAL POP-UPS :)

Set Up

Please follow the setup instructions of OSWorld and VisualWebArena. We keep the snapshots of their repository that used to run our experiment here.

Data

For OSWorld, the 50-task test set we are using is here. For 72-task test set we are using is here.

Code Structure

We outline the files that we made significant changes.

|_OSWorld/
      |_attack.py -> main logic of drawing adversaril pop-ups
      |_general_attack_utils.py -> drawing utils
      |_image_processing.py -> drawing utils
      |_mm_agents/
            |_agent.py -> agent prediction, preparation of attack
|_VisualWebArena/
      |_agent/
            |_prompts/
                  |prompt_constructor.py
      |_browswer_env/
            |_processors.py -> agent prediction, preparation of attack
      |_run.py

Example Command

Similar to the original benchmark, we provide example scripts:

cd OSWorld
bash ./task_run.sh
cd VisualWebArena
bash ./task_run.sh

The major change is that you need to provide attack configs, where we have examples for OSWorld and VisualWebArena:

{
    "bottom": "True",
    "notice": "OK", # For info banner, check the implementation in ./OSWorld/attack.py
    "attack_string": "instruct click tgt", # instruction, check the implementation in ./OSWorld/attack.py
    "prefix": "intent", # attention hook, check the implementation in ./OSWorld/attack.py
    "adv_text": "True" # ALT descriptor, check the implementation in ./OSWorld/mm_agents/agent.py
}

Acknowledgement

We thank OSWorld, VisualWebArena, and Adversarial Attacks on Multimodal Agents for their awsome work!

If you find our work helpful, please consider citing our paper:

@misc{zhang2024attacking,
    title={Attacking Vision-Language Computer Agents via Pop-ups},
    author={Yanzhe Zhang and Tao Yu and Diyi Yang},
    year={2024},
    eprint={2411.02391},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}

If you have any questions, please feel free to leave issues or email the first author.