SeeAct is a system for generalist web agents that autonomously carry out tasks on any given website, with a focus on large multimodal models (LMMs) such as GPT-4V(ision). It consists of two main components: (1) A robust codebase that supports running web agents on live websites, and (2) an innovative framework that utilizes LMMs as generalist web agents.
Website • Paper • Dataset • Twitter
- 2024/3/18: Multimodal-Mind2Web dataset released. We have paired each HTML document with the corresponding webpage screenshot image and saved the trouble of downloading Mind2Web Raw Dump.
The SeeAct tool enables running web agents on live websites through PlayWright, serving as an interface between an agent and a web browser. It efficiently tunnels inputs from the browser to the agent, and translates predicted actions of the agent into browser events for execution. This tool can be used for running web agent demos and evaluating their performance on live websites.
- Create a conda environment and install dependency:
conda create -n seeact python=3.10
conda activate seeact
pip install -r requirements.txt
- Set up PlayWright and install the browser kernels.
playwright install
Please fill in the OpenAI API Key in the configuration file at src/config/demo_mode.toml
before running SeeAct.
Your API key is available through your OpenAI account page.
Note that the key is only stored locally and will NOT be shared anywhere.
In the demo mode, SeeAct takes task
and website
from user terminal input. Run SeeAct in demo mode with the following command:
cd src
python seeact.py
Demo mode will use the default configuration file at src/config/demo_mode.toml
.
SeeAct is configurable through TOML files in src/config/
.
These files enable you to customize various aspects of the system's behavior
via the following parameters:
is_demo
: Settrue
to allow task and website from user terminal input, setfalse
to run tasks and websites from a JSON file (useful for batch evaluation).default_task
anddefault_website
: Default task and website used in the demo mode.max_op
: Maximum number of actions the agent can take for a task.api_key
: OpenAI API key.save_file_dir
: Directory path to save output results, including terminal logs and screenshot images.
After starting SeeAct, you'll be required to enter a task description
or you can press Enter
to use the default task of finding our paper on arXiv.
Next, you need to input the website URL
(please ensure it includes all necessary prefixes (https, www))
or you can press Enter
to use the default Google homepage (https://www.google.com/).
You can also automatically run SeeAct on a list of tasks and websites in a JSON file. Run SeeAct with the following command:
cd src
python seeact.py -c config/auto_mode.toml
In the configuration file, task_file_path
defines the path of the JSON file.
It is default to ../data/online_tasks/sample_tasks.json
, which contains a variety of task examples.
For custom scenarios, modify the configuration files to adapt the tool to your specific requirements. This includes setting up custom tasks, adjusting experiment parameters, and configuring Playwright options for more precise control over the web browsing experience.
The current version is research/experimental in nature and by no means perfect. Please always be very cautious of safety risks and closely monitor the agent.
In the default setting (monitor = true
), the agent will prompt for confirmation before executing every operation.
This setting pauses the agent before each operation, allowing for close examination, action rejection, and other human intervention like manually doing some operation when needed.
You should always monitor the agent's predictions before execution to prevent harmful outcomes. Please reject any action that may cause any potential harm.
You can monitor and intervene actions through terminal input before each execution:
Y
orEnter
: Accept this action.n
: Reject this action and record it in the action history.i
: Reject this action and pause for human intervention.- During the pause, you can do anything, such as opening or closing tabs, opening another link, and so on, except for directly closing the browser. If the current active tab is closed, the active tab will default to the last tab in the browser. If all tabs are closed, the browser will reopen a Google page.
- You can leave a message after manual operations, which will be injected into the prompt of the agent, for better human-agent cooperation.
e
: Terminate the session and save results.
We do not support direct login actions to safeguard your personal information and prevent exposure to potential safety and legal risks. To prevent unintended consequential errors, we advise against using SeeAct for tasks that require account login.
Multimodal-Mind2Web is the multimodal version of Mind2Web dataset hosted on Huggingface under OpenRAIL License. In this dataset, we align each HTML document in the dataset with its corresponding webpage screenshot image from the Mind2Web Raw Dump. This multimodal version addresses the inconvenience of loading images from the ~300GB Mind2Web Raw Dump.
- train: 7775 actions from 1009 tasks.
- test_task: 1339 actions from 177 tasks. Tasks from the same website are seen during training.
- test_website: 1019 actions from 142 tasks. Websites are not seen during training.
- test_domain: 4060 actions from 694 tasks. Entire domains are not seen during training.
The train set may include some screenshot images not properly rendered caused by rendering issues during Mind2Web annotation. The three test splits (test_task, test_website, test_domain) have undergone human verification to confirm element visibility and correct rendering for action prediction.
Each line in the dataset is an action consisting of screenshot image, HTML text and other fields required for action prediction, for the convenience of inference.
- "annotation_id" (str): unique id for each task
- "website" (str): website name
- "domain" (str): website domain
- "subdomain" (str): website subdomain
- "confirmed_task" (str): task description
- "action_reprs" (list[str]): human readable string representation of the action sequence
- "screenshot" (str): path to the webpage screenshot image corresponding to the HTML.
- "action_uid" (str): unique id for each action (step)
- "raw_html" (str): raw html of the page before the action is performed
- "cleaned_html" (str): cleaned html of the page before the action is performed
- "operation" (dict): operation to perform
- "op" (str): operation type, one of CLICK, TYPE, SELECT
- "original_op" (str): original operation type, contain additional HOVER and ENTER that are mapped to CLICK, not used
- "value" (str): optional value for the operation, e.g., text to type, option to select
- "pos_candidates" (list[dict]): ground truth elements. Here we only include positive elements that exist in "cleaned_html" after our preprocessing, so "pos_candidates" might be empty. The original labeled element can always be found in the "raw_html".
- "tag" (str): tag of the element
- "is_original_target" (bool): whether the element is the original target labeled by the annotator
- "is_top_level_target" (bool): whether the element is a top level target find by our algorithm. please see the paper for more details.
- "backend_node_id" (str): unique id for the element
- "attributes" (str): serialized attributes of the element, use
json.loads
to convert back to dict
- "neg_candidates" (list[dict]): other candidate elements in the page after preprocessing, has similar structure as "pos_candidates"
You can also generate screenshot image and query text data from the Mind2Web raw dump. Run the following commands to generate screenshot images and overlay image annotation for each grounding method:
cd src/offline_experiments/screenshot_generation
# Textual Choices
python textual_choices.py
# Element Attributes
python element_attributes.py
# Image Annotation
python image_annotation.py
Please download the Mind2Web raw dump from this link and the query source data from here. After downloading, please place both files in the ../data/
directory.
To reproduce the online evaluation experiments in the paper, run the following command to run SeeAct in auto mode:
python src/seeact.py -c config/online_exp.toml
Note: Some tasks may require manual updates to the task descriptions due to time sensitivity.
We followed the 2-stage strategy of MindAct for a fair comparison. You can find the trained ranker model DeBERTa-v3-base in the Huggingface Model Hub.
The code under this repo is licensed under an OPEN RAIL-S License.
The data under this repo is licensed under an OPEN RAIL-D License.
The model weight and parameters under this repo are licensed under an OPEN RAIL-M License.
The code was released solely for research purposes, with the goal of making the web more accessible via language technologies. The authors are strongly against any potentially harmful use of the data or technology by any party.
We extend our heartfelt thanks to Xiang Deng for his original contributions to the SeeAct system. Additionally, we are grateful to our colleagues from the OSU NLP group for testing the SeeAct system and offering valuable feedback.
Questions or issues? File an issue or contact Boyuan Zheng, Boyu Gou, Huan Sun, Yu Su, The Ohio State University
If you find this work useful, please consider starring our repos and citing our papers:
@article{zheng2024seeact,
title={GPT-4V(ision) is a Generalist Web Agent, if Grounded},
author={Boyuan Zheng and Boyu Gou and Jihyung Kil and Huan Sun and Yu Su},
journal={arXiv preprint arXiv:2401.01614},
year={2024},
}
@inproceedings{deng2023mindweb,
title={Mind2Web: Towards a Generalist Agent for the Web},
author={Xiang Deng and Yu Gu and Boyuan Zheng and Shijie Chen and Samuel Stevens and Boshi Wang and Huan Sun and Yu Su},
booktitle={Thirty-seventh Conference on Neural Information Processing Systems},
year={2023},
url={https://openreview.net/forum?id=kiYqbO3wqw}
}