How Capable are Web Agents at Solving Common Knowledge Work Tasks? (2024)

Alexandre Drouin  Maxime Gasse  Massimo Caccia  Issam H. Laradji  Manuel Del Verme  Tom Marty  Léo Boisvert  Megh Thakkar  Quentin Cappart  David Vazquez  Nicolas Chapados  Alexandre Lacoste

Abstract

We study the use of large language model-based agents for interacting with software via web browsers. Unlike prior work, we focus on measuring the agents’ ability to perform tasks that span the typical daily work of knowledge workers utilizing enterprise software systems. To this end, we propose WorkArena, a remote-hosted benchmark of 33 tasks based on the widely-used ServiceNow platform. We also introduce BrowserGym, an environment for the design and evaluation of such agents, offering a rich set of actions as well as multimodal observations. Our empirical evaluation reveals that while current agents show promise on WorkArena, there remains a considerable gap towards achieving full task automation. Notably, our analysis uncovers a significant performance disparity between open and closed-source LLMs, highlighting a critical area for future exploration and development in the field.

Machine Learning, ICML

How Capable are Web Agents at Solving Common Knowledge Work Tasks? (1)
How Capable are Web Agents at Solving Common Knowledge Work Tasks? (2)

1 Introduction

Graphical User Interfaces (UIs) are the predominant medium through which people interact with software, serving as a crucial gateway to the digital world.While they have evolved to become more intuitive, featuring generic and universal components like forms, lists, and buttons, UIs can still make complex or repetitive tasks burdensome for users. While being more and more intuitive, those complex UIs also became inadvertedly more and more discriminative for visually-impaired users.An ideal user experience would involve automated assistants that can streamline these tasks ensuring accessibility for everyone.While Application Programming Interfaces (APIs) have facilitated programmatic interactions with software, the resulting automated assistants often lack transparency, are difficult for users to inspect, and are not universally available.In contrast, assistants that directly manipulate UIs (UI assistants) offer greater transparency and are more amenable to human oversight.Most notably, because the user can give and take back control over the UI at any point, UI assistants can provide varying levels of automation ranging from partial assistance (such as finding a menu or filling a form) to complete task execution (like placing an order), akin to the six levels of automation in autonomous driving(SAE, 2021).

Recent advancements in the fields of large language and vision models have seen the rapid development of UI assistants, particularly web agents acting through a browser(Furuta etal., 2023; Kim etal., 2023; Gur etal., 2023b). The range of web tasks explored in the literature varies from simple UI commands such as selecting specific menu elements on toy web pages(Liu etal., 2018; Shi etal., 2017a), to more complex requests such as “Checkout merge requests assigned to me”, on real-world websites like Reddit and GitLab(Zhou etal., 2023). Yet, one area in which web agents can be particularly impactful and remain unexplored is enterprise software. In the workplace, where repetitive tasks are common, enterprise software often prioritizes functionality over user experience, leading to inefficiencies and long learning curves for workers.Our work addresses this gap, and investigates the potential of web agents in enterprise settings to improve accessibility, user experience, and worker productivity.

To this end, we introduce WorkArena, a benchmark developed on the widely-used ServiceNow platform(ServiceNow, 2023). ServiceNow is a comprehensive cloud-based platform that offers solutions for automating and managing digital workflows across various enterprise functions, including IT service management, human resources, customer service, and security operations.In 2023 their customer base counted over 7,000 companies worldwide, including 85% of the Fortune 500 companies(Mastantuono, 2023).Within these firms alone, the ServiceNow platform potentially impacts over 12 million individuals, not including broader public interactions,such as the 500,000 daily users of Disney+’s customer help center(Maas, 2020).ServiceNow’s extensive reach makes it an ideal real-world environment for evaluating the potential impact of UI assistants in the workplace.

Our contributions are as follows:

  • WorkArena: A realistic benchmark of enterprise-related tasks for web agents comprising 19,912unique task instances (§3, Figure1a);

  • BrowserGym: A new framework for the development and evaluation of web agents, compatible with previous benchmarks like WebArena(Zhou etal., 2023), MiniWoB(Liu etal., 2018; Shi etal., 2017a) and WebShop(Yao etal., 2022), that offers a richer set of multimodal observations (e.g., screenshot, accessibility tree, screen coordinates), a broader set of actions (e.g., Python code and high-level primitives), and supports chat-based interactions (§4, Figure1b); Surprisingly, these features contribute to bringing our GPT-4 agent at the top of the leaderboard on WebArena, with a score of 25.4%, contrasting with the score of the original paper, 14.4%.

  • Empirical study: We report a collection of experiments to assess the ability of state-of-the-art large language model (LLM)-based agents to solve WorkArena, as well as an analysis of the impact of the different BrowserGym features on WorkArena and MiniWoB (§5).

2 Related Works

Benchmarks for web agents:Early benchmarks for web agents were based on synthetic web environments where agents were tasked with performing low-level keyboard and mouse actions(Shi etal., 2017b). Notable examples are MiniWoB(Shi etal., 2017a; Liu etal., 2018), which offer a collection of 125 toy web tasks ranging from clicking a specific button to using a basic text editor, and WebShop(Yao etal., 2022), a simulated e-commerce website with shopping tasks that require searching and browsing a catalog of items. More recently, Zhou etal. (2023) introduced WebArena, a collection of 190 tasks based on realistic websites that emulate real-world domains such as e-commerce, social forums, collaborative software development, and content management. WebArena is a notoriously challenging benchmark, with a success rate of 14% for a state-of-the-art web agent based on GPT-4, and 78% for human agents.Deng etal. (2023) proposed Min2Web, a large-scale dataset of 2,000 web interactions from 137 websites curated by human annotators.Similarly, Lù etal. (2024) propose WebLINX, a curated dataset of web interactions composed of 2337 expert demonstrations from 155 different real-world websites. In WebLINX, each task is composed of a turn-based chat dialogue averaging 43 interactions per task.Recently, He etal. (2024) propose 300 information-retrieval tasks, from 15 real-world consumer websites (e.g., Amazon, Coursera, Booking), which are used to evaluate WebVoyager, a vision-based web agent.

Worth mentioning is the related body of work on mobile UI agent benchmarks, which includes the early PixelHelp dataset (Li etal., 2020) and more recent Android in the Wild (Rawles etal., 2023) and Macro Mining (Huang etal., 2024).

Our proposed benchmark, WorkArena, is designed to complement existing work by specifically focusing on real-world enterprise software applications. It includes a wide range of tasks that collectively encompass several end-to-end workflows typically performed by knowledge workers.Additionally, it poses a series of technical challenges, such as pages with very large document object models (DOMs), non-standard HTML, and complex UI elements, which we outline in §3.2. This benchmark integrates into BrowserGym, a new environment that we propose for the evaluation of web agents, which aggregates all features proposed in previous work, such as multimodal observations and code-based actions while being the first to support chat-based agent-user interactions (§4.1).

LLM-based Agents: The scope of our experimental contributions is limited to web agents that rely on language models for reasoning.Recent studies include the seminal work of Nakano etal. (2021) that introduces WebGPT, an agent capable of browsing the web and answering questions via information retrieval.Other works have also explored web agents that receive HTML as input and produce a series of high-level actions such as click, type, select(Deng etal., 2023; Liu etal., 2023a, b; Yao etal., 2023).

Other works have shown that using only textual information as input is sometimes limiting and have considered multimodal observations that combine visual information (screenshots of a page) with text(Humphreys etal., 2022; He etal., 2024).Instead of directly interacting with a website, recent works have proposed methods that can act on websites using Python-generated code from task-specific instructions (Gur etal., 2023a, b).Our proposed environment, BrowserGym is flexible in that it supports all observations and actions spaces utilized in prior research.

3 WorkArena – An Enterprise Benchmark

WorkArena consists of a suite of 33tasks and 19,912unique instances that cover core interactions with the ServiceNow platform, such as filtering lists, filling forms, searching knowledge bases, utilizing service catalogs, reading dashboards, and navigating the workspace via menus (see Figure1(a)). Collectively, these tasks are representative of a wide array of common operations that employees, like IT, administrative, and white-collar staff, perform on a daily basis.

As a guiding example, consider an IT support agent tasked with onboarding new hires. Each day, this agent logs into the ServiceNow platform. On the landing page, they see a dashboard showing key statistics of the work assigned to them. Using the workspace’s menu, they navigate to a list of requests to be fulfilled. They then filter the list to extract all requests assigned to them and sort them by priority. Finally, they process each request by filling out forms to create new user profiles and using the service catalog to order laptops for them.As we will see, all of the interactions listed above are included in WorkArena, and this is only one of the many realistic user trajectories that the benchmark covers.

How Capable are Web Agents at Solving Common Knowledge Work Tasks? (3)

3.1 WorkArena Tasks

How Capable are Web Agents at Solving Common Knowledge Work Tasks? (4)

In WorkArena, each task is coupled with a natural language goal that provides instructions to the agent. Each goal is automatically generated from a human-designed template filled with pre-defined values (menu name, field value, item specifications etc.), and explicitely provides all the information required to solve the task (examples in Figure2). Other key components of tasks include: i) validation functions and ii) oracle functions, a unique feature that distinguishes WorkArena from previous work. The validation functions offer real-time feedback to agents, identifying errors ranging from minor (e.g., unfilled mandatory fields) to critical (such as pushing invalid data to the database).The oracle functions are hand-crafted solutions that automatically complete the tasks using Playwright browser automation(Microsoft, 2023).They serve three purposes: (i)they ensure the feasibility of each task, (ii)they act as ground truth for agents that have learning capabilities, and (iii)they help maintain the benchmark’s longevity by making it easier to identify and adjust tasks impacted by future updates to the ServiceNow platform.Below, we outline the main categories of tasks included in the benchmark.

Lists: We consider 12 list-based tasks, which can be grouped into two categories: filtering and sorting. The former consists of using the UI to construct a complex filter with 1 to 5 conditions. The latter consists of using the UI to sort the list based on up to 3 columns. In both cases, the interaction with the UI is non-trivial and requires opening a hidden menu, adding the right number of conditions, and filling them out accordingly. There are 6 tasks of each type, each corresponding to a different data table (e.g., users, incidents). In both cases, client-side validation is used to verify that the resulting list satisfies the expected conditions. Together, these tasks yield 6,900 instances.

Forms: We consider 5 form-based tasks, which each consist of creating a new entry in a given data table. These tasks vary in complexity based on the number of fields that must be filled (from 1 to 26) and on the intricate properties of each form’s UI. For example, some forms require navigating through a set of tabs to expose hidden fields. Others use dynamic auto-completions fields, which require careful handling (see Figure3 for an example). In all cases, validation proceeds by querying the database to retrieve entries created by the agent and verifying that their values are as expected. Together, these tasks yield 5,000 instances.

Knowledge bases: The benchmark includes an information retrieval task that consists of searching the platform’s knowledge base to answer a question. Concretely, this requires searching with appropriate keywords and browsing the resulting articles to find specific information. These tasks are constructed by starting from a list of facts, generating articles containing each fact with GPT-4(OpenAI, 2023), and generating a series of questions that unambiguously ask for this fact. Then, validation proceeds by verifying if the answer returned by the agent is within a set of acceptable answers. For example, if the question is “What is the level of customer satisfaction?” and the answer is “8.5/10”, alternative answers such as “85%” or “8.5 out of 10” would be accepted. In total, this task yields 1,000 instances. Details on article generation and validation are given in §A.3.

Service catalogs: The benchmark includes 9 tasks that require navigating a catalog of products and ordering items with given specifications. Such tasks vary in complexity based on the number of item configuration options. In all cases, validation is done by querying the database to verify that the order request created by the agent includes the right items in the right amounts, with the expected specifications. Together, these tasks yield 3,550 instances.

Dashboards: The benchmark includes 4 retrieval tasks that require extracting numerical data from charts and (optionally) performing simple reasoning over them to answer a question. The complexity of such tasks depends on i) whether there are one or many charts on the page and ii) the reasoning to be done (e.g., extract the minimum).In all cases, validation consists of verifying if the agent’s response contains key numbers and labels referring to chart elements.Together, these tasks yield 1,862 instances.

Menus: We consider 2 menu-based tasks: i) navigating via the “All” menu and ii) impersonating users. The first consists of using the platform’s main menu to navigate to a given application. In this case, validation simply verifies that the agent has arrived at the expected location. The second consists of impersonating a user, a task commonly performed by IT support agents, where the agent logs into the platform as a given user to diagnose an issue. In this case, validation verifies that the expected user is logged in. Together, these tasks yield 1,600 instances.

A detailed list of all the tasks implemented in WorkArena is available in §A.1 Table6.

3.2 Challenges: the World Wild Web of Work

The ServiceNow platform poses a specific set of challenges for UI assistants, which we believe make WorkArena a complementary and meaningful benchmark for the community.

Non-standard, dynamic UIs: First, the web pages are heavily dynamic, and exhibit complex UI elements and ways of interacting with them. For example, the form to create an incident contains specific rules that can make some fields required or hidden, depending on the value of other fields (e.g., setting an incident’s status to “Resolved” requires filling its “Resolution notes”). Or, on some pages, the right-click can be overloaded to display a dynamic menu in certain areas. While these UI behaviors are not necessarily standard or best practices in web development, they are fairly common and representative of real-world enterprise software, which is not always designed with user accessibility in mind.

Non-standard, exotic HTML: Second, the ServiceNow platform relies on a complex combination of web technologies to implement its web pages, with nested iFrames, shadow DOMs, and proprietary Javascript APIs and HTML tags that do not necessarily adhere to existing web standards.111https://www.w3.org/standards/ This specificity would require strong out-of-distribution generalization for a UI assistant to successfully solve a task.

Large HTML: Third, the Document Object Model (DOM) of rendered web pages in the ServiceNow platform can be prohibitively large even for state-of-the-art language models, with a flat HTML text size that ranges between 40k and 500k tokens, even after a basic cleaning (removing scripts, styles and empty elements). Thus, even a conceptually simple task, such as finding the next element to click on the current page requires long context understanding, which is an active area of research in language models.

3.3 Availability

WorkArena is open-source and designed to be easily extended by the scientific community.222https://github.com/ServiceNow/WorkArenaTasks are executed on Personal Developer Instances, which are real cloud-based instances of the ServiceNow product, accessible for free through their developer program.It is important to note that although our benchmark is built on top of ServiceNow’s platform, it operates independently of any proprietary code.

4 BrowserGym

Along with the WorkArena benchmark, we introduce BrowserGym, a generic browser environment that facilitates the design of new benchmarks, and provides a solid platform for the evaluation of multi-modal web agents.333https://github.com/ServiceNow/BrowserGym BrowserGym (Figure1(b)) is implemented as an OpenAI Gym environment(Brockman etal., 2016) and follows a Partially-Observable Markov Decision Process (POMDP) paradigm.444BrowserGym extracts at each time step an observation that is based only on the current view of the page. This emphasizes the need for web agents to implement some kind of memory mechanism, which BrowserGym does not provide. It relies on Chromium and uses the Chrome DevTools Protocol (CDP)(Google, 2023) and the Playwright library(Microsoft, 2023) to interact with the web browser.

4.1 Capabilities

BrowserGym implements the following capabilities.

Chat-based user interaction: One of the interaction modalities is a chat interface where the user and the web agent can exchange messages. In WorkArena, the goal of each task is provided as the initial user message, to which the agent can reply at any time. This allows for information retrieval tasks where a specific answer is expected from the agent, but also sequential tasks where user instructions change over time and are delivered sequentially in a way that mimics real-world use cases.

Augmented attributes: For every element on the current page BrowserGym provides a unique identifier (bid), its bounding box coordinates (left,top,right,bottom), and flags indicating whether it is visible and clickable. These attributes both provide a crude summary of the visual rendering of the UI and allow for unambiguous interaction with individual elements using their identifiers.

Rich observation space: at each time step, the observation space contains the content of the chat (list of messages), the currently open pages (list of URLs), the error message from the last action if any (stack trace), and a multi-modal view of the active web page: its HTML DOM snapshot (structured object), its accessibility tree or AXTree (structured object) as originally proposed byZhou etal. (2023), and a viewport screenshot (image). Both DOM and AXTree are obtained through Chrome’s CDP, and are enriched with augmented attributes (bid, coords, visible, clickable). These structured objects can typically be rendered as text for processing by a language model and can be combined with the screenshot for a vision-language model.

Rich action space: the action space is customizable and includes Python code, which can be restricted to specific set of pre-defined high-level primitives, such as bid-based actions (click(bid), type(bid,text), …), and coord-based actions (mouse_click(x,y), keyboard_type(text), …). Alternatively, the action space can allow for the execution of arbitrary Python code, including the entire Playwright API which gives the web agent maximum flexibility in interacting with the browser.For the complete list of high-level primitives available in Browsergym, refer to §B, Table10.

Multi-page navigation: BrowserGym natively supports web tasks that require multiple open pages (tabs, popups) and is also robust to web pages that employ nested encapsulation techniques such as iFrames and shadow DOMs. This robustness is essential to handle the heterogeneity of real-world websites and is missing in existing web environments.

4.2 An Ideal Experimental Framework

Flexible agent design: BrowserGym offers an extensive list of features but does not impose any restriction on how web agents should be implemented. The agent is responsible for using the provided observations or not (HTML, AXTree, screenshot, error message), deciding how to handle the history (past observations and actions), or deciding which action space it should be using (python, bid, coord, coord+bid). As such, with BrowserGym, researchers can easily experiment with new ideas and evaluate and compare a wide variety of web agents on the same set of tasks, such as text-only agents, vision-augmented agents, memory-augmented agents, and so on.

Minimal task design: BrowserGym reduces the burden of creating new benchmarks to a minimum. Implementing a new task in BrowserGym boils down to implementing four functions: setup(), teardown(), validate() and cheat() (an optional oracle function). The setup() function is responsible for initializing anything the task needs beforehand, such as creating database entries, navigating to the starting URL, authenticating, etc. Likewise, teardown() is responsible for cleaning up any resource that might have been created during the task’s execution. validate() is responsible for checking if the task’s goal was fulfilled, which can involve operations such as querying a database, validating the URL and the content of the current page, or looking at the messages sent by the agent in the chat. The method returns a reward, an optional user message for the chat, and a done flag indicating the end of the task. Finally, each task can optionally implement an oracle function, cheat(), that automatically solves the task using a hard-coded Playwright solution. This can notably be used to assess the feasibility of tasks.

Extensibility:BrowserGymis easily extensible to additional benchmarks. We currently support MiniWoB §B.2), WebArena (Zhou etal., 2023) and WorkArena. We encourage the community to add new benchmarks or new agents to this platform. We also offer support to reduce the friction of adding new benchmarks to the platform. Please reach out on GitHub for assistance.

5 Experiments

We present a series of empirical experiments to assess the performance of state-of-the-art, general-purpose LLMs at solving work-related web tasks, using WorkArena and BrowserGym. The aim of these experiments is two-fold. First, we situate the level of difficulty of WorkArena by comparing it across baselines and benchmarks. Second, we propose an ablation study to quantify the impact of the different features offered in BrowserGym. All experiments are run with BrowserGym v0.3.5 and WorkArena v0.3.0.555Code: https://github.com/ServiceNow/AgentLab.

TypeFlagGPT-4oGPT-3.5Llama3
Agentuse_thinking
use_action_history
use_error_history
use_think_history
Obs. Spaceuse_focused_element
use_last_error
coords
extract_visible_tag
extract_clickable_tag
only_visible_elements
Act. Spacemulti_actions
action_setbidbidbid
individual_examples
long_description

5.1 Agent Design

We implement a simple web agent with chain-of-thought prompting(Wei etal., 2022b), and evaluate its performance across two axes: (1) the underlying LLM, and (2) the use of BrowserGym features. To study the effect of different design choices in our agents, we use flags to activate or deactivate certain features, such as use_thinking for chain-of-thought.

Observation space: Our observation space is composed of the goal, the current page’s HTML and/or AXTree,666On WebArena and WorkArena we only use AXTrees because HTML is prohibitively large. On MiniWoB we use both AXTree and HTML as it consistently gave the best performance. the currently focused element, and the error from the previous action if any. We activate / deactivate these features with flags use_last_error and use_focused_element, and we decide whether or not to augment each element with extra information using flags coords=center or coords=box for screen coordinates, and extract_visible_tag and extract_clickable_tag for whether elements are visible and/or clickable. Last, the flag filter_visible_only allow us to only include elements that are visible on the screen.

Action space: We leverage the flexibility of BrowserGym with a multi_action flag, which either allows agents to execute multiple actions per step (e.g., click(12) click(52)) or not. The action_set=bid flag only permits primitives that use element identifiers, while action_set=bid+coord also includes primitives operating with 2D coordinates (x,y). This is useful in MiniWoB, where certain tasks require clicking at specific positions in SVG images. Finally, we also experiment with the textual description of the action space which can either include a short or long description for each primitive (long_description), as well as examples of valid function calls or not (individual_examples).

History: To extend the horizon window of our agents, we experiment with use_action_history which includes in the prompt the history of actions since the beginning of the episode, use_error_history which includes all previous error messages, and use_think_history which re-injects the previous chain-of-though outputs into the prompt, thus creating an artificial memory that gives agents a chance to recall their previous thoughs.

Zero-shot examples: In the prompt, we provide a single generic example of how the chain of thought and action outputs should be formatted. This contrasts with other methods(Kim etal., 2023) where task-specific few-shot examples are provided, yet aligns with our objective of developing zero-shot agents able to solve a large range of new tasks.

Parse and retry: Once the LLM provides an answer, we have a parsing loop that can re-prompt the agent up to 4 times to make it aware of a parsing mistake. This can save the agent from making basic mistakes and is mainly useful for less capable LLMs such as GPT-3.5. Once parsed, the action is executed via BrowserGym, which moves to the next step.

Language models: Our study distinguishes between closed- and open-source LLMs. For the closed-source segment, we evaluate GPT-3.5 (gpt-3.5-turbo-1106, 16K context) and GPT-4o(OpenAI, 2023) (gpt-4o-2024-05-13, 128K context), through OpenAI’s API. In the realm of open-source LLMs, we sought a model that 1) understands code and HTML, 2) can manage a substantial context size, and 3) is instruction-finetuned. Our choice fell on Llama3-70b(Meta, 2024) (meta-llama-3-70B-instruct, 8K context), a recently released model that exhibits performances close to GPT-4. This model was deployed using Hugging Face’s Text Generation Inference (TGI) library on 4 A100 GPUs. We also explore the effect of providing the screenshot of the page using GPT-4o vision and Set-of-Mark(Yang etal., 2023) as proposed in WebVoyager(He etal., 2024).

Prompt truncation: We use a maximum prompt length of 40K tokens for GPT-4o, 15K for GPT-3.5 and 8K for Llama3. When the prompt is too large (Figure4), we progressively truncate the HTML and AXTree from the end until it fits the maximum allowed number of tokens.

Task CategoryGPT-4oGPT-4o-VGPT-3.5Llama3
SR %±plus-or-minus\pm±SESR %±plus-or-minus\pm±SESR %±plus-or-minus\pm±SESR %±plus-or-minus\pm±SE
WorkArena (33 tasks)42.7±plus-or-minus\pm±1.541.8±plus-or-minus\pm±1.76.1±plus-or-minus\pm±1.317.9±plus-or-minus\pm±1.5
 Dashboard (4)62.5±plus-or-minus\pm±6.872.5±plus-or-minus\pm±6.020.0±plus-or-minus\pm±4.837.5±plus-or-minus\pm±6.0
 Form (5)40.0±plus-or-minus\pm±5.934.0±plus-or-minus\pm±4.82.0±plus-or-minus\pm±2.532.0±plus-or-minus\pm±4.6
 Knowledge (1)80.0±plus-or-minus\pm±12.270.0±plus-or-minus\pm±13.90.0±plus-or-minus\pm±4.330.0±plus-or-minus\pm±12.3
 List-filter (6)0.0±plus-or-minus\pm±1.60.0±plus-or-minus\pm±1.70.0±plus-or-minus\pm±1.60.0±plus-or-minus\pm±1.8
 List-sort (6)10.0±plus-or-minus\pm±3.813.3±plus-or-minus\pm±4.08.3±plus-or-minus\pm±3.71.7±plus-or-minus\pm±2.5
 Menu (2)60.0±plus-or-minus\pm±8.090.0±plus-or-minus\pm±6.05.0±plus-or-minus\pm±4.70.0±plus-or-minus\pm±2.9
 Service catalog (9)77.8±plus-or-minus\pm±3.265.6±plus-or-minus\pm±3.65.6±plus-or-minus\pm±2.326.7±plus-or-minus\pm±3.4
MiniWoB (125 tasks)66.1±plus-or-minus\pm±1.067.7±plus-or-minus\pm±1.038.9±plus-or-minus\pm±1.162.6±plus-or-minus\pm±0.6
 WebGum Subset (56)82.9±plus-or-minus\pm±1.583.2±plus-or-minus\pm±1.553.6±plus-or-minus\pm±1.480.5±plus-or-minus\pm±1.0
WebArena (812 tasks)23.5±plus-or-minus\pm±0.724.0±plus-or-minus\pm±0.66.7±plus-or-minus\pm±0.611.0±plus-or-minus\pm±0.6
 Content-and-config (411)25.8±plus-or-minus\pm±1.026.8±plus-or-minus\pm±0.98.8±plus-or-minus\pm±0.812.7±plus-or-minus\pm±0.9
 Information-seeking (325)22.5±plus-or-minus\pm±1.022.5±plus-or-minus\pm±0.94.3±plus-or-minus\pm±0.99.8±plus-or-minus\pm±1.1
 Navigation (76)15.8±plus-or-minus\pm±2.215.8±plus-or-minus\pm±1.85.3±plus-or-minus\pm±1.96.6±plus-or-minus\pm±1.9

5.2 Experimental Protocol

Standard Error: To be able to run a range of experiments under a fixed budget, we limit the number of seeds to 10 per task for MiniWoB and WorkArena, and 1 per task for WebArena. After averaging results over each benchmark (or subset of it), we usually observe a sufficiently low standard error to draw the needed conclusions. We use stratified bootstrapto obtain 1,000 samples of the mean and report the average and standard deviation of these means as success rate and standard error.

Max step: In all tasks we give our agents a maximum of 15 steps per episode, as in He etal. (2024). This ensures a low-performing agent will not wander for too long if they are incapable of solving the task. Note that 15 steps is considered sufficient for MiniWoB, but on WorkArena some tasks might require more than 15 steps to complete unless the agent runs in multi-action mode (see §A.1 Table6 for an upper bound).

Model selection: We evaluate all LLMs using a common agent code-base that we tune by activating / deactivating flags (see §5.1). For each LLM we find the best agent configuration via random search on MiniWoB and WorkArena. We then fix the final configuration (one per LLM), and re-run a full evaluation on all benchmarks with a different seed.777WebArena being deterministic, we don’t use it for tuning. Final configurations are reported in Table1.

5.3 Results

We report the final performance of our agents on MiniWoB(Liu etal., 2018), WebArena(Zhou etal., 2023) and WorkArena in Table2. We emphasize our key findings below.

GPT-4o’s Superiority: The data unequivocally demonstrate GPT-4o’s dominance over GPT-3.5 and Llama3, on all benchmarks. The performance gap is particularly striking on WorkArena, where GPT-4o achieves 43% success, in stark contrast to GPT-3.5’s 6.1% and Llama3’s 17.9%. The performance disparity between LLMs is significantly more pronounced in WorkArena and WebArena than in MiniWoB, aligning with findings on the emergent properties of AI systems(Wei etal., 2022a). As the complexity of task increases, the necessity for a more advanced LLM to achieve any score becomes apparent, with noticeable improvements in performance correlating with the enhancement of capabilities.

WorkArena poses a great challenge: Consistent with our expectations, our newly proposed benchmark, WorkArena, proves to be a significant challenge for current LLMs. While the tasks remain simple at a high-level (navigating menus, filling forms) and mostly require general knowledge, the difficulty arises primarily from the use of complex user interfaces from real-world software environments, resulting in long contexts and non-trivial interactions. Consequently, all agents exhibit low performance levels, with no agent achieving 100% success in one specific task. Notably, the list-based tasks which require interacting with a non-standard HTML widget (§A.2 Figure5) prove to be the most challenging ones with a clear 0% success rate for all LLMs, while being fairly simple for a human to complete.

State-of-the-art on MiniWoB and WebArena: Our GPT-4o agent demonstrates notably high performance on MiniWoB, achieving significantly greater success compared to other agents. The outcomes on the full benchmark and the WebGum subset (82.9%) surpass those of prior studies on zero-shot web agents (Assouel etal., 2023; Zeng etal., 2023), underscoring the effectiveness of both our agent design and the features offered by BrowserGym. On WebArena, while the score obtained by our GPT-4o agent is only 23.5%, it constitutes the best zero-shot performance reported so far, way above the 14.4% succes rate obtained using GPT-4 in the original paper(Zhou etal., 2023).

Open-source LLMs: While our Llama3 agent still performs poorer than a closed-source GPT-4o, its performance is remarkably high compared to a GPT-3.5 agent despite its shorter context window (8K vs 16K). Most notably, in preliminary experiments with a Llama2 model we could not obtain any success (flat 0%) on both WebArena and WorkArena, which makes us hopeful for the future of open-source LLMs. These results also indicate that web agent benchmarks, and WorkArena in particular, offer a great tool for evaluating emerging capabilities in LLMs, with larger performance gaps and more room for improvement compared to regular benchmarks (Touvron etal., 2023).

ConfigurationMiniWoBWorkArena
SR %±plus-or-minus\pm±SESR %±plus-or-minus\pm±SE
Initial configuration68.2±plus-or-minus\pm±1.045.5±plus-or-minus\pm±2.2
+multi_actions68.5±plus-or-minus\pm±1.040.6±plus-or-minus\pm±2.0
+coords=box,action_set=bid+coord72.6±plus-or-minus\pm±1.041.2±plus-or-minus\pm±1.8
+use_think_history66.7±plus-or-minus\pm±0.942.4±plus-or-minus\pm±2.3
+use_error_history67.2±plus-or-minus\pm±0.943.6±plus-or-minus\pm±2.1
-extract_visible_tag68.8±plus-or-minus\pm±1.043.0±plus-or-minus\pm±2.2

ConfigurationMiniWoBWorkArena
SR %±plus-or-minus\pm±SESR %±plus-or-minus\pm±SE
Initial configuration41.3±plus-or-minus\pm±1.18.5±plus-or-minus\pm±1.3
-use_thinking30.2±plus-or-minus\pm±1.06.1±plus-or-minus\pm±1.2
-use_action_history34.1±plus-or-minus\pm±1.15.2±plus-or-minus\pm±1.3
+use_think_history36.2±plus-or-minus\pm±1.23.0±plus-or-minus\pm±1.0
+use_error_history37.9±plus-or-minus\pm±1.07.0±plus-or-minus\pm±1.4
+multi_actions41.4±plus-or-minus\pm±1.19.4±plus-or-minus\pm±1.5
+only_visible_elements38.1±plus-or-minus\pm±1.05.8±plus-or-minus\pm±1.3
+long_description39.5±plus-or-minus\pm±1.06.4±plus-or-minus\pm±1.2
-individual_examples34.4±plus-or-minus\pm±1.08.2±plus-or-minus\pm±1.3
+coords=center, action_set=bid+coord40.8±plus-or-minus\pm±1.14.5±plus-or-minus\pm±1.1
+coords=box, action_set=bid+coord40.5±plus-or-minus\pm±1.26.1±plus-or-minus\pm±1.4
+use_focused_element40.3±plus-or-minus\pm±1.19.4±plus-or-minus\pm±1.5
-extract_visible_tag37.8±plus-or-minus\pm±1.06.7±plus-or-minus\pm±1.3
+extract_clickable_tag37.9±plus-or-minus\pm±1.17.6±plus-or-minus\pm±1.2

ConfigurationMiniWoBWorkArena
SR %±plus-or-minus\pm±SESR %±plus-or-minus\pm±SE
Initial configuration59.8±plus-or-minus\pm±1.020.0±plus-or-minus\pm±2.3
-use_thinking48.6±plus-or-minus\pm±1.18.5±plus-or-minus\pm±1.7
-use_action_history49.9±plus-or-minus\pm±1.18.5±plus-or-minus\pm±1.7
-use_think_history52.2±plus-or-minus\pm±1.118.8±plus-or-minus\pm±2.1
+multi_actions63.0±plus-or-minus\pm±1.017.6±plus-or-minus\pm±2.1
+long_description55.5±plus-or-minus\pm±1.017.6±plus-or-minus\pm±2.0
+use_last_error60.2±plus-or-minus\pm±1.119.4±plus-or-minus\pm±1.8
-extract_visible_tag61.3±plus-or-minus\pm±0.915.8±plus-or-minus\pm±2.0

5.4 Ablation Study

We report in Tables3, 4 and5 an ablation study of each LLM on MiniWoB and WorkArena, for the most important agent flags.888The ablation study was done with a different seed, hence numbers differ slightly from those in Table2. We present our findings below.

Chain-of-though is crucial: In all of our 3 agents, asking the LLM to directly produce the next best action without chain-of-though is very detrimental to performance. This is clearly seen in Table5, with a drop of 10 points for Llama3 on both MiniWoB and WorkArena. This conclusion holds for all three LLMs very clearly.

More is not always better: As we gradually add more features, the prompt becomes longer and seems to overwhelm the LLM. This is mostly observed with GPT-3.5 and Llama3, the weaker LLMs, where seemingly harmless features such as a long_description or individual_examples in the action space description actually degrade performance. We hypothesise that since the AXTree is already large, adding too much non-crucial features is likely to distract the agent instead of helping it. We also note that in WorkArena, more features imply more prompt truncation for GPT-3.5 and Llama3 due to their limited context length (Figure4).

Memory can hurt: The use_think_history flag provides an interesting capacity to our agents, that of remembering their thoughts from previous time steps. We observed in practice that this features works, in the sense that the agents do re-use information about previous pages. However, such a feature is mostly useful when the task requires some form of memory, such as navigating a first page to find some information (e.g., an address), and then going to another page (e.g., Google maps) to type this information. On both WorkArena and MiniWoB the vast majority of tasks does not require navigation, and all the necessary information is always present on the current page. What we observe instead, is that agents with this feature activated tend to stick to decisions decided in early time steps, even erroneous ones, and are somehow less prone to self-correction. This is mostly observed with the GPT agents, Tables3 and4.

Deceiving multimodal performance: We report in Table2 a complimentary experiment with GPT-4o-V, a vision-augmented GPT-4o model, to assess the capacity of multi-modal models at utilizing the screenshot of the current web page for solving web tasks. We augment the screenshot with Set-of-Mark(Yang etal., 2023) as proposed in WebVoyager(He etal., 2024), and we keep the rest of the flags identical to the GPT-4o best performing agent. The results are somewhat deceiving, with very minor performance improvements on MiniWoB and WebArena. Similarly poor multimodal performances are reported by Xie etal. (2024), which could be explained by the lack of screen-related training data in current vision-language models. This leaves room for future development in this area.

2D features do not always help: Several MiniWoB tasks require precise 2D understanding of the UI and interaction with x,y𝑥𝑦x,yitalic_x , italic_y coordinates. While we observe a notable performance gain on this benchmark when giving GPT-4o additional coordinate features and actions (Table3), this does not translate to WorkArena where the tasks seem best solved using bid only. This seems to indicate that, while 2D actions are clearly required in specific cases (i.e., drawing on a web whiteboard), most web interactions actually do not require it, which can greatly simplify the agent design.

How Capable are Web Agents at Solving Common Knowledge Work Tasks? (5)

6 Conclusion

In this work, we introduced WorkArena, a new benchmark for the evaluation of web agents on tasks inspired by the day-to-day workflow of knowledge workers in the ServiceNow platform. WorkArena underscores the challenge of navigating real-world enterprise websites, notably dealing with large observations and complex, non-standard interfaces.

We also introduced BrowserGym, a robust, general-purpose environment for automated web agents, which encompasses an extensive list of features previously proposed in the literature (HTML, AXTree, screenshot, set-of-mark, code and high-level action space), as well as novel capabilities such as an interactive chat. BrowserGym allows for the flawless evaluation of web agents on multiple benchmarks within the same unified framework, and currently supports MiniWoB, WebArena and WorkArena.Both contributions are open source and meant to serve as catalysts to accelerate the development of new web agents and their evaluation in terms of capability and potential impact on the real world.

We presented an empirical evaluation of GPT-3.5, GPT-4o and Llama3 – among the most advanced general-purpose Large Language Models (LLMs) currently available for intruction-following and coding.Specifically, we investigated their generalization performance as web agents in MiniWoB, WebArena and WorkArena.Our results validate WorkArena as an unsolved, challenging benchmark that requires advanced reasoning capabilities over long contexts (HTML or AXTree), which seem to emerge only in very large models.

In future work we plan to integrate additional standard benchmarks into BrowserGym, such as WebShop(Yao etal., 2022) and WebVoyager(He etal., 2024).We also plan to expand WorkArena significantly by creating compositional tasks built upon the present benchmark, to obtain realistic complex workflows that require additional skills like retrieval, memorization, visual perception, and advanced reasoning.

Due to the multi-modal nature of web observations (textual and visual), and the potentially unlimited complexity of web tasks that can be designed (from toy setups like MiniWoB to harder benchmarks like WebArena and WorkArena), we believe that browser-based task automation provides the perfect testbed to evaluate the emergent capabilities of multimodal large language models.We hope that the present work will stimulate further progress in the community.

Impact Statement

This research presents contributions aimed at facilitating the development of UI assistants, with a particular focus on browser-based web agents and their application in the workspace.The emergence of such agents is bound to disrupt the way humans interact with digital software, and the way workers perform digital tasks.Below, we reflect on the potential positive and negative societal impacts of this work.

Positive impacts

  • Productivity. A prominent benefit of UI assistants / web agents is the significant boost in worker productivity that would be achieved by automating repetitive and monotonous tasks. This automation would not only streamline digital workflows but also free up valuable time for employees to engage in complex problem-solving and creative tasks, thereby enhancing overall work quality and innovation.

  • Accessibility Further, UI assistants / web agents could lead to a substantial leap in digital accessibility, opening new employment opportunities for individuals who may have previously been excluded from certain roles due to disabilities, such as visual impairments. This advancement could also mitigate labor shortages by expanding the talent pool, benefiting both employers and workers in the job market.

Negative impacts

  • Labor displacement. One concern is job market disruption, including labor displacement. However, we believe this disruption is more likely to lead to the evolution of job roles rather than their outright replacement. A prime example can be observed in the translation industry, which has been significantly disrupted by automatic translationvander Meer (2021), but continues to thrive nonetheless. Interestingly, benchmarks such as WorkArena may help in forecasting which job roles are more likely to be affected by automation and help take preventive measures to minimize downsides (e.g., preemptive human reskilling).

  • Cybersecurity. On another note, human-like web agents the potential for increased and elaborate cyberattacks via agents mimicking human interactions. This necessitates the implementation of security measures, such as the use of constrained language models (e.g., GPT-4(OpenAI, 2023)).

  • Privacy. The deployment of LLM-based web agents in the workplace also raises privacy concerns due to the necessity of transmitting sensitive information, requiring further security research.

  • Environment. Additionally, the significant energy consumption associated with the extensive use of LLMs for inference presents environmental challenges that must be addressed before their widespread implementation.

Web agents, as a disruptive technology, introduce several potential negative impacts. We believe that all of the above points warrant careful consideration by the research community.

References

  • Assouel etal. (2023)Assouel, R., Marty, T., Caccia, M., Laradji, I., Drouin, A., Rajeswar, S., Palacios, H., Cappart, Q., Vazquez, D., Chapados, N., Gasse, M., and Lacoste, A.The unsolved challenges of LLMs in open-ended web tasks: A case study.In NeurIPS 2023 Foundation Models for Decision Making Workshop, 2023.URL https://openreview.net/forum?id=jt3il4fC5B.
  • Brockman etal. (2016)Brockman, G., Cheung, V., Pettersson, L., Schneider, J., Schulman, J., Tang, J., and Zaremba, W.OpenAI gym, 2016.
  • Deng etal. (2023)Deng, X., Gu, Y., Zheng, B., Chen, S., Stevens, S., Wang, B., Sun, H., and Su, Y.Mind2Web: Towards a generalist agent for the web.arXiv, abs/2306.06070, 2023.
  • Furuta etal. (2023)Furuta, H., Nachum, O., Lee, K.-H., Matsuo, Y., Gu, S.S., and Gur, I.Multimodal web navigation with instruction-finetuned foundation models.arXiv, abs/2305.11854, 2023.URL https://arxiv.org/abs/2305.11854.
  • Google (2023)Google.Chrome devtools protocol, 2023.URL https://chromedevtools.github.io/devtools-protocol/.
  • Gur etal. (2023a)Gur, I., Furuta, H., Huang, A., Safdari, M., Matsuo, Y., Eck, D., and Faust, A.A real-world webagent with planning, long context understanding, and program synthesis.arXiv preprint arXiv:2307.12856, 2023a.
  • Gur etal. (2023b)Gur, I., Furuta, H., Huang, A., Safdari, M., Matsuo, Y., Eck, D., and Faust, A.A real-world WebAgent with planning, long context understanding, and program synthesis.arXiv, abs/2307.12856, 2023b.URL https://arxiv.org/abs/2307.12856.
  • He etal. (2024)He, H., Yao, W., Ma, K., Yu, W., Dai, Y., Zhang, H., Lan, Z., and Yu, D.WebVoyager: Building an end-to-end web agent with large multimodal models.arXiv, abs/2401.13919, 2024.URL https://arxiv.org/abs/2401.13919.
  • Huang etal. (2024)Huang, F., Li, G., Li, T., and Li, Y.Automatic macro mining from interaction traces at scale.In Proceedings of the CHI Conference on Human Factors in Computing Systems, CHI ’24, New York, NY, USA, 2024. Association for Computing Machinery.ISBN 9798400703300.doi: 10.1145/3613904.3642074.URL https://doi.org/10.1145/3613904.3642074.
  • Humphreys etal. (2022)Humphreys, P.C., Raposo, D., Pohlen, T., Thornton, G., Chhaparia, R., Muldal, A., Abramson, J., Georgiev, P., Santoro, A., and Lillicrap, T.A data-driven approach for learning to control computers.In International Conference on Machine Learning (ICML), 2022.
  • Kim etal. (2023)Kim, G., Baldi, P., and McAleer, S.Language models can solve computer tasks.arXiv, abs/2303.17491, 2023.URL https://arxiv.org/abs/2303.17491.
  • Li etal. (2020)Li, Y., He, J., Zhou, X., Zhang, Y., and Baldridge, J.Mapping natural language instructions to mobile ui action sequences.In Annual Conference of the Association for Computational Linguistics (ACL 2020), 2020.URL https://www.aclweb.org/anthology/2020.acl-main.729.pdf.
  • Liu etal. (2018)Liu, E.Z., Guu, K., Pasupat, P., Shi, T., and Liang, P.Reinforcement learning on web interfaces using workflow-guided exploration.In International Conference on Learning Representations (ICLR), 2018.
  • Liu etal. (2023a)Liu, X., Yu, H., Zhang, H., Xu, Y., Lei, X., Lai, H., Gu, Y., Ding, H., Men, K., Yang, K., Zhang, S., Deng, X., Zeng, A., Du, Z., Zhang, C., Shen, S., Zhang, T., Su, Y., Sun, H., Huang, M., Dong, Y., and Tang, J.AgentBench: Evaluating LLMs as agents.arXiv, abs/2308.03688, 2023a.URL https://arxiv.org/abs/2308.03688.
  • Liu etal. (2023b)Liu, Z., Yao, W., Zhang, J., Xue, L., Heinecke, S., Murthy, R., Feng, Y., Chen, Z., Niebles, J.C., Arpit, D., Xu, R., Mui, P., Wang, H., Xiong, C., and Savarese, S.BOLAA: Benchmarking and orchestrating LLM-augmented autonomous agents.arXiv, abs/2308.05960, 2023b.
  • Lù etal. (2024)Lù, X.H., Kasner, Z., and Reddy, S.Weblinx: Real-world website navigation with multi-turn dialogue.arXiv preprint arXiv:2402.05930, 2024.
  • Maas (2020)Maas, M.Knowledge 2020: “The digital workflow revolution has just begun”.Technical report, Sprinklr, 2020.URL https://www.linkedin.com/pulse/knowledge-2020-digital-workflow-revolution-has-just-begun-maas/.
  • Mastantuono (2023)Mastantuono, G.ServiceNow joins the prestigious Fortune 500 list.https://www.servicenow.com/blogs/2023/servicenow-joins-fortune-500-list.html, 2023.Accessed: 2024-01-29.
  • Meta (2024)Meta.Llama 3: Meta’s latest large language model.https://github.com/meta-llama/llama3, 2024.Accessed: 2024-06-03.
  • Microsoft (2023)Microsoft.Playwright for Python documentation, 2023.URL https://playwright.dev/python/.
  • Nakano etal. (2021)Nakano, R., Hilton, J., Balaji, S., Wu, J., Ouyang, L., Kim, C., Hesse, C., Jain, S., Kosaraju, V., Saunders, W., Jiang, X., Cobbe, K., Eloundou, T., Krueger, G., Button, K., Knight, M., Chess, B., and Schulman, J.WebGPT: Browser-assisted question-answering with human feedback.arXiv, abs/2112.09332, 2021.URL https://arxiv.org/abs/2112.09332.
  • OpenAI (2023)OpenAI.GPT-4 technical report.ArXiv, abs/2303.08774, 2023.URL https://arxiv.org/abs/2303.08774.
  • Rawles etal. (2023)Rawles, C., Li, A., Rodriguez, D., Riva, O., and Lillicrap, T.Androidinthewild: A large-scale dataset for android device control.In Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., and Levine, S. (eds.), Advances in Neural Information Processing Systems, volume36, pp. 59708–59728, 2023.
  • SAE (2021)SAE.Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles.Technical report, Society of Automotive Engineers (SAE), 04 2021.URL https://doi.org/10.4271/J3016_202104.
  • ServiceNow (2023)ServiceNow.Vancouver release notes.Online, 2023.Available at: https://docs.servicenow.com/bundle/vancouver-release-notes/.
  • Shi etal. (2017a)Shi, T., Karpathy, A., Fan, L., Hernandez, J., and Liang, P.World of bits: An open-domain platform for web-based agents.In International Conference on Machine Learning (ICML), 2017a.
  • Shi etal. (2017b)Shi, T., Karpathy, A., Fan, L., Hernandez, J., and Liang, P.World of bits: An open-domain platform for web-based agents.ICML, 2017b.
  • Touvron etal. (2023)Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., Bikel, D., Blecher, L., Canton-Ferrer, C., Chen, M., Cucurull, G., Esiobu, D., Fernandes, J., Fu, J., Fu, W., Fuller, B., Gao, C., Goswami, V., Goyal, N., Hartshorn, A., Hosseini, S., Hou, R., Inan, H., Kardas, M., Kerkez, V., Khabsa, M., Kloumann, I., Korenev, A., Koura, P.S., Lachaux, M.-A., Lavril, T., Lee, J., Liskovich, D., Lu, Y., Mao, Y., Martinet, X., Mihaylov, T., Mishra, P., Molybog, I., Nie, Y., Poulton, A., Reizenstein, J., Rungta, R., Saladi, K., Schelten, A., Silva, R., Smith, E.M., Subramanian, R., Tan, X.E., Tang, B., Taylor, R., Williams, A., Kuan, J.X., Xu, P., Yan, Z., Zarov, I., Zhang, Y., Fan, A., Kambadur, M., Narang, S., Rodriguez, A., Stojnic, R., Edunov, S., and Scialom, T.Llama 2: Open foundation and fine-tuned chat models.CoRR, abs/2307.09288, 2023.
  • vander Meer (2021)vander Meer, J.A journey into the future of the translation industry, 2021.URL https://www.taus.net/resources/blog/a-journey-into-the-future-of-the-translation-industry.Accessed: 2024-02-01.
  • Wei etal. (2022a)Wei, J., Tay, Y., Bommasani, R., Raffel, C., Zoph, B., Borgeaud, S., Yogatama, D., Bosma, M., Zhou, D., Metzler, D., etal.Emergent abilities of large language models.arXiv preprint arXiv:2206.07682, 2022a.
  • Wei etal. (2022b)Wei, J., Wang, X., Schuurmans, D., Bosma, M., ichter, b., Xia, F., Chi, E., Le, Q.V., and Zhou, D.Chain-of-thought prompting elicits reasoning in large language models.In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (eds.), Advances in Neural Information Processing Systems, volume35, pp. 24824–24837. Curran Associates, Inc., 2022b.URL https://proceedings.neurips.cc/paper_files/paper/2022/file/9d5609613524ecf4f15af0f7b31abca4-Paper-Conference.pdf.
  • Xie etal. (2024)Xie, T., Zhang, D., Chen, J., Li, X., Zhao, S., Cao, R., Hua, T.J., Cheng, Z., Shin, D., Lei, F., Liu, Y., Xu, Y., Zhou, S., Savarese, S., Xiong, C., Zhong, V., and Yu, T.Osworld: Benchmarking multimodal agents for open-ended tasks in real computer environments, 2024.
  • Yang etal. (2023)Yang, J., Zhang, H., Li, F., Zou, X., Li, C., and Gao, J.Set-of-mark prompting unleashes extraordinary visual grounding in gpt-4v.arXiv preprint arXiv:2310.11441, 2023.
  • Yao etal. (2022)Yao, S., Chen, H., Yang, J., and Narasimhan, K.WebShop: Towards scalable real-world web interaction with grounded language agents.In Advances in Neural Information Processing Systems (NeurIPS), 2022.
  • Yao etal. (2023)Yao, S., Zhao, J., Yu, D., Du, N., Shafran, I., Narasimhan, K., and Cao, Y.ReAct: Synergizing reasoning and acting in language models.arXiv, abs/2210.03629, 2023.URL https://arxiv.org/abs/2210.03629.
  • Zeng etal. (2023)Zeng, A., Liu, M., Lu, R., Wang, B., Liu, X., Dong, Y., and Tang, J.Agenttuning: Enabling generalized agent abilities for llms.arXiv preprint arXiv:2310.12823, 2023.
  • Zhou etal. (2023)Zhou, S., Xu, F.F., Zhu, H., Zhou, X., Lo, R., Sridhar, A., Cheng, X., Bisk, Y., Fried, D., Alon, U., and Neubig, G.Webarena: A realistic web environment for building autonomous agents.ArXiv, abs/2307.13854, 2023.URL https://arxiv.org/abs/2307.13854.

Appendix A WorkArena – Additional Details

A.1 Tasks

This section provides additional details on each type of task included in the benchmark.

CategoryTask NameInstancesOracle Actions
Lists(12 tasks) FilterAssetList1,00017.3 ±plus-or-minus\pm± 6.5
FilterChangeRequestList1,00018.7 ±plus-or-minus\pm± 4.7
FilterHardwareList1,00018.4 ±plus-or-minus\pm± 5.6
FilterIncidentList1,00016.2 ±plus-or-minus\pm± 4.2
FilterServiceCatalogItemList1,00019.9 ±plus-or-minus\pm± 5.9
FilterUserList1,00012.7 ±plus-or-minus\pm± 3.2
SortAssetList1507.4 ±plus-or-minus\pm± 2.3
SortChangeRequestList1507.7 ±plus-or-minus\pm± 1.6
SortHardwareList1508 ±plus-or-minus\pm± 2.3
SortIncidentList1508 ±plus-or-minus\pm± 2.7
SortServiceCatalogItemList1508.3 ±plus-or-minus\pm± 2.5
SortUserList1507.7 ±plus-or-minus\pm± 2.1
Forms(5 tasks) CreateChangeRequest1,00021.5 ±plus-or-minus\pm± 6.2
CreateIncident1,00023 ±plus-or-minus\pm± 7.9
CreateHardwareAsset1,00047.1 ±plus-or-minus\pm± 10.9
CreateProblem1,00010 ±plus-or-minus\pm± 3.4
CreateUser1,00017.9 ±plus-or-minus\pm± 5.2
Knowledge Bases (1 task) KnowledgeBaseSearch1,0004.0 ±plus-or-minus\pm± 0.0
Service Catalogs(9 tasks) OrderDeveloperLaptopMac1,0008.7 ±plus-or-minus\pm± 0.9
OrderIpadMini806.0 ±plus-or-minus\pm± 0.0
OrderIpadPro606.0 ±plus-or-minus\pm± 0.0
OrderSalesLaptop1,0009.0 ±plus-or-minus\pm± 0.8
OrderStandardLaptop1,0008.0 ±plus-or-minus\pm± 0.6
OrderAppleWatch104.0 ±plus-or-minus\pm± 0.0
OrderAppleMacBookPro15104.0 ±plus-or-minus\pm± 0.0
OrderDevelopmentLaptopPC406.0 ±plus-or-minus\pm± 0.0
OrderLoanerLaptop3508.0 ±plus-or-minus\pm± 0.0
Menus(2 tasks) AllMenu1,0003.0 ±plus-or-minus\pm± 0.0
Impersonation6007.0 ±plus-or-minus\pm± 0.0
Dashboards(4 tasks) SingleChartValueRetrieval10001.0 ±plus-or-minus\pm± 0.0
SingleChartMinMaxRetrieval3461.0 ±plus-or-minus\pm± 0.0
MultiChartValueRetrieval4442.0 ±plus-or-minus\pm± 0.0
MultiChartMinMaxRetrieval722.0 ±plus-or-minus\pm± 0.0
Total (33 tasks)19,912

A.2 Task User Interface Examples

In this section, we provide an example of the typical user interface encountered for each category of task (Figures5, 6, 7, 9 and8). We omit “form” tasks, as such an example has already been presented in Figure3.

How Capable are Web Agents at Solving Common Knowledge Work Tasks? (6)
How Capable are Web Agents at Solving Common Knowledge Work Tasks? (7)
How Capable are Web Agents at Solving Common Knowledge Work Tasks? (8)
How Capable are Web Agents at Solving Common Knowledge Work Tasks? (9)
How Capable are Web Agents at Solving Common Knowledge Work Tasks? (10)

A.3 Knowledge Base Tasks – Additional Details

This task consists of searching a company knowledge base for specific information to answer a given question.Here, we explain how the knowledge base included in WorkArena is generated, how we produce the questions and answers used in each task instance, and how validation is performed.

Generating the knowledge base:

The knowledge base included in WorkArena consists of 100 articles generated using GPT-4(OpenAI, 2023). To achieve this, we start from a list of 100 facts, which are each composed of an item and a value. Table7 shows a few examples. Then, for each fact, we use GPT-4 to produce an article in HTML format and make sure that the exact string “the {fact} is {item}” is included in the article. An example article is shown in Figure10.

FactItem
Password to conference room A-561roo918k
Address of office #45642, Pizza street, New York, USA
CEO’s nameAlex Johnson

How Capable are Web Agents at Solving Common Knowledge Work Tasks? (11)

Generating questions:

For each fact, i.e., (item, value) pair, we produce a list of questions that ask about item and whose answers are exactly value. We achieve this by prompting GPT-4 with the initial question “What is {item}?” and ask it to produce 10 alternative wordings for the question, as well as formatting instructions, to ensure that the answer is exactly value. Then, we prompt GPT-3.5 with the generated article and each question, ensuring that every single one is answered correctly. If the model fails to answer a question, we ask GPT-4 to improve it and we repeat the process. Note that we use GPT-3.5 to answer the questions to avoid the pitfall where GPT-4 would cater to itself, producing ambiguous questions that it somehow succeeds in answering correctly. Example questions and formatting instructions are shown in Table8.

Answer validation:

Despite the precise formatting instructions included with each question, we allow for slight variations in formatting and wording by verifying if the answer produced by the agent is within a set of acceptable answers. To produce such alternative answers, we provide the expected value to GPT-4 and ask it to produce 10 alternative formats. We include multiple examples in the prompt and inspect the results to ensure coherence. An example is shown in Table9.

QuestionFormatting InstructionsCould you provide the street location for Office #456?Make sure to include the Street Number, Street Name, City, and Country.Where can Office #456 be found?Provide the exact street address, city, and country.Where should one go to visit Office #456?Please respond with the format: Number, Street, City, Country.What’s the precise location of Office #456?Answer with the Street Number, Street Name, City, and Country.

Alternative Answer
42 Pizza Street, New York, USA
42, Pizza St., NY, United States
#42 Pizza Street, New York, U.S.
42 Pizza St, New York City, United States of America

Appendix B BrowserGym – Additional Details

B.1 Action Space

CategoryPrimitiveDescription
bidfill(bid, text)Fill an input field with text.
click(bid, button)Click an element.
dblclick(bid, button)Double-click an element.
hover(bid)Hover the mouse over an element.
press(bid, key_comb)Focus an element and press a combination of keys.
focus(bid)Focus an element.
clear(bid)Clear an input field.
select_option(bid, options)Select one or multiple options in a drop-down element.
drag_and_drop(from_bid, to_bid)Drag and drop one element to another.
coordmouse_move(x, y)Move the mouse to a location.
mouse_down(x, y, button)Move the mouse to a location then press and hold a mouse button.
mouse_up(x, y, button)Move the mouse to a location then release a mouse button.
mouse_click(x, y, button)Move the mouse to a location and click a mouse button.
mouse_dblclick(x, y, button)Move the mouse to a location and double-click a mouse button.
mouse_drag_and_drop(from_x, from_y, to_x, to_y)Drag and drop from a location to a location.
keyboard_down(key)Press and holds a keyboard key.
keyboard_up(key)Release a keyboard key.
keyboard_press(key_comb)Press a combination of keys.
keyboard_type(text)Types a string of text through the keyboard.
keyboard_insert_text(text)Insert a string of text in the currently focused element.
tabnew_tab()Open a new tab.
tab_close()Close the current tab.
tab_focus(index)Bring a tab to front (activate tab).
navgo_back()Navigate to the previous page in history.
go_forward()Navigate to the next page in history.
goto(url)Navigate to a url.
miscscroll(dx, dy)Scroll pixels in X and/or Y direction.
send_msg_to_user(text)Send a message to the user in the chat.
noop()Do nothing.
pythonAny python code (unsafe!)Executes code with playwright, the active page and the send_msg_to_user(text) primitive available.

B.2 MiniWoB

As part of BrowserGym, we provide a port of the MiniWoB benchmark(Shi etal., 2017a; Liu etal., 2018). Whereas in MiniWoB the goal of each task is embedded in the HTML of the web page, in BrowserGym the goal is provided in a separate chat window, accessible to the agent. We therefore run a minimal Javascript snippet in the setup() function of MiniWoB tasks to extract the goal from the web page and place it in the chat instead. Apart from that change and removing the hard time limit per episode, porting MiniWoB to BrowserGym required minimal effort.

How Capable are Web Agents at Solving Common Knowledge Work Tasks? (12)
How Capable are Web Agents at Solving Common Knowledge Work Tasks? (2024)
Top Articles
Latest Posts
Article information

Author: Gregorio Kreiger

Last Updated:

Views: 5299

Rating: 4.7 / 5 (77 voted)

Reviews: 84% of readers found this page helpful

Author information

Name: Gregorio Kreiger

Birthday: 1994-12-18

Address: 89212 Tracey Ramp, Sunside, MT 08453-0951

Phone: +9014805370218

Job: Customer Designer

Hobby: Mountain biking, Orienteering, Hiking, Sewing, Backpacking, Mushroom hunting, Backpacking

Introduction: My name is Gregorio Kreiger, I am a tender, brainy, enthusiastic, combative, agreeable, gentle, gentle person who loves writing and wants to share my knowledge and understanding with you.