GUI-WORLD: A Dataset for GUI-oriented Multimodal LLM-based Agents
Recently, Multimodal Large Language Models (MLLMs) have been used as agents to control keyboard and mouse inputs by directly perceiving the Graphical User Interface (GUI) and generating corresponding code. However, current agents primarily exhibit excellent understanding capabilities in static envir...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , , , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Chen, Dongping Huang, Yue Wu, Siyuan Tang, Jingyu Chen, Liuyi Bai, Yilin He, Zhigang Wang, Chenlong Zhou, Huichi Li, Yiqiang Zhou, Tianshuo Yu, Yue Gao, Chujie Zhang, Qihui Gui, Yi Li, Zhen Wan, Yao Zhou, Pan Gao, Jianfeng Sun, Lichao |
description | Recently, Multimodal Large Language Models (MLLMs) have been used as agents
to control keyboard and mouse inputs by directly perceiving the Graphical User
Interface (GUI) and generating corresponding code. However, current agents
primarily exhibit excellent understanding capabilities in static environments
and are predominantly applied in relatively simple domains, such as Web or
mobile interfaces. We argue that a robust GUI agent should be capable of
perceiving temporal information on the GUI, including dynamic Web content and
multi-step tasks. Additionally, it should possess a comprehensive understanding
of various GUI scenarios, including desktop software and multi-window
interactions. To this end, this paper introduces a new dataset, termed
GUI-World, which features meticulously crafted Human-MLLM annotations,
extensively covering six GUI scenarios and eight types of GUI-oriented
questions in three formats. We evaluate the capabilities of current
state-of-the-art MLLMs, including ImageLLMs and VideoLLMs, in understanding
various types of GUI content, especially dynamic and sequential content. Our
findings reveal that ImageLLMs struggle with dynamic GUI content without
manually annotated keyframes or operation history. On the other hand, VideoLLMs
fall short in all GUI-oriented tasks given the sparse GUI video dataset. Based
on GUI-World, we take the initial step of leveraging a fine-tuned VideoLLM as a
GUI agent, demonstrating an improved understanding of various GUI tasks.
However, due to the limitations in the performance of base LLMs, we conclude
that using VideoLLMs as GUI agents remains a significant challenge. We believe
our work provides valuable insights for future research in dynamic GUI content
understanding. The code and dataset are publicly available at our project
homepage: https://gui-world.github.io/. |
doi_str_mv | 10.48550/arxiv.2406.10819 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2406_10819</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2406_10819</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-2632d618bcb6d02b4eff6d4d341ff9dd18dfa62c13d9783ada1335e64339e5fa3</originalsourceid><addsrcrecordid>eNotj8FKAzEURbNxIbUf4Mr8QKZJXiZN3A2t1kJKQSouhzd9iQxMO5KZiv69bXV1F-dy4DB2r2RhXFnKGebv9qvQRtpCSaf8LatWb2vxvn0Ny0de8SWOOMSRpz7zC-hzG49jJL45dWN76Ak7HsJGNOcX8erjDIc7dpOwG-L0fyds9_y0W7yIsF2tF1UQaOdeaAuarHLNvrEkdWNiSpYMgVEpeSLlKKHVewXk5w6QUAGU0RoAH8uEMGEPf9prQ_2Z2wPmn_rSUl9b4BfdvEJW</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>GUI-WORLD: A Dataset for GUI-oriented Multimodal LLM-based Agents</title><source>arXiv.org</source><creator>Chen, Dongping ; Huang, Yue ; Wu, Siyuan ; Tang, Jingyu ; Chen, Liuyi ; Bai, Yilin ; He, Zhigang ; Wang, Chenlong ; Zhou, Huichi ; Li, Yiqiang ; Zhou, Tianshuo ; Yu, Yue ; Gao, Chujie ; Zhang, Qihui ; Gui, Yi ; Li, Zhen ; Wan, Yao ; Zhou, Pan ; Gao, Jianfeng ; Sun, Lichao</creator><creatorcontrib>Chen, Dongping ; Huang, Yue ; Wu, Siyuan ; Tang, Jingyu ; Chen, Liuyi ; Bai, Yilin ; He, Zhigang ; Wang, Chenlong ; Zhou, Huichi ; Li, Yiqiang ; Zhou, Tianshuo ; Yu, Yue ; Gao, Chujie ; Zhang, Qihui ; Gui, Yi ; Li, Zhen ; Wan, Yao ; Zhou, Pan ; Gao, Jianfeng ; Sun, Lichao</creatorcontrib><description>Recently, Multimodal Large Language Models (MLLMs) have been used as agents
to control keyboard and mouse inputs by directly perceiving the Graphical User
Interface (GUI) and generating corresponding code. However, current agents
primarily exhibit excellent understanding capabilities in static environments
and are predominantly applied in relatively simple domains, such as Web or
mobile interfaces. We argue that a robust GUI agent should be capable of
perceiving temporal information on the GUI, including dynamic Web content and
multi-step tasks. Additionally, it should possess a comprehensive understanding
of various GUI scenarios, including desktop software and multi-window
interactions. To this end, this paper introduces a new dataset, termed
GUI-World, which features meticulously crafted Human-MLLM annotations,
extensively covering six GUI scenarios and eight types of GUI-oriented
questions in three formats. We evaluate the capabilities of current
state-of-the-art MLLMs, including ImageLLMs and VideoLLMs, in understanding
various types of GUI content, especially dynamic and sequential content. Our
findings reveal that ImageLLMs struggle with dynamic GUI content without
manually annotated keyframes or operation history. On the other hand, VideoLLMs
fall short in all GUI-oriented tasks given the sparse GUI video dataset. Based
on GUI-World, we take the initial step of leveraging a fine-tuned VideoLLM as a
GUI agent, demonstrating an improved understanding of various GUI tasks.
However, due to the limitations in the performance of base LLMs, we conclude
that using VideoLLMs as GUI agents remains a significant challenge. We believe
our work provides valuable insights for future research in dynamic GUI content
understanding. The code and dataset are publicly available at our project
homepage: https://gui-world.github.io/.</description><identifier>DOI: 10.48550/arxiv.2406.10819</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language ; Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2024-06</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2406.10819$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.10819$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Dongping</creatorcontrib><creatorcontrib>Huang, Yue</creatorcontrib><creatorcontrib>Wu, Siyuan</creatorcontrib><creatorcontrib>Tang, Jingyu</creatorcontrib><creatorcontrib>Chen, Liuyi</creatorcontrib><creatorcontrib>Bai, Yilin</creatorcontrib><creatorcontrib>He, Zhigang</creatorcontrib><creatorcontrib>Wang, Chenlong</creatorcontrib><creatorcontrib>Zhou, Huichi</creatorcontrib><creatorcontrib>Li, Yiqiang</creatorcontrib><creatorcontrib>Zhou, Tianshuo</creatorcontrib><creatorcontrib>Yu, Yue</creatorcontrib><creatorcontrib>Gao, Chujie</creatorcontrib><creatorcontrib>Zhang, Qihui</creatorcontrib><creatorcontrib>Gui, Yi</creatorcontrib><creatorcontrib>Li, Zhen</creatorcontrib><creatorcontrib>Wan, Yao</creatorcontrib><creatorcontrib>Zhou, Pan</creatorcontrib><creatorcontrib>Gao, Jianfeng</creatorcontrib><creatorcontrib>Sun, Lichao</creatorcontrib><title>GUI-WORLD: A Dataset for GUI-oriented Multimodal LLM-based Agents</title><description>Recently, Multimodal Large Language Models (MLLMs) have been used as agents
to control keyboard and mouse inputs by directly perceiving the Graphical User
Interface (GUI) and generating corresponding code. However, current agents
primarily exhibit excellent understanding capabilities in static environments
and are predominantly applied in relatively simple domains, such as Web or
mobile interfaces. We argue that a robust GUI agent should be capable of
perceiving temporal information on the GUI, including dynamic Web content and
multi-step tasks. Additionally, it should possess a comprehensive understanding
of various GUI scenarios, including desktop software and multi-window
interactions. To this end, this paper introduces a new dataset, termed
GUI-World, which features meticulously crafted Human-MLLM annotations,
extensively covering six GUI scenarios and eight types of GUI-oriented
questions in three formats. We evaluate the capabilities of current
state-of-the-art MLLMs, including ImageLLMs and VideoLLMs, in understanding
various types of GUI content, especially dynamic and sequential content. Our
findings reveal that ImageLLMs struggle with dynamic GUI content without
manually annotated keyframes or operation history. On the other hand, VideoLLMs
fall short in all GUI-oriented tasks given the sparse GUI video dataset. Based
on GUI-World, we take the initial step of leveraging a fine-tuned VideoLLM as a
GUI agent, demonstrating an improved understanding of various GUI tasks.
However, due to the limitations in the performance of base LLMs, we conclude
that using VideoLLMs as GUI agents remains a significant challenge. We believe
our work provides valuable insights for future research in dynamic GUI content
understanding. The code and dataset are publicly available at our project
homepage: https://gui-world.github.io/.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8FKAzEURbNxIbUf4Mr8QKZJXiZN3A2t1kJKQSouhzd9iQxMO5KZiv69bXV1F-dy4DB2r2RhXFnKGebv9qvQRtpCSaf8LatWb2vxvn0Ny0de8SWOOMSRpz7zC-hzG49jJL45dWN76Ak7HsJGNOcX8erjDIc7dpOwG-L0fyds9_y0W7yIsF2tF1UQaOdeaAuarHLNvrEkdWNiSpYMgVEpeSLlKKHVewXk5w6QUAGU0RoAH8uEMGEPf9prQ_2Z2wPmn_rSUl9b4BfdvEJW</recordid><startdate>20240616</startdate><enddate>20240616</enddate><creator>Chen, Dongping</creator><creator>Huang, Yue</creator><creator>Wu, Siyuan</creator><creator>Tang, Jingyu</creator><creator>Chen, Liuyi</creator><creator>Bai, Yilin</creator><creator>He, Zhigang</creator><creator>Wang, Chenlong</creator><creator>Zhou, Huichi</creator><creator>Li, Yiqiang</creator><creator>Zhou, Tianshuo</creator><creator>Yu, Yue</creator><creator>Gao, Chujie</creator><creator>Zhang, Qihui</creator><creator>Gui, Yi</creator><creator>Li, Zhen</creator><creator>Wan, Yao</creator><creator>Zhou, Pan</creator><creator>Gao, Jianfeng</creator><creator>Sun, Lichao</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240616</creationdate><title>GUI-WORLD: A Dataset for GUI-oriented Multimodal LLM-based Agents</title><author>Chen, Dongping ; Huang, Yue ; Wu, Siyuan ; Tang, Jingyu ; Chen, Liuyi ; Bai, Yilin ; He, Zhigang ; Wang, Chenlong ; Zhou, Huichi ; Li, Yiqiang ; Zhou, Tianshuo ; Yu, Yue ; Gao, Chujie ; Zhang, Qihui ; Gui, Yi ; Li, Zhen ; Wan, Yao ; Zhou, Pan ; Gao, Jianfeng ; Sun, Lichao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-2632d618bcb6d02b4eff6d4d341ff9dd18dfa62c13d9783ada1335e64339e5fa3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Dongping</creatorcontrib><creatorcontrib>Huang, Yue</creatorcontrib><creatorcontrib>Wu, Siyuan</creatorcontrib><creatorcontrib>Tang, Jingyu</creatorcontrib><creatorcontrib>Chen, Liuyi</creatorcontrib><creatorcontrib>Bai, Yilin</creatorcontrib><creatorcontrib>He, Zhigang</creatorcontrib><creatorcontrib>Wang, Chenlong</creatorcontrib><creatorcontrib>Zhou, Huichi</creatorcontrib><creatorcontrib>Li, Yiqiang</creatorcontrib><creatorcontrib>Zhou, Tianshuo</creatorcontrib><creatorcontrib>Yu, Yue</creatorcontrib><creatorcontrib>Gao, Chujie</creatorcontrib><creatorcontrib>Zhang, Qihui</creatorcontrib><creatorcontrib>Gui, Yi</creatorcontrib><creatorcontrib>Li, Zhen</creatorcontrib><creatorcontrib>Wan, Yao</creatorcontrib><creatorcontrib>Zhou, Pan</creatorcontrib><creatorcontrib>Gao, Jianfeng</creatorcontrib><creatorcontrib>Sun, Lichao</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chen, Dongping</au><au>Huang, Yue</au><au>Wu, Siyuan</au><au>Tang, Jingyu</au><au>Chen, Liuyi</au><au>Bai, Yilin</au><au>He, Zhigang</au><au>Wang, Chenlong</au><au>Zhou, Huichi</au><au>Li, Yiqiang</au><au>Zhou, Tianshuo</au><au>Yu, Yue</au><au>Gao, Chujie</au><au>Zhang, Qihui</au><au>Gui, Yi</au><au>Li, Zhen</au><au>Wan, Yao</au><au>Zhou, Pan</au><au>Gao, Jianfeng</au><au>Sun, Lichao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>GUI-WORLD: A Dataset for GUI-oriented Multimodal LLM-based Agents</atitle><date>2024-06-16</date><risdate>2024</risdate><abstract>Recently, Multimodal Large Language Models (MLLMs) have been used as agents
to control keyboard and mouse inputs by directly perceiving the Graphical User
Interface (GUI) and generating corresponding code. However, current agents
primarily exhibit excellent understanding capabilities in static environments
and are predominantly applied in relatively simple domains, such as Web or
mobile interfaces. We argue that a robust GUI agent should be capable of
perceiving temporal information on the GUI, including dynamic Web content and
multi-step tasks. Additionally, it should possess a comprehensive understanding
of various GUI scenarios, including desktop software and multi-window
interactions. To this end, this paper introduces a new dataset, termed
GUI-World, which features meticulously crafted Human-MLLM annotations,
extensively covering six GUI scenarios and eight types of GUI-oriented
questions in three formats. We evaluate the capabilities of current
state-of-the-art MLLMs, including ImageLLMs and VideoLLMs, in understanding
various types of GUI content, especially dynamic and sequential content. Our
findings reveal that ImageLLMs struggle with dynamic GUI content without
manually annotated keyframes or operation history. On the other hand, VideoLLMs
fall short in all GUI-oriented tasks given the sparse GUI video dataset. Based
on GUI-World, we take the initial step of leveraging a fine-tuned VideoLLM as a
GUI agent, demonstrating an improved understanding of various GUI tasks.
However, due to the limitations in the performance of base LLMs, we conclude
that using VideoLLMs as GUI agents remains a significant challenge. We believe
our work provides valuable insights for future research in dynamic GUI content
understanding. The code and dataset are publicly available at our project
homepage: https://gui-world.github.io/.</abstract><doi>10.48550/arxiv.2406.10819</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2406.10819 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2406_10819 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computation and Language Computer Science - Computer Vision and Pattern Recognition |
title | GUI-WORLD: A Dataset for GUI-oriented Multimodal LLM-based Agents |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-10T13%3A37%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=GUI-WORLD:%20A%20Dataset%20for%20GUI-oriented%20Multimodal%20LLM-based%20Agents&rft.au=Chen,%20Dongping&rft.date=2024-06-16&rft_id=info:doi/10.48550/arxiv.2406.10819&rft_dat=%3Carxiv_GOX%3E2406_10819%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |