Learning to Denoise Raw Mobile UI Layouts for Improving Datasets at Scale

The layout of a mobile screen is a critical data source for UI design research and semantic understanding of the screen. However, UI layouts in existing datasets are often noisy, have mismatches with their visual representation, or consists of generic or app-specific types that are difficult to anal...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Li, Gang, Baechler, Gilles, Tragut, Manuel, Li, Yang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Li, Gang
Baechler, Gilles
Tragut, Manuel
Li, Yang
description The layout of a mobile screen is a critical data source for UI design research and semantic understanding of the screen. However, UI layouts in existing datasets are often noisy, have mismatches with their visual representation, or consists of generic or app-specific types that are difficult to analyze and model. In this paper, we propose the CLAY pipeline that uses a deep learning approach for denoising UI layouts, allowing us to automatically improve existing mobile UI layout datasets at scale. Our pipeline takes both the screenshot and the raw UI layout, and annotates the raw layout by removing incorrect nodes and assigning a semantically meaningful type to each node. To experiment with our data-cleaning pipeline, we create the CLAY dataset of 59,555 human-annotated screen layouts, based on screenshots and raw layouts from Rico, a public mobile UI corpus. Our deep models achieve high accuracy with F1 scores of 82.7% for detecting layout objects that do not have a valid visual representation and 85.9% for recognizing object types, which significantly outperforms a heuristic baseline. Our work lays a foundation for creating large-scale high quality UI layout datasets for data-driven mobile UI research and reduces the need of manual labeling efforts that are prohibitively expensive.
doi_str_mv 10.48550/arxiv.2201.04100
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2201_04100</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2201_04100</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-bfb55794520b06d5dc1353d8eb03e66b5d008fa50373dbe3892523659f60b0663</originalsourceid><addsrcrecordid>eNotj81OwzAQhH3hgAoPwAm_QMImzjrJEbX8REqFBO05WtfrylIaV04o9O0hhdNIo_lG-oS4yyAtKkR4oPjtT2meQ5ZCkQFci6ZlioMf9nIKcsVD8CPLd_qS62B8z3LbyJbO4XMapQtRNodjDKd5vqKJRv6taZIfO-r5Rlw56ke-_c-F2Dw_bZavSfv20iwf24R0CYlxBrGsC8zBgLZod5lCZSs2oFhrgxagcoSgSmUNq6rOMVcaa6dnQKuFuP-7vbh0x-gPFM_d7NRdnNQPFO5FIA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Learning to Denoise Raw Mobile UI Layouts for Improving Datasets at Scale</title><source>arXiv.org</source><creator>Li, Gang ; Baechler, Gilles ; Tragut, Manuel ; Li, Yang</creator><creatorcontrib>Li, Gang ; Baechler, Gilles ; Tragut, Manuel ; Li, Yang</creatorcontrib><description>The layout of a mobile screen is a critical data source for UI design research and semantic understanding of the screen. However, UI layouts in existing datasets are often noisy, have mismatches with their visual representation, or consists of generic or app-specific types that are difficult to analyze and model. In this paper, we propose the CLAY pipeline that uses a deep learning approach for denoising UI layouts, allowing us to automatically improve existing mobile UI layout datasets at scale. Our pipeline takes both the screenshot and the raw UI layout, and annotates the raw layout by removing incorrect nodes and assigning a semantically meaningful type to each node. To experiment with our data-cleaning pipeline, we create the CLAY dataset of 59,555 human-annotated screen layouts, based on screenshots and raw layouts from Rico, a public mobile UI corpus. Our deep models achieve high accuracy with F1 scores of 82.7% for detecting layout objects that do not have a valid visual representation and 85.9% for recognizing object types, which significantly outperforms a heuristic baseline. Our work lays a foundation for creating large-scale high quality UI layout datasets for data-driven mobile UI research and reduces the need of manual labeling efforts that are prohibitively expensive.</description><identifier>DOI: 10.48550/arxiv.2201.04100</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Human-Computer Interaction ; Computer Science - Learning</subject><creationdate>2022-01</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2201.04100$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2201.04100$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Li, Gang</creatorcontrib><creatorcontrib>Baechler, Gilles</creatorcontrib><creatorcontrib>Tragut, Manuel</creatorcontrib><creatorcontrib>Li, Yang</creatorcontrib><title>Learning to Denoise Raw Mobile UI Layouts for Improving Datasets at Scale</title><description>The layout of a mobile screen is a critical data source for UI design research and semantic understanding of the screen. However, UI layouts in existing datasets are often noisy, have mismatches with their visual representation, or consists of generic or app-specific types that are difficult to analyze and model. In this paper, we propose the CLAY pipeline that uses a deep learning approach for denoising UI layouts, allowing us to automatically improve existing mobile UI layout datasets at scale. Our pipeline takes both the screenshot and the raw UI layout, and annotates the raw layout by removing incorrect nodes and assigning a semantically meaningful type to each node. To experiment with our data-cleaning pipeline, we create the CLAY dataset of 59,555 human-annotated screen layouts, based on screenshots and raw layouts from Rico, a public mobile UI corpus. Our deep models achieve high accuracy with F1 scores of 82.7% for detecting layout objects that do not have a valid visual representation and 85.9% for recognizing object types, which significantly outperforms a heuristic baseline. Our work lays a foundation for creating large-scale high quality UI layout datasets for data-driven mobile UI research and reduces the need of manual labeling efforts that are prohibitively expensive.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Human-Computer Interaction</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj81OwzAQhH3hgAoPwAm_QMImzjrJEbX8REqFBO05WtfrylIaV04o9O0hhdNIo_lG-oS4yyAtKkR4oPjtT2meQ5ZCkQFci6ZlioMf9nIKcsVD8CPLd_qS62B8z3LbyJbO4XMapQtRNodjDKd5vqKJRv6taZIfO-r5Rlw56ke-_c-F2Dw_bZavSfv20iwf24R0CYlxBrGsC8zBgLZod5lCZSs2oFhrgxagcoSgSmUNq6rOMVcaa6dnQKuFuP-7vbh0x-gPFM_d7NRdnNQPFO5FIA</recordid><startdate>20220111</startdate><enddate>20220111</enddate><creator>Li, Gang</creator><creator>Baechler, Gilles</creator><creator>Tragut, Manuel</creator><creator>Li, Yang</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220111</creationdate><title>Learning to Denoise Raw Mobile UI Layouts for Improving Datasets at Scale</title><author>Li, Gang ; Baechler, Gilles ; Tragut, Manuel ; Li, Yang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-bfb55794520b06d5dc1353d8eb03e66b5d008fa50373dbe3892523659f60b0663</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Human-Computer Interaction</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Li, Gang</creatorcontrib><creatorcontrib>Baechler, Gilles</creatorcontrib><creatorcontrib>Tragut, Manuel</creatorcontrib><creatorcontrib>Li, Yang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Li, Gang</au><au>Baechler, Gilles</au><au>Tragut, Manuel</au><au>Li, Yang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Learning to Denoise Raw Mobile UI Layouts for Improving Datasets at Scale</atitle><date>2022-01-11</date><risdate>2022</risdate><abstract>The layout of a mobile screen is a critical data source for UI design research and semantic understanding of the screen. However, UI layouts in existing datasets are often noisy, have mismatches with their visual representation, or consists of generic or app-specific types that are difficult to analyze and model. In this paper, we propose the CLAY pipeline that uses a deep learning approach for denoising UI layouts, allowing us to automatically improve existing mobile UI layout datasets at scale. Our pipeline takes both the screenshot and the raw UI layout, and annotates the raw layout by removing incorrect nodes and assigning a semantically meaningful type to each node. To experiment with our data-cleaning pipeline, we create the CLAY dataset of 59,555 human-annotated screen layouts, based on screenshots and raw layouts from Rico, a public mobile UI corpus. Our deep models achieve high accuracy with F1 scores of 82.7% for detecting layout objects that do not have a valid visual representation and 85.9% for recognizing object types, which significantly outperforms a heuristic baseline. Our work lays a foundation for creating large-scale high quality UI layout datasets for data-driven mobile UI research and reduces the need of manual labeling efforts that are prohibitively expensive.</abstract><doi>10.48550/arxiv.2201.04100</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2201.04100
ispartof
issn
language eng
recordid cdi_arxiv_primary_2201_04100
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
Computer Science - Human-Computer Interaction
Computer Science - Learning
title Learning to Denoise Raw Mobile UI Layouts for Improving Datasets at Scale
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T17%3A51%3A41IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Learning%20to%20Denoise%20Raw%20Mobile%20UI%20Layouts%20for%20Improving%20Datasets%20at%20Scale&rft.au=Li,%20Gang&rft.date=2022-01-11&rft_id=info:doi/10.48550/arxiv.2201.04100&rft_dat=%3Carxiv_GOX%3E2201_04100%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true