Webui: A dataset for enhancing visual ui understanding with web semantics

J Wu, S Wang, S Shen, YH Peng, J Nichols… - Proceedings of the …, 2023 - dl.acm.org
Modeling user interfaces (UIs) from visual information allows systems to make inferences
about the functionality and semantics needed to support use cases in accessibility, app …

Vision-based mobile app gui testing: A survey

S Yu, C Fang, Z Tuo, Q Zhang, C Chen, Z Chen… - arxiv preprint arxiv …, 2023 - arxiv.org
Graphical User Interface (GUI) has become one of the most significant parts of mobile
applications (apps). It is a direct bridge between mobile apps and end users, which directly …

Never-ending learning of user interfaces

J Wu, R Krosnick, E Schoop, A Swearngin… - Proceedings of the 36th …, 2023 - dl.acm.org
Machine learning models have been trained to predict semantic information about user
interfaces (UIs) to make apps more accessible, easier to test, and to automate. Currently …

Unblind text inputs: predicting hint-text of text input in mobile apps via LLM

Z Liu, C Chen, J Wang, M Chen, B Wu… - Proceedings of the …, 2024 - dl.acm.org
Mobile apps have become indispensable for accessing and participating in various
environments, especially for low-vision users. Users with visual impairments can use screen …

Llamatouch: A faithful and scalable testbed for mobile ui task automation

L Zhang, S Wang, X Jia, Z Zheng, Y Yan… - Proceedings of the 37th …, 2024 - dl.acm.org
The emergent large language/multimodal models facilitate the evolution of mobile agents,
especially in mobile UI task automation. However, existing evaluation approaches, which …

Uicoder: Finetuning large language models to generate user interface code through automated feedback

J Wu, E Schoop, A Leung, T Barik, JP Bigham… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) struggle to consistently generate UI code that compiles and
produces visually relevant designs. Existing approaches to improve generation rely on …

Iluvui: Instruction-tuned language-vision modeling of uis from machine conversations

Y Jiang, E Schoop, A Swearngin, J Nichols - arxiv preprint arxiv …, 2023 - arxiv.org
Multimodal Vision-Language Models (VLMs) enable powerful applications from their fused
understanding of images and language, but many perform poorly on UI tasks due to the lack …

Interactive flexible style transfer for vector graphics

J Warner, KW Kim, B Hartmann - Proceedings of the 36th Annual ACM …, 2023 - dl.acm.org
Vector graphics are an industry-standard way to represent and share visual designs.
Designers frequently source and incorporate styles from existing designs into their work …

FinerMe: Examining App-level and Feature-level Interventions to Regulate Mobile Social Media Use

A Orzikulova, H Cho, HY Chung, H Hong… - Proceedings of the …, 2023 - dl.acm.org
Many digital wellbeing tools help users monitor and control social media use on their
smartphones by tracking and setting limits on their usage time. Tracking is typically done at …

Parammacros: Creating ui automation leveraging end-user natural language parameterization

R Krosnick, S Oney - … on Visual Languages and Human-Centric …, 2022 - ieeexplore.ieee.org
Prior work in programming-by-demonstration (PBD) has explored ways to enable end-users
to create custom automation without needing to write code. We propose a new end-user …