Mobile GUI Agent Privacy Personalization with Trajectory Induced Preference Optimization
Zhixin Lin, Jungang Li, Dongliang Xu, Shidong Pan, Yibo Shi + 3 more
TLDR
TIPO enables mobile GUI agents to personalize privacy preferences by optimizing execution trajectories, improving persona alignment and task success.
Key contributions
- Identifies the overlooked problem of privacy personalization in mobile GUI agents and its impact on trajectories.
- Proposes Trajectory Induced Preference Optimization (TIPO) for handling variable-length, structurally diverse trajectories.
- TIPO uses preference-intensity weighting and padding gating to emphasize privacy-related steps and reduce alignment noise.
- Achieves improved persona alignment, distinction, and strong task executability on a new Privacy Preference Dataset.
Why it matters
This paper addresses a critical gap in mobile GUI agent development by focusing on user privacy personalization. It introduces a novel optimization method, TIPO, that allows agents to adapt to diverse user preferences without sacrificing task performance. This work is crucial for building more user-centric and trustworthy AI assistants.
Original Abstract
Mobile GUI agents powered by Multimodal Large Language Models (MLLMs) can execute complex tasks on mobile devices. Despite this progress, most existing systems still optimize task success or efficiency, neglecting users' privacy personalization. In this paper, we study the often-overlooked problem of agent personalization. We observe that personalization can induce systematic structural heterogeneity in execution trajectories. For example, privacy-first users often prefer protective actions, e.g., refusing permissions, logging out, and minimizing exposure, leading to logically different execution trajectories from utility-first users. Such variable-length and structurally different trajectories make standard preference optimization unstable and less informative. To address this issue, we propose Trajectory Induced Preference Optimization (TIPO), which uses preference-intensity weighting to emphasize key privacy-related steps and padding gating to suppress alignment noise. Results on our Privacy Preference Dataset show that TIPO improves persona alignment and distinction while preserving strong task executability, achieving 65.60% SR, 46.22 Compliance, and 66.67% PD, outperforming existing optimization methods across various GUI tasks. The code and dataset will be publicly released at https://github.com/Zhixin-L/TIPO.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.