SusBench: An Online Benchmark for Evaluating Dark Pattern Susceptibility of Computer-Use Agents
Longjie Guo, Chenjie Yuan, Mingyuan Zhong, Robert Wolfe, Ruican Zhong, Yue Xu, Bingbing Wen, Hua Shen, Lucy Lu Wang, Alexis Hiniker
Audit your AI agents against SusBench's nine dark pattern categories before deployment. If your agent handles transactions or account management, test specifically for hidden costs and preselection vulnerabilities—these are where autonomous systems cause the most user harm.
LLM-based agents autonomously clicking through interfaces inherit human vulnerabilities to manipulative UI patterns—confirmshaming, hidden costs, trick questions—but at machine scale.
Method: SusBench tests nine dark pattern types (confirmshaming, forced action, hidden costs, misdirection, nagging, obstruction, preselection, sneaking, trick questions) across believable e-commerce scenarios. The benchmark constructs realistic interfaces with manipulative elements embedded in checkout flows, subscription cancellations, and account settings. Agents are evaluated on whether they fall for these patterns or correctly resist them.
Caveats: Benchmark uses simulated e-commerce scenarios. Real-world dark patterns evolve faster than static test suites can capture.
Reflections: Can agents learn to recognize novel dark pattern variants not in the training set? · What's the tradeoff between dark pattern resistance and task completion speed? · Should agents warn users about detected dark patterns or silently resist them?