Publications
2025
UnsafeBench: Benchmarking Image Safety Classifiers on Real-World and AI-Generated Images
Yiting Qu, Xinyue Shen, Yixin Wu, Michael Backes, Savvas Zannettou, Yang Zhang; CCS 2025
[arXiv] [Website] [Code]
Hate in Plain Sight: On the Risks of Moderating AI-Generated Hateful Illusions
Yiting Qu, Ziqing Yang, Yihan Ma, Michael Backes, Savvas Zannettou, Yang Zhang; ICCV 2025
[arXiv] [Code]
Bridging the Gap in Vision Language Models in Identifying Unsafe Concepts Across Modalities
Yiting Qu, Michael Backes, Yang Zhang; USENIX Security 2025
[arXiv] [Code]
HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate Campaigns
Xinyue Shen, Yixin Wu, Yiting Qu, Michael Backes, Savvas Zannettou, Yang Zhang; USENIX Security 2025
[arXiv] [Code]
From Meme to Threat: On the Hateful Meme Understanding and Induced Hateful Content Generation in Open-Source Vision Language Models
Yihan Ma, Xinyue Shen, Yiting Qu, Ning Yu, Michael Backes, Savvas Zannettou, Yang Zhang; USENIX Security 2025
[Code]
2024
FAKEPCD: Fake Point Cloud Detection via Source Attribution
Yiting Qu, Zhikun Zhang, Yun Shen, Michael Backes, Yang Zhang; ACM ASIACCS 2024
[arXiv] [Code]
Prompt Stealing Attacks Against Text-to-Image Generation Models
Xinyue Shen, Yiting Qu, Michael Backes, Yang Zhang; USENIX Security 2024
[arXiv] [Code]
2023
Unsafe Diffusion: On the Generation of Unsafe Images and Hateful Memes From Text-To-Image Models
Yiting Qu, Xinyue Shen, Xinlei He, Michael Backes, Savvas Zannettou, Yang Zhang; ACM CCS 2023
[arXiv] [Code] [Media Coverage]
On the Evolution of (Hateful) Memes by Means of Multimodal Contrastive Learning
Yiting Qu, Xinlei He, Shannon Pierson, Michael Backes, Yang Zhang, Savvas Zannettou; IEEE S&P 2023
[PDF] [arXiv] [Code]
2020
Automatic Permission Optimization Framework for Privacy Enhancement of Mobile Applications
Yiting Qu, Suguo Du, Shaofeng Li, Yan Meng, Le Zhang, Haojin Zhu; IoTJ 2020
[PDF]