Portrait of Xiangtao Meng

孟祥涛

Xiangtao Meng

2nd-year Ph.D., School of Cyber Science and Technology

Shandong University

Qingdao, China

Exploring secure and trustworthy AI, from deepfake detection to robust large language models.

Trustworthy Machine Learning

Researching safety, robustness, and privacy across generative models and LLM agents.

Deepfake Forensics

Building attacks and defenses for facial forgery detection in practical pipelines.

Secure LLM Systems

Designing evaluation frameworks that expose risk interactions and support safer deployments.

Bio

Profile

Exploring secure and trustworthy AI, from deepfake detection to robust large language models.

Based in Qingdao, China. 2nd-year Ph.D., School of Cyber Science and Technology at Shandong University.

  • Trustworthy Machine Learning

    Researching safety, robustness, and privacy across generative models and LLM agents.

  • Deepfake Forensics

    Building attacks and defenses for facial forgery detection in practical pipelines.

  • Secure LLM Systems

    Designing evaluation frameworks that expose risk interactions and support safer deployments.

Updates

Latest News

2025-11-03
Featured by MIT Technology Review China

Media coverage of our latest LLM defense study.

Read more
2025-10-10
Preprint: From Defender to Devil?

Investigating unintended risk interactions introduced by LLM defenses.

Read more
2025-09-18
ErrorTrace accepted at NeurIPS 2025 (spotlight)

Black-box traceability based on model family error space.

Read more
2025-09-06
Industry collaboration launched

Joint research project on LLM security testing and risk assessment with Topsec.

Read more
2025-08-28
Preprint: Safe-Control

Safety patch for mitigating unsafe content in text-to-image generation models.

Read more
2025-08-13
DCMI accepted at CCS 2025

Differential calibration membership inference against RAG.

Read more
2025-03-11
Fuzz-testing meets LLM-based agents accepted at IEEE S&P 2025

Automated framework for jailbreaking text-to-image generation models.

Read more
2024-11-15
Outstanding master's thesis

Recognized for thesis on robustness research for deepfake detection.

Read more

Selected Works

Publications

2025
From Defender to Devil? Unintended Risk Interactions Induced by LLM Defenses
Xiangtao Meng, Tianshuo Cong, Li Wang, Wenyu Chen, Zheng Li✉, Shanqing Guo✉, Xiaoyun Wang✉
arXiv · LLM Safety Risk Analysis
2025
ErrorTrace: A Black-Box Traceability Mechanism Based on Model Family Error Space
Chuanchao Zang, Xiangtao Meng, Wenyu Chen, Tianshuo Cong, Zha Yaxing, Dong Qi, Zheng Li, Shanqing Guo
NeurIPS (Spotlight) · Model Provenance NeurIPS 2025
2025
Safe-Control: A Safety Patch for Mitigating Unsafe Content in Text-to-Image Generation Models
Xiangtao Meng, Yingkai Dong, Ning Yu, Li Wang, Zheng Li✉, Shanqing Guo✉
arXiv · T2I Safety Defense
2025
DCMI: A Differential Calibration Membership Inference Attack Against Retrieval-Augmented Generation
Xinyu, Xiangtao Meng✉, Yingkai Dong, Zheng Li✉, Shanqing Guo✉
CCS · RAG Security CCS 2025
2025
Fuzz-testing meets LLM-based agents: An automated and efficient framework for jailbreaking text-to-image generation models
Yingkai Dong, Xiangtao Meng, Ning Yu, Li Wang, Zheng Li✉, Shanqing Guo✉
IEEE S&P · Adversarial Testing IEEE S&P
2024
AVA: Inconspicuous Attribute Variation-based Adversarial Attack bypassing DeepFake Detection
Xiangtao Meng, Li Wang, Shanqing Guo✉, Lei Ju, Qingchuan Zhao
IEEE S&P · Deepfake Attack Code Released
2024
DEEPFAKER: A Unified Evaluation Platform for Facial Deepfake and Detection Models
Li Wang, Xiangtao Meng, Dan Li, Xuhong Zhang, Shouling Ji, Shanqing Guo✉
ACM TOPS · Benchmark CCF B