← Back to search
paper reviewed open access llmsec-2023-00006

Multi-step Jailbreaking Privacy Attacks on ChatGPT

Haoran Li, Dadi Guo, Wei Fan, Mingshi Xu, Jie Huang, Fanpu Meng, Yangqiu Song

2023-04 — EMNLP 2023 Findings 175 citations

Abstract

Demonstrates multi-step jailbreaking attacks to extract personal information from ChatGPT, showing how sequential prompting can bypass safety measures.

Categories

Tags

multi-stepprivacyPII-extraction

Framework Mappings

OWASP LLM: LLM01 OWASP LLM: LLM02 MITRE ATLAS: AML.T0054 MITRE ATLAS: AML.T0056

Cite This Resource

@article{llmsec202300006,
  title = {Multi-step Jailbreaking Privacy Attacks on ChatGPT},
  author = {Haoran Li and Dadi Guo and Wei Fan and Mingshi Xu and Jie Huang and Fanpu Meng and Yangqiu Song},
  year = {2023},
  journal = {EMNLP 2023 Findings},
  url = {https://arxiv.org/abs/2304.05197},
}

Metadata

Added
2026-04-14
Added by
manual
Source
manual
arxiv_id
2304.05197