← Back to search
paper reviewed open access llmsec-2023-00006
Multi-step Jailbreaking Privacy Attacks on ChatGPT
Haoran Li, Dadi Guo, Wei Fan, Mingshi Xu, Jie Huang, Fanpu Meng, Yangqiu Song
2023-04 — EMNLP 2023 Findings 175 citations
Abstract
Demonstrates multi-step jailbreaking attacks to extract personal information from ChatGPT, showing how sequential prompting can bypass safety measures.
Categories
Tags
multi-stepprivacyPII-extraction
Framework Mappings
OWASP LLM: LLM01 OWASP LLM: LLM02 MITRE ATLAS: AML.T0054 MITRE ATLAS: AML.T0056
Cite This Resource
@article{llmsec202300006,
title = {Multi-step Jailbreaking Privacy Attacks on ChatGPT},
author = {Haoran Li and Dadi Guo and Wei Fan and Mingshi Xu and Jie Huang and Fanpu Meng and Yangqiu Song},
year = {2023},
journal = {EMNLP 2023 Findings},
url = {https://arxiv.org/abs/2304.05197},
} Metadata
- Added
- 2026-04-14
- Added by
- manual
- Source
- manual
- arxiv_id
- 2304.05197