← Back to search
paper reviewed open access llmsec-2024-00015
Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory
Niloofar Mireshghallah, Hyunwoo Kim, Xuhui Zhou, Yulia Tsvetkov, Maarten Sap, Reza Shokri, Yejin Choi
2024-01 — ICLR 2024 110 citations
Abstract
Evaluates LLM privacy behavior through the lens of contextual integrity theory, finding significant mismatches between LLM norms and human privacy expectations.
Framework Mappings
OWASP LLM: LLM02 NIST AI RMF: GOVERN NIST AI RMF: MAP
Cite This Resource
@article{llmsec202400015,
title = {Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory},
author = {Niloofar Mireshghallah and Hyunwoo Kim and Xuhui Zhou and Yulia Tsvetkov and Maarten Sap and Reza Shokri and Yejin Choi},
year = {2024},
journal = {ICLR 2024},
url = {https://arxiv.org/abs/2310.17884},
} Metadata
- Added
- 2026-04-14
- Added by
- manual
- Source
- manual
- arxiv_id
- 2310.17884