← Back to search
paper reviewed open access llmsec-2025-00024

Can Sensitive Information Be Deleted From LLMs? Objectives for Defending Against Extraction Attacks

Vaidehi Patil, Peter Hase, Mohit Bansal

2024 — ICLR 2024 70 citations

Abstract

Evaluates methods for deleting sensitive information from trained LLMs, finding current unlearning approaches insufficient against determined adversaries.

Categories

Tags

knowledge-deletionunlearningextraction-defense

Framework Mappings

OWASP LLM: LLM02

Cite This Resource

@article{llmsec202500024,
  title = {Can Sensitive Information Be Deleted From LLMs? Objectives for Defending Against Extraction Attacks},
  author = {Vaidehi Patil and Peter Hase and Mohit Bansal},
  year = {2024},
  journal = {ICLR 2024},
  url = {https://arxiv.org/abs/2309.17410},
}

Metadata

Added
2026-04-14
Added by
manual
Source
manual
arxiv_id
2309.17410