← Back to search
paper reviewed open access llmsec-2025-00024
Can Sensitive Information Be Deleted From LLMs? Objectives for Defending Against Extraction Attacks
Vaidehi Patil, Peter Hase, Mohit Bansal
2024 — ICLR 2024 70 citations
Abstract
Evaluates methods for deleting sensitive information from trained LLMs, finding current unlearning approaches insufficient against determined adversaries.
Categories
Tags
knowledge-deletionunlearningextraction-defense
Framework Mappings
OWASP LLM: LLM02
Cite This Resource
@article{llmsec202500024,
title = {Can Sensitive Information Be Deleted From LLMs? Objectives for Defending Against Extraction Attacks},
author = {Vaidehi Patil and Peter Hase and Mohit Bansal},
year = {2024},
journal = {ICLR 2024},
url = {https://arxiv.org/abs/2309.17410},
} Metadata
- Added
- 2026-04-14
- Added by
- manual
- Source
- manual
- arxiv_id
- 2309.17410