← Back to search
paper reviewed open access llmsec-2024-00052
Purple Llama CyberSecEval: A Secure Coding Benchmark for Language Models
Manish Bhatt, Sahana Chennabasappa, Cyrus Nikolaidis, Shengye Wan, Ivan Evtimov
2024 — arXiv preprint 140 citations
Abstract
Introduces CyberSecEval, a benchmark for evaluating the cybersecurity risks of LLM code generation, including insecure code suggestions.
Framework Mappings
OWASP LLM: LLM05 NIST AI RMF: MEASURE
Cite This Resource
@article{llmsec202400052,
title = {Purple Llama CyberSecEval: A Secure Coding Benchmark for Language Models},
author = {Manish Bhatt and Sahana Chennabasappa and Cyrus Nikolaidis and Shengye Wan and Ivan Evtimov},
year = {2024},
journal = {arXiv preprint},
url = {https://arxiv.org/abs/2312.04724},
} Metadata
- Added
- 2026-04-14
- Added by
- manual
- Source
- manual
- arxiv_id
- 2312.04724