← Back to search
paper reviewed open access llmsec-2024-00052

Purple Llama CyberSecEval: A Secure Coding Benchmark for Language Models

Manish Bhatt, Sahana Chennabasappa, Cyrus Nikolaidis, Shengye Wan, Ivan Evtimov

2024 — arXiv preprint 140 citations

Abstract

Introduces CyberSecEval, a benchmark for evaluating the cybersecurity risks of LLM code generation, including insecure code suggestions.

Categories

Tags

CyberSecEvalcode-securityMetabenchmark

Framework Mappings

OWASP LLM: LLM05 NIST AI RMF: MEASURE

Cite This Resource

@article{llmsec202400052,
  title = {Purple Llama CyberSecEval: A Secure Coding Benchmark for Language Models},
  author = {Manish Bhatt and Sahana Chennabasappa and Cyrus Nikolaidis and Shengye Wan and Ivan Evtimov},
  year = {2024},
  journal = {arXiv preprint},
  url = {https://arxiv.org/abs/2312.04724},
}

Metadata

Added
2026-04-14
Added by
manual
Source
manual
arxiv_id
2312.04724