|
| 1 | + |
| 2 | + |
| 3 | +[![licence badge]][licence] |
| 4 | +[![stars badge]][stars] |
| 5 | +[![forks badge]][forks] |
| 6 | +[![issues badge]][issues] |
| 7 | + |
| 8 | + |
| 9 | + |
| 10 | +[licence badge]:https://img.shields.io/badge/license-New%20BSD-blue.svg |
| 11 | +[stars badge]:https://img.shields.io/github/stars/NetSPI/Open-LLM-Security-Benchmark.svg |
| 12 | +[forks badge]:https://img.shields.io/github/forks/NetSPI/Open-LLM-Security-Benchmark.svg |
| 13 | +[issues badge]:https://img.shields.io/github/issues/NetSPI/Open-LLM-Security-Benchmark.svg |
| 14 | + |
| 15 | + |
| 16 | +[licence]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/blob/master/LICENSE.txt |
| 17 | +[stars]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/stargazers |
| 18 | +[forks]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/network |
| 19 | +[issues]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/issues |
| 20 | + |
| 21 | + |
| 22 | +# NetSPI’s Open LLM Security Benchmark: Balancing Security & Usability of Large Language Models (LLMs) |
| 23 | + |
| 24 | +## Download the PDF: [open-llm-security-benchmark.pdf](./open-LLM-Security-Benchmark.pdf) |
| 25 | + |
| 26 | +## TL;DR |
| 27 | + |
| 28 | +- Large Language Models (LLMs) have become more integrated into critical systems, applications, and processes, posing a |
| 29 | + challenge for potential security risks. |
| 30 | + |
| 31 | +- Increasing security measures in LLMs can negatively affect usability, requiring the right balance. But these behaviors |
| 32 | + may be desired depending on the business use case. |
| 33 | + |
| 34 | +- Our LLM benchmarking framework shows how different LLMs handle adversarial conditions, testing their jailbreakability, |
| 35 | + while measuring any impact on usability. |
| 36 | + |
| 37 | +## Security Concerns |
| 38 | + |
| 39 | +As LLMs are integrated into critical systems, vulnerabilities like jailbreaks, model extraction, and data leakage pose |
| 40 | +growing risks like harmful content generation, data exposure, or unauthorized actions. These threats can lead to |
| 41 | +proprietary data loss, reputational harm, and legal issues, emphasizing the urgent need for rigorous benchmarks to |
| 42 | +assess and improve their security. |
| 43 | + |
| 44 | +## Balancing Security and Usability |
| 45 | + |
| 46 | +While enhancing security of an LLM is important, usability is equally important. The model should still perform its |
| 47 | +intended functions effectively. Oftentimes, security and usability is a balancing act. This challenge is well-documented |
| 48 | +in software and system design – overly strict filters may limit useful responses, while insufficient security poses |
| 49 | +risks. |
| 50 | + |
| 51 | +## LLM Benchmarking Framework |
| 52 | + |
| 53 | +These challenges and concerns are not going away anytime soon. So, what can be done? We’ve created a benchmarking |
| 54 | +framework that evaluates both the security and usability of LLMs. Our systematic assessment shows how different LLMs |
| 55 | +handle adversarial conditions, testing their jailbreakability, while measuring any impact on usability. This dual |
| 56 | +evaluation helps balance security with functionality, crucial for AI applications in cybersecurity. |
| 57 | + |
| 58 | +Our intent is that the benchmark can provide some level of transparency so that it can be used by organizations to make |
| 59 | +more informed choices that better align to their use cases and risk appetite. |
| 60 | + |
| 61 | +## How We Approached LLM Security & Usability Research |
| 62 | + |
| 63 | +1. Understand Current Benchmarks: Our research reflects the current understanding of LLM security and usability using |
| 64 | + established benchmarks. |
| 65 | + |
| 66 | +2. Recognize Ongoing Evolution: This work is part of a growing field. As new techniques and security practices emerge, |
| 67 | + benchmarks will be refined and expanded. |
| 68 | + |
| 69 | +3. Engage with Feedback: Reader feedback and constructive critiques are welcomed to improve the methodology's robustness |
| 70 | + and scope. |
| 71 | + |
| 72 | +4. Commit to High Standards: We remain dedicated to maintaining the highest standards in evaluation as the field |
| 73 | + advances. |
| 74 | + |
| 75 | +## How to Participate |
| 76 | + |
| 77 | +### Providing Feedback |
| 78 | + |
| 79 | +If you have feedback on the paper, please create an issue on GitHub with the following information: |
| 80 | + |
| 81 | +1. A clear and descriptive title. |
| 82 | + |
| 83 | +2. A detailed description of your feedback. |
| 84 | + |
| 85 | +3. Specific sections or lines of the paper that your feedback pertains to. |
| 86 | + |
| 87 | +4. Any relevant suggestions or improvements. |
| 88 | + |
| 89 | +### Suggesting Enhancements |
| 90 | + |
| 91 | +We welcome suggestions for new sections, topics, or improvements to the paper. To suggest an enhancement, please create |
| 92 | +an issue on GitHub with the following information: |
| 93 | + |
| 94 | +1. A clear and descriptive title. |
| 95 | + |
| 96 | +2. A detailed description of the proposed enhancement. |
| 97 | + |
| 98 | +3. Any relevant examples, references, or mockups. |
| 99 | + |
| 100 | +### Writing and Editing |
| 101 | + |
| 102 | +We appreciate contributions to the writing and editing of the paper. If you would like to contribute, please follow |
| 103 | +these steps: |
| 104 | + |
| 105 | +1. Fork the repository. |
| 106 | + |
| 107 | +2. Create a new branch (git checkout -b feature/your-feature-name). |
| 108 | + |
| 109 | +3. Make your changes to the markdown file [open-LLM-security-benchmark](./docs/open-LLM-security-benchmark.md). |
| 110 | + |
| 111 | +4. Commit your changes (git commit -m 'Add new section on topic'). |
| 112 | + |
| 113 | +5. Push to the branch (git push origin feature/your-feature-name). |
| 114 | + |
| 115 | +6. Open a pull request on GitHub. |
| 116 | + |
| 117 | +Please ensure your pull request includes: |
| 118 | + |
| 119 | +1. A clear and descriptive title. |
| 120 | + |
| 121 | +2. A detailed description of the changes. |
| 122 | + |
| 123 | +3. Any relevant issue numbers (e.g., "Addresses feedback from #123"). |
| 124 | + |
| 125 | + |
| 126 | + |
| 127 | + |
| 128 | + |
| 129 | + |
| 130 | + |
| 131 | + |
| 132 | + |
| 133 | + |
0 commit comments