Skip to content

Commit b9584bf

Browse files
committed
init
0 parents  commit b9584bf

15 files changed

+4855
-0
lines changed

.github/workflows/main.yml

+39
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,39 @@
1+
name: Docs to PDF
2+
# This workflow is triggered on pushes to the repository.
3+
on:
4+
push:
5+
branches:
6+
- main
7+
# Paths can be used to only trigger actions when you have edited certain files, such as a file within the /docs directory
8+
paths:
9+
- 'docs/**.md'
10+
- 'docs/images/**'
11+
12+
jobs:
13+
converttopdf:
14+
name: Build PDF
15+
runs-on: ubuntu-latest
16+
steps:
17+
- uses: actions/checkout@v4
18+
- uses: baileyjm02/markdown-to-pdf@v1
19+
with:
20+
input_dir: docs
21+
output_dir: output/
22+
images_dir: docs/images
23+
# for example <img src="./images/file-name.png">
24+
image_import: ./images
25+
# Default is true, can set to false to only get PDF files
26+
build_html: true
27+
- uses: actions/upload-artifact@v4
28+
with:
29+
name: docs
30+
path: output
31+
- name: Commit and push output files
32+
env:
33+
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
34+
run: |
35+
git config --global user.name 'github-actions[bot]'
36+
git config --global user.email '41898282+github-actions[bot]@users.noreply.github.com'
37+
git add .
38+
git commit -m "[Auto] Add generated PDF and HTML files"
39+
git push origin main

LICENSE.txt

+31
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,31 @@
1+
Open LLM Security Benchmark is provided under the 3-clause BSD license below.
2+
3+
*************************************************************
4+
5+
Copyright (c) 2024, NetSPI
6+
All rights reserved.
7+
8+
Redistribution and use in source and binary forms, with or without
9+
modification, are permitted provided that the following conditions are met:
10+
11+
* Redistributions of source code must retain the above copyright notice, this
12+
list of conditions and the following disclaimer.
13+
14+
* Redistributions in binary form must reproduce the above copyright notice,
15+
this list of conditions and the following disclaimer in the documentation
16+
and/or other materials provided with the distribution.
17+
18+
* Neither the name of Open LLM Security Benchmark nor the names of its
19+
contributors may be used to endorse or promote products derived from
20+
this software without specific prior written permission.
21+
22+
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
23+
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24+
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
25+
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
26+
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
27+
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
28+
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
29+
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
30+
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31+
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

README.md

+133
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,133 @@
1+
![Banner](./docs/images/open-LLM-security-benchmark-banner.png)
2+
3+
[![licence badge]][licence]
4+
[![stars badge]][stars]
5+
[![forks badge]][forks]
6+
[![issues badge]][issues]
7+
![Twitter Follow](https://img.shields.io/twitter/follow/NetSPI.svg?style=social)
8+
9+
10+
[licence badge]:https://img.shields.io/badge/license-New%20BSD-blue.svg
11+
[stars badge]:https://img.shields.io/github/stars/NetSPI/Open-LLM-Security-Benchmark.svg
12+
[forks badge]:https://img.shields.io/github/forks/NetSPI/Open-LLM-Security-Benchmark.svg
13+
[issues badge]:https://img.shields.io/github/issues/NetSPI/Open-LLM-Security-Benchmark.svg
14+
15+
16+
[licence]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/blob/master/LICENSE.txt
17+
[stars]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/stargazers
18+
[forks]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/network
19+
[issues]:https://github.com/NetSPI/Open-LLM-Security-Benchmark/issues
20+
21+
22+
# NetSPI’s Open LLM Security Benchmark: Balancing Security & Usability of Large Language Models (LLMs)
23+
24+
## Download the PDF: [open-llm-security-benchmark.pdf](./open-LLM-Security-Benchmark.pdf)
25+
26+
## TL;DR
27+
28+
- Large Language Models (LLMs) have become more integrated into critical systems, applications, and processes, posing a
29+
challenge for potential security risks.
30+
31+
- Increasing security measures in LLMs can negatively affect usability, requiring the right balance. But these behaviors
32+
may be desired depending on the business use case.
33+
34+
- Our LLM benchmarking framework shows how different LLMs handle adversarial conditions, testing their jailbreakability,
35+
while measuring any impact on usability.
36+
37+
## Security Concerns
38+
39+
As LLMs are integrated into critical systems, vulnerabilities like jailbreaks, model extraction, and data leakage pose
40+
growing risks like harmful content generation, data exposure, or unauthorized actions. These threats can lead to
41+
proprietary data loss, reputational harm, and legal issues, emphasizing the urgent need for rigorous benchmarks to
42+
assess and improve their security.
43+
44+
## Balancing Security and Usability
45+
46+
While enhancing security of an LLM is important, usability is equally important. The model should still perform its
47+
intended functions effectively. Oftentimes, security and usability is a balancing act. This challenge is well-documented
48+
in software and system design – overly strict filters may limit useful responses, while insufficient security poses
49+
risks.
50+
51+
## LLM Benchmarking Framework
52+
53+
These challenges and concerns are not going away anytime soon. So, what can be done? We’ve created a benchmarking
54+
framework that evaluates both the security and usability of LLMs. Our systematic assessment shows how different LLMs
55+
handle adversarial conditions, testing their jailbreakability, while measuring any impact on usability. This dual
56+
evaluation helps balance security with functionality, crucial for AI applications in cybersecurity.
57+
58+
Our intent is that the benchmark can provide some level of transparency so that it can be used by organizations to make
59+
more informed choices that better align to their use cases and risk appetite.
60+
61+
## How We Approached LLM Security & Usability Research
62+
63+
1. Understand Current Benchmarks: Our research reflects the current understanding of LLM security and usability using
64+
established benchmarks.
65+
66+
2. Recognize Ongoing Evolution: This work is part of a growing field. As new techniques and security practices emerge,
67+
benchmarks will be refined and expanded.
68+
69+
3. Engage with Feedback: Reader feedback and constructive critiques are welcomed to improve the methodology's robustness
70+
and scope.
71+
72+
4. Commit to High Standards: We remain dedicated to maintaining the highest standards in evaluation as the field
73+
advances.
74+
75+
## How to Participate
76+
77+
### Providing Feedback
78+
79+
If you have feedback on the paper, please create an issue on GitHub with the following information:
80+
81+
1. A clear and descriptive title.
82+
83+
2. A detailed description of your feedback.
84+
85+
3. Specific sections or lines of the paper that your feedback pertains to.
86+
87+
4. Any relevant suggestions or improvements.
88+
89+
### Suggesting Enhancements
90+
91+
We welcome suggestions for new sections, topics, or improvements to the paper. To suggest an enhancement, please create
92+
an issue on GitHub with the following information:
93+
94+
1. A clear and descriptive title.
95+
96+
2. A detailed description of the proposed enhancement.
97+
98+
3. Any relevant examples, references, or mockups.
99+
100+
### Writing and Editing
101+
102+
We appreciate contributions to the writing and editing of the paper. If you would like to contribute, please follow
103+
these steps:
104+
105+
1. Fork the repository.
106+
107+
2. Create a new branch (git checkout -b feature/your-feature-name).
108+
109+
3. Make your changes to the markdown file [open-LLM-security-benchmark](./docs/open-LLM-security-benchmark.md).
110+
111+
4. Commit your changes (git commit -m 'Add new section on topic').
112+
113+
5. Push to the branch (git push origin feature/your-feature-name).
114+
115+
6. Open a pull request on GitHub.
116+
117+
Please ensure your pull request includes:
118+
119+
1. A clear and descriptive title.
120+
121+
2. A detailed description of the changes.
122+
123+
3. Any relevant issue numbers (e.g., "Addresses feedback from #123").
124+
125+
126+
127+
128+
129+
130+
131+
132+
133+
Loading

docs/images/aaa-vs-breakability.png

195 KB
Loading
98.4 KB
Loading

docs/images/agentic-system.png

108 KB
Loading
Loading

docs/images/usability-formula.png

7.93 KB
Loading
169 KB
Loading

0 commit comments

Comments
 (0)