
Balancing Security and Usability of Large Language Models: An LLM Benchmarking Framework
By 2026, Gartner predicts that “80% of all enterprises will have used or deployed generative AI applications.” However, many of these organizations have yet to find a way to balance usability and security in their deployments. As a result, consumer-facing LLM capabilities introduce a new and less understood set of risks for organizations. The mission of this article, along with the first release of the NetSPI Open Large Language Model (LLM) Security Benchmark, is to clarify some of the ambiguity around LLM security and highlight the visible trade-offs between security and usability.
TLDR;
- Large Language Models (LLMs) have become more integrated into critical systems, applications, and processes, posing a challenge for potential security risks.
- Increasing security measures in LLMs can negatively affect usability, requiring the right balance. But these behaviors may be desired depending on the business use case.
- Our LLM benchmarking framework shows how different LLMs handle adversarial conditions, testing their jailbreakability, while measuring any impact on usability.
Security Concerns in Large Language Models
As LLMs become integral to critical systems, the risk of vulnerabilities like model extraction, data leakage, membership inference, direct prompt injection, and jailbreakability increases. Jailbreaking refers to manipulating a model to bypass safety filters, potentially generating harmful content, exposing sensitive data, or performing unauthorized actions.
These vulnerabilities have significant implications. In business, a compromised LLM could leak proprietary information or become an attack vector. In public applications, there is a risk of harmful or biased content causing reputational damage and legal issues. Therefore, ensuring LLM security is crucial, highlighting the need for robust benchmarks to test their resilience against attacks, including jailbreakability.
Balancing Security and Usability
While enhancing security of an LLM is important, usability is equally important. The model should still perform its intended functions effectively. Oftentimes, security and usability is a balancing act. This challenge is well-documented in software and system design – overly strict filters may limit useful responses, while insufficient security poses risks.
LLM Benchmarking Framework
These challenges and concerns are not going away anytime soon. So, what can be done? We’ve created a benchmarking framework that evaluates both the security and usability of LLMs. Our systematic assessment shows how different LLMs handle adversarial conditions, testing their jailbreakability, while measuring any impact on usability. This dual evaluation helps balance security with functionality, crucial for AI applications in cybersecurity.
Our intent is that the benchmark can provide some level of transparency so that it can be used by organizations to make more informed choices that better align to their use cases and risk appetite.
While the findings and benchmarks presented in this paper reflect our current understanding of LLM security and usability, it is important to note that this research is part of an evolving body of work. As advancements in model evaluation techniques and security practices emerge, we expect to refine and expand upon these benchmarks. We encourage feedback and constructive critique from readers, as it will help to further improve the robustness and comprehensiveness of our methodology. We remain committed to ensuring that these evaluations continue to meet the highest standards as the field develops.
We invite you to participate in this research and contribute your insights to the paper, helping shape the future of AI security.

Explore More Blog Posts

CVE-2025-21299 and CVE-2025-29809: Unguarding Microsoft Credential Guard
Learn more about the January 2025 Patch Tuesday that addresses a critical vulnerability where Kerberos canonicalization flaws allow attackers to bypass Virtualization Based Security and extract protected TGTs from Windows systems.

CVE-2025-27590 – Oxidized Web: Local File Overwrite to Remote Code Execution
Learn about a critical security vulnerability (CVE-2025-27590) in Oxidized Web v0.14 that allows attackers to overwrite local files and execute remote code execution.

Is It Worth It? Let Me Work It: Calculating the Cost Savings of Proactive Security
Discover the cost savings of proactive security solutions to support your shift from traditional vulnerability management to a risk-based approach to exposure management.