[ISM] LLM risk mitigation (OWASP top 10):

The OWASP Top 10 for Large Language Model Applications are mitigated in the development of large language model applications.

[OWASP] Protect and prevent prompt injection (SSS-02-05-01)

Ensure that large language models (LLMs) are protected from prompt injection vulnerabilities that can manipulate model behavior, bypass safety protocols, and generate unintended or harmful outputs. Implement input validation, privilege restrictions, and adversarial testing to minimize the risk of direct and indirect prompt injections.

[OWASP] Implement direct/indirect input validation and verification (SSS-02-05-01-01)

Establish strict controls on user input and model processing. Implement structured prompt validation to detect and reject adversarial prompts before they reach the model. Apply content filtering mechanisms, such as semantic analysis and string-based checks, to identify and block malicious inputs. Enforce least privilege access by restricting API tokens and external integrations to only necessary functions. Segregate external and untrusted content to prevent model behavior from being altered by indirect injections. Require human verification for high-risk actions where model outputs could lead to significant decisions. Conduct regular adversarial testing to simulate real-world attacks and continuously update safety protocols. Ensure that multimodal AI models, which handle different data types, have cross-modal security measures in place. Develop clear output formatting guidelines to prevent response manipulation and improve detection of injection attempts.

Operations

ID Operation Description Phase Agent
SSS-02-05-01-01-01 Implement input validation and sanitization Apply filtering mechanisms to detect adversarial prompts and remove potentially harmful instructions before processing. Development Security team, AI engineers
SSS-02-05-01-01-02 Enforce privilege control and access restrictions Limit the model’s access to external APIs and system functionalities by implementing role-based access control (RBAC) and API token segregation. Deployment Security team, Infrastructure team
SSS-02-05-01-01-03 Apply structured output formatting and validation Define expected output formats and use deterministic validation techniques to verify model responses before they are returned to users. Post-deployment AI engineers, Product team
SSS-02-05-01-01-04 Conduct adversarial testing and attack simulations Regularly perform security assessments by simulating real-world attacks to evaluate model vulnerabilities and improve response mechanisms. Post-deployment Security team, Red team, AI engineers
SSS-02-05-01-01-05 Segregate and identify external content sources Clearly label and separate trusted and untrusted content to prevent unauthorized influence on model responses. Development AI engineers, Legal team

References

Industry framework Academic work Real-world case
Information Security Manual (ISM-1923)
OWASP Top 10 for LLM
OWASP Top 10 for LLM (LLM01:2025)