Files
www-project-ai-testing-guide/Document/content/3.1_AI_Application_Testing.md
2025-11-15 16:57:29 +01:00

3.5 KiB

3.1 AI Application Testing

The AI Application Testing category addresses security, safety, and trust risks arising specifically from interactions between AI systems, end-users, and external data sources. This testing category evaluates the behavior of AI applications when processing user inputs, generating outputs, and handling runtime interactions, with the goal of uncovering and mitigating vulnerabilities unique to AI-driven interactions, such as prompt injections, sensitive data leaks, and unsafe or biased outputs.

Given the direct exposure of AI applications to users and external environments, testing at this layer is critical to prevent unauthorized access, manipulation of AI behavior, and compliance violations. The category covers comprehensive evaluation against well-defined threat scenarios, including adversarial prompt manipulation, unsafe outputs, agentic misbehavior, and risks related to model extraction or embedding manipulation.

Scope of This Testing Category

This category evaluates whether the AI application:

Each test within the AI Application Testing category contributes to the holistic security posture of AI systems by systematically addressing application-level risks, ensuring robust operation in real-world environments, and helping organizations comply with ethical standards and regulations.