Prisma AIRS
AI Red Teaming
Table of Contents
Expand All
|
Collapse All
Prisma AIRS Docs
AI Red Teaming
Learn about new features for Prisma AIRS AI Red Teaming.
Here are the new Prisma AIRS AI Red Teaming features.
Enhanced AI Red Teaming with Brand and Reputational Risk Detection
|
January 2026
Supported for:
|
You can now assess and protect your AI systems against brand and
reputational risks using Prisma AIRS enhanced AI Red Teaming capabilities. This
feature addresses a critical gap in AI security by identifying vulnerabilities that
could damage your organization's reputation when AI systems interact with users in
production environments. Beyond the existing security, safety, and compliance risk
categories, you can now scan for brand and reputational threats including disallowed
topics, competitor endorsements, brand tarnishing content, discriminating claims,
and political endorsements.
When you configure your AI Red Teaming assessment, it leverages
comprehensive target profiling that considers your
industry context, specific use cases, competitive landscape, and company policies to
generate contextually relevant attack scenarios. This approach ensures that
vulnerability testing aligns with your organization's unique brand requirements and
regulatory constraints. You can provide company policy documents and other relevant
materials to further customize the assessment parameters, enabling more precise
detection of content that could violate your brand guidelines or corporate
standards.
The enhanced agent assessment capabilities automatically generate goals
focused on brand and reputational risk scenarios, particularly around disallowed
topics that could expose your organization to public relations challenges or
regulatory scrutiny. You benefit from specialized evaluation methods designed to
detect subtle forms of reputational risk, including false claims and inappropriate
endorsements that traditional security scanning might miss. This comprehensive
approach allows you to proactively identify and address potential brand
vulnerabilities before deploying AI systems to production environments, protecting
both your technical infrastructure and corporate reputation in an increasingly
AI-driven business landscape.
Advanced Target Profiling for Context-Aware AI Red Teaming
|
January 2026
Supported for:
|
Target Profiling enhances your AI security assessments by automatically
gathering comprehensive contextual information about your AI endpoints, enabling
more accurate and relevant vulnerability discoveries. When you conduct AI Red
Teaming assessments without proper context, you receive only generic baseline risk
evaluations that may not reflect real-world threats specific to your environment.
With Target Profiling, you can leverage both
user-provided information and intelligent agent-based discovery to build detailed
profiles of your AI models, applications, and agents.
Target Profiling automatically collects critical background information
about your AI systems, including industry context, use cases, competitive landscape,
and technical foundations such as base models, architecture patterns, and
accessibility requirements. AI Red Teaming's agentic profiling capability
interrogates your endpoints to discover configuration details like rate limiting,
guardrails, and system prompts without requiring manual input. This automated
approach saves you time while ensuring comprehensive coverage of contextual factors
that influence security risks.
The feature provides you with a centralized Target Profile page where you
can visualize all gathered context, review assessment history, and track risk scores
across multiple scans over time. You can distinguish between user-provided
information and agent-discovered data, giving you full transparency into how your
target profiles are constructed. When you need to update target information due to
system changes, you can easily modify profiles or trigger new agentic profiling
sessions.
Target Profiling directly improves your AI Red Teaming effectiveness by
enabling context-aware assessments that identify vulnerabilities specific to your
industry, use case, and technical implementation. AI Red Teaming uses your target's
industry and competitive context to evaluate brand and reputational risks more
accurately, while technical configuration details help identify
implementation-specific vulnerabilities. By maintaining detailed profiles and
assessment histories, you can track your security posture improvements over time and
ensure that your AI systems remain protected as they evolve in production
environments.
Enhanced AI Red Teaming for AI Agents and Multi-Agent Systems
|
January 2026
Supported for:
|
You can now leverage Prisma AIRS AI Red Teaming's enhanced capabilities to
comprehensively assess the security posture of your autonomous AI agents and
multi-agent systems. As your organization deploys agentic systems that extend beyond
traditional AI applications to include tool calling, instruction execution, and
system interactions, you face an expanded and more complex attack surface that
requires specialized security assessment approaches. This advanced AI Red Teaming
solution addresses the unique vulnerabilities inherent in both pro-code agents and
supported no-code/low-code frameworks by employing agent-led testing methodologies that craft
targeted goals and attacks specifically designed to exploit agentic system
weaknesses.
When you configure your AI Red Teaming assessments, the system
automatically tailors its approach based on your target endpoint type, enabling you
to uncover critical vulnerabilities such as tool misuse where malicious actors
manipulate your AI agents to abuse their integrated tools through deceptive prompts
while operating within authorized permissions. The solution also identifies intent
breaking and goal manipulation vulnerabilities where attackers redirect your agent's
objectives and reasoning to perform unintended tasks. Through comprehensive target
profiling, the agent-led AI Red Teaming capability gathers complete contextual
information about your endpoints and develops sophisticated attack scenarios that
traditional security testing approaches might miss.
Your security reports now provide enhanced visibility into agent-specific
vulnerabilities versus generic security issues, with risk scoring algorithms adapted
to properly weight the unique threats facing agentic systems. The automated
reporting includes specialized summaries and recommendations that emphasize the
specific vulnerabilities discovered in your agentic deployments, helping you
understand not just what risks exist but how they relate to your broader AI security
strategy. This targeted approach ensures you can confidently deploy AI agents in
production environments while maintaining robust security controls against the
evolving threat landscape targeting autonomous AI systems.
Remediation Recommendations for AI Red Teaming Risk Assessment
|
December 2025
Supported for:
|
The Remediation Recommendations feature enables you to seamlessly
transition from identifying AI system vulnerabilities through Red Teaming
assessments to implementing targeted runtime policy configurations that address
your specific risks.
When you conduct AI Red Teaming evaluations on your AI models,
applications, or agents, this integrated solution automatically analyzes the
discovered security, safety, brand reputation, and compliance risks to generate
contextual remediation recommendations that directly address your specific
vulnerabilities. Rather than configuring runtime security policies through trial and
error, you receive intelligent guidance that maps each identified risk category to
appropriate guardrail configurations, such as enabling prompt injection protection
for security vulnerabilities or activating toxic content moderation for safety
concerns.
You can leverage this capability to close the critical gap between risk
assessment and mitigation in your AI deployment lifecycle. The feature provides you
with detailed security profile recommendations that specify optimal runtime
protection settings based on your AI Red Teaming results, eliminating the guesswork
typically associated with configuring AI security controls. For organizations
deploying AI systems in production environments, this capability ensures that your
runtime security configurations are informed by actual risk insights rather than
generic best practices, resulting in more effective protection against the specific
threats your AI systems face.
The remediation recommendations appear directly in your AI Red Teaming
reports, providing you with actionable guidance on creating appropriate security
profiles with the necessary parameters to protect against identified threats. You
can manually create and attach these recommended security profiles to your desired
workloads, transforming AI risk management from a reactive, disconnected process
into a proactive workflow that directly connects vulnerability discovery with
targeted protection measures.
AI Red Teaming Executive Reports
|
December 2025
Supported for:
|
You can now generate executive-ready AI Red Teaming assessment reports that provide
comprehensive security insights tailored for top executives of your organizations
((like CEO, CFO, COO, CTO). This enhanced reporting capability transforms technical
vulnerability data into strategic intelligence that you can easily share with Chief
Information Security Officers (CISOs), Chief Information Officers (CIOs), and other
executives who need to understand your organization's security posture at a high
level. The feature leverages advanced LLM technology to automatically generate
concise summaries that distill complex attack simulation results into clear verdicts
about your target endpoint's safety, security, and compliance alignment, while
identifying the most critical attack vectors and their business implications.
When you complete an AI Red Teaming scan, you receive an AI-powered
overview that synthesizes target profiling data with vulnerability findings to
provide contextual insights about your security risks. This executive summary
eliminates the need for manual interpretation of technical data, allowing you to
quickly understand which attack methods pose the greatest threats to your systems
and what the potential business impact might be. You can then export these
comprehensive insights as professional PDF reports that maintain the detailed
technical information security teams require while presenting the strategic overview
that executives need for decision-making.
This capability is particularly valuable when you need to communicate
security assessment results across different organizational levels or when preparing
briefings for leadership meetings. Rather than struggling to translate technical
vulnerability reports into business language, you can rely on the AI Red Teaming
generated executive report to articulate security, safety, compliance, brand, and
business risks in terms that resonate with executive audiences.
The exportable PDF format ensures you can easily share findings in board
presentations, compliance documentation, or strategic planning sessions while
maintaining the granular attack details that technical teams use for remediation
efforts.
Error Logs and Partial Scan Reports
|
December 2025
Supported for:
|
When you conduct AI Red Teaming scans using Prisma AIRS, you may encounter
situations where scans fail completely or complete only partially due to target
system issues or connectivity problems. The Error Logs and Partial Scan Reports feature provides you
with comprehensive visibility into scan failures and enables you to generate
actionable reports even when your scans don't complete successfully. You can access
detailed error logs directly within the scan interface, both during active scans on
the progress page and after completion in the scan logs section, allowing you to
quickly identify whether issues stem from your target AI system or the Prisma AIRS
platform itself.
This feature particularly benefits you when conducting Red Teaming
assessments against enterprise AI systems that may have intermittent availability or
response issues. When your scan completes the full simulation but doesn’t receive
valid responses for all attacks, AI Red Teaming marks it as partially complete
rather than failed. You can then choose to generate a comprehensive report based on
the available test results, giving you valuable security insights even from
incomplete assessments. AI Red Teaming transparently informs you about credit
consumption before report generation and clearly marks any generated reports as
partial scans, indicating the percentage of attacks that received responses.
By leveraging this capability, you can maximize the value of your Red
Teaming efforts, troubleshoot scanning issues more effectively, and maintain
continuous security assessment workflows even when facing target system limitations
or temporary connectivity challenges during your AI security evaluations.
Automated AI Red Teaming
|
October 2025
Supported for:
|
Palo Alto Networks' is an automated solution designed to scan any AI
system—including LLMs and LLM-powered applications—for safety and security
vulnerabilities.
The tool performs a Scan against a specified Target (model,
application, or agent) by sending carefully crafted attack prompts to
simulate real-world threats. The findings are compiled into a comprehensive Scan
Report that includes an overall Risk Score (ranging from 0 to 100),
indicating the system's susceptibility to attacks.
Prisma AIRS offers three distinct scanning modes for thorough assessment:
- Attack Library Scan: Uses a curated, proprietary library of predefined attack scenarios, categorized by Security (e.g., Prompt Injection, Jailbreak), Safety (e.g., Bias, Cybercrime), and Compliance (e.g., OWASP LLM Top 10).
- Agent Scan: Utilizes a dynamic LLM attacker to generate and adapt attacks in real-time, enabling full-spectrum Black-box, Grey-box, and White-box testing.
- Custom Attack Scan: Allows users to upload and execute their own custom prompt sets alongside the built-in library.
A key feature of the service is its single-tenant deployment model, which
ensures complete isolation of compute resources and data for enhanced security and
privacy.