Detect Malicious Code in LLM Outputs
Focus
Focus
What's New in the NetSec Platform

Detect Malicious Code in LLM Outputs

Table of Contents

Detect Malicious Code in LLM Outputs

Malicious Code Detection scans LLM-generated code snippets across multiple languages to prevent security threats and supply chain attacks.
Code snippets generated by Large Language Models (LLMs) can be protected with Malicious Code Detection feature for potential security threats. This feature is crucial for preventing supply chain attacks, enhancing application security, maintaining code integrity, and mitigating AI risks.
The system supports scanning for malicious code in multiple languages, including JavaScript, Python, VBScript, PowerShell, Batch, Shell, and Perl.
To activate this protection, you need to enable it within the API Security Profile. When configured, this feature can block the execution of potentially malicious code or be set to allow, depending on your security needs. This capability is vital for organizations that are increasingly leveraging generative AI for development, as it helps to secure against the risks of LLM poisoning, where adversaries intentionally introduce malicious data into training datasets to manipulate model outputs.