conf.json alongside your provider credentials, then referenced in configs using the <plugin-id>.<functionId> format.
Enabling a partner plugin
- Add the plugin ID to
plugins_enabledinconf.json. - Add your credentials under the matching key in
credentials. - Run
npm run build-pluginsto recompile. - Reference the function in your config.
conf.json
Security & content moderation
Aporia
Plugin ID:
aporiaMitigates LLM hallucinations, inappropriate responses, prompt injection attacks, and other unintended behaviors in real-time by validating requests and responses against your Aporia project policies.Credentials: apiKey — find it in the Aporia dashboard.Functions:aporia.validateProject— checks all policies in a given Aporia project (beforeRequestHook,afterRequestHook)
Pangea
Plugin ID:
pangeaPangea AI Guard scans LLM inputs and outputs for malicious content, prompt manipulation, and sensitive data. Supports recipe-based configuration via the Pangea User Console.Credentials: apiKey (AI Guard token), domain (e.g. aws.us.pangea.cloud).Functions:pangea.textGuard— AI Guard scan with recipe support and optional redaction (beforeRequestHook,afterRequestHook)pangea.pii— PII Guard with optional redaction (beforeRequestHook,afterRequestHook)
Pillar
Plugin ID:
pillarPillar Security scans prompts and responses for prompt injection, PII, secrets, toxic language, and invisible characters.Credentials: apiKey — find it in the Pillar dashboard.Functions:pillar.scanPrompt— scans user prompts (beforeRequestHook)pillar.scanResponse— scans model responses (afterRequestHook)
Walled AI
Plugin ID:
walledaiEnsures the safety and compliance of LLM inputs and outputs with configurable PII, safety, and compliance checks.Credentials: apiKey — find it in the Walled AI dashboard.Functions:walledai.walledprotect— safety and compliance check (beforeRequestHook,afterRequestHook)
Prompt Security
Plugin ID:
promptsecurityProtects user prompts before they reach the LLM and sanitizes model responses before they reach the user.Credentials: apiDomain (API domain), apiKey.Functions:promptsecurity.protectPrompt— guards the user prompt (beforeRequestHook)promptsecurity.protectResponse— guards the LLM response (afterRequestHook)
Acuvity
Plugin ID:
acuvityComprehensive content safety and security checks with configurable thresholds for prompt injection, toxicity, jailbreak, malicious URLs, bias, harmful content, PII, and secrets detection.Credentials: apiKey — find it in the Acuvity dashboard.Functions:acuvity.Acuvity— multi-category scan with per-category thresholds (beforeRequestHook,afterRequestHook)
Lasso Security
Plugin ID:
lassoGenAI security platform that autonomously monitors interactions and detects risks in real-time using AI-powered Deputies for jailbreaks, custom policy violations, hate speech, illegal content, and more.Credentials: apiKey — find it in the Lasso dashboard.Functions:lasso.classify— classifies prompts for security risks (beforeRequestHook)
Qualifire
Plugin ID:
qualifireContent moderation, hallucination detection, PII checks, prompt injection detection, grounding checks, and tool use quality evaluation.Credentials: apiKey — create one in Qualifire settings.Functions:qualifire.contentModeration— dangerous/sexual/harassment content (beforeRequestHook,afterRequestHook)qualifire.hallucinations— hallucination detection (afterRequestHook)qualifire.pii— PII detection (beforeRequestHook,afterRequestHook)qualifire.promptInjections— injection detection (beforeRequestHook)qualifire.grounding— context grounding (afterRequestHook)qualifire.toolUseQuality— tool call quality (afterRequestHook)qualifire.policy— custom policy violations (beforeRequestHook,afterRequestHook)
Cloud provider guardrails
Azure AI Content Safety
Plugin ID:
azureAzure AI Content Safety and PII Detection with support for API key, Entra ID, and Managed Identity authentication.Credentials: Per-function credentials block with azureAuthMode, resourceName, and either apiKey or Entra/Managed Identity fields.Functions:azure.contentSafety— hate, self-harm, sexual, violence detection (beforeRequestHook,afterRequestHook)azure.pii— PII detection with optional redaction (beforeRequestHook,afterRequestHook)azure.shieldPrompt— jailbreak and prompt injection detection via Prompt Shields (beforeRequestHook)azure.protectedMaterial— detects copyrighted content in LLM outputs (afterRequestHook)
conf.json credentials block
AWS Bedrock Guardrails
Plugin ID:
bedrockApplies AWS Bedrock Guardrails to filter harmful content, prompt injections, and PII, using guardrail configurations managed in your AWS account.Credentials: accessKeyId, accessKeySecret, region, optionally awsSessionToken.Functions:bedrock.guard— applies a named Bedrock guardrail (beforeRequestHook,afterRequestHook)
LLM-powered evaluation
Patronus AI
Plugin ID:
patronusLeading LLM automated testing and evaluation platform. Checks for PHI, PII, toxicity, bias, hallucination, and behavioral quality (conciseness, helpfulness, politeness).Credentials: apiKey — find it in the Patronus dashboard.Functions:patronus.phi— Protected Health Information detection with optional redaction (afterRequestHook)patronus.pii— PII detection with optional redaction (afterRequestHook)patronus.isConcise— conciseness check (afterRequestHook)patronus.isHelpful— helpfulness check (afterRequestHook)patronus.isPolite— politeness check (afterRequestHook)patronus.noApologies— no-apology check (afterRequestHook)patronus.noGenderBias— gender bias check (afterRequestHook)patronus.noRacialBias— racial bias check (afterRequestHook)patronus.retrievalAnswerRelevance— answer relevance check (afterRequestHook)patronus.retrievalHallucination— hallucination check (afterRequestHook)patronus.toxicity— toxicity/hate speech check (afterRequestHook)patronus.custom— custom evaluator using a named profile (afterRequestHook)
Mistral Moderation
Plugin ID:
mistralMistral’s content moderation classifier covering sexual content, hate/discrimination, violence, dangerous/criminal content, self-harm, health, financial, legal, and PII categories.Credentials: apiKey — find it in Mistral la-plateforme.Functions:mistral.moderateContent— multi-category moderation check (beforeRequestHook,afterRequestHook)
Promptfoo
Plugin ID:
promptfooPromptfoo’s Red Team and Guardrails API detects security risks, PII, and harmful content. No API credentials required.Credentials: None required.Functions:promptfoo.guard— injection and jailbreak detection (beforeRequestHook)promptfoo.pii— PII detection with optional redaction (beforeRequestHook,afterRequestHook)promptfoo.harm— harmful content detection (beforeRequestHook,afterRequestHook)
Portkey Advanced
Plugin ID:
portkeyAdvanced guardrails from Portkey including content moderation (via OpenAI moderation API), language detection, PII detection with redaction, and gibberish detection. Supports 50+ languages.Credentials: apiKey, optionally baseURL for a custom deployment.Functions:portkey.moderateContent— multi-category moderation (beforeRequestHook)portkey.language— language detection for 50+ languages (beforeRequestHook)portkey.pii— PII detection with optional redaction for 7 categories (beforeRequestHook,afterRequestHook)portkey.gibberish— gibberish detection (beforeRequestHook,afterRequestHook)
Enterprise security
CrowdStrike AIDR
Plugin ID:
crowdstrike-aidrCrowdStrike AI Detection and Response scans LLM inputs and outputs, blocking or sanitizing content according to rules configured in your Falcon console.Credentials: apiKey (AIDR token from the Falcon console), baseUrl.Functions:crowdstrike-aidr.guardChatCompletions— guards chat completions (beforeRequestHook,afterRequestHook)
Palo Alto Networks Prisma AIRS
Plugin ID:
panwPrismaAirsPalo Alto Networks Prisma AI Runtime Security provides real-time scanning for prompt injections, malicious content, PII leakage, and policy violations using security profiles from Strata Cloud Manager.Credentials: AIRS_API_KEY — find it in Strata Cloud Manager.Functions:panwPrismaAirs.intercept— scans prompts and responses against a named AIRS profile (beforeRequestHook,afterRequestHook)
F5 Guardrails
Plugin ID:
f5-guardrailsF5 Guardrails powered by Calypso AI provides advanced content moderation and PII detection with optional redaction.Credentials: apiKey, optionally calypsoUrl (defaults to https://us1.calypsoai.app).Functions:f5-guardrails.scan— content moderation and PII detection (beforeRequestHook,afterRequestHook)
Javelin
Plugin ID:
javelinJavelin’s AI security platform applies all guardrails configured in your application policy — trust and safety, prompt injection detection, language detection, and more — in a single call.Credentials: apiKey, application (application name for policy-specific guardrails), optionally domain.Functions:javelin.guardrails— applies all enabled guardrails in the named application policy (beforeRequestHook,afterRequestHook)
Sydelabs SydeGuard
Plugin ID:
sydelabsSydeGuard assesses the risk of inputs for prompt injections, toxicity, and evasion attempts, with configurable per-category thresholds.Credentials: apiKey — find it in the Sydelabs dashboard.Functions:sydelabs.sydeguard— risk assessment with configurable thresholds (beforeRequestHook,afterRequestHook)
Web-augmented transformers
Exa Online Search
Plugin ID:
exaEnhances prompts with real-time web search results from Exa’s search API before the request reaches the LLM. This is a transformer, not a guardrail — it modifies the request rather than blocking it.Credentials: apiKey — your Exa.ai API key.Functions:exa.online— injects web search context into prompts (beforeRequestHook)