AI LLM App Security Scanner
PromptGuard Scan
Run prompt injection, jailbreak, leakage, and tool-misuse checks before an LLM app reaches production.
cvss: 8.2 gate: fail recommended_fix: - enforce server-side tool authorization - isolate retrieved content from developer policy - redact secrets before model context
Release gate
Make LLM security feel like unit tests.
Configure once, run on every pull request, and block releases when high-risk prompt injection or jailbreak findings return.
- Attack every changed prompt, RAG connector, model route, and tool policy.
- Fail builds on hidden instruction leaks, unsafe tool calls, PII exposure, or multi-turn bypasses.
- Keep a report trail for customer security reviews and internal governance.
name: LLM Security Scan
on: [pull_request]
jobs:
promptguard:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- uses: promptguard/scan-action@v1
with:
app: support-agent
fail-on: high
report: promptguard-report.pdf
Coverage
Security checks for the parts of LLM apps that actually break.
Prompt Injection Auto Scan
Direct instructions, malicious retrieved content, tool-output poisoning, and multi-turn policy erosion.
Jailbreak Library Detection
Continuously updated attack templates for roleplay, encoding, translation, coercion, and refusal bypass patterns.
Sensitive Data Leakage
Detect PII, API keys, internal instructions, hidden prompts, credentials, and unsafe log exposure.
CI/CD Integrations
GitHub Actions and GitLab CI checks with PR status, severity thresholds, and machine-readable output.
Risk Reports
CVSS-style scoring, evidence, reproduction steps, remediation guidance, and PDF-ready export.
Multi-Model Support
Run the same suite across OpenAI, Anthropic, Gemini, and self-hosted LLM endpoints.
Buyer confidence
Turn risky AI behavior into fixable engineering work.
Every finding includes exploit evidence, affected surface, severity, recommended control, and a retest path. Security teams get the proof they need. Engineers get the next patch to make.
Pricing
Team annual is selected by default.
Annual billing is 50% off. All plans use NOWPayments checkout and keep the product page open.
Dev
For solo builders validating one product before launch.
- Prompt injection scans
- Jailbreak template checks
- PII and key leak detection
- HTML risk report
- Email support
Team
For engineering teams shipping AI apps through pull requests.
- Everything in Dev
- GitHub Actions and GitLab CI gates
- Multi-turn bypass testing
- PDF export and CVSS scoring
- Shared workspaces and API access
Enterprise
For platform teams, private deployments, and audit-heavy AI systems.
- Everything in Team
- Private deployment path
- Custom test packs
- Compliance evidence exports
- Priority security review support
Security playbooks