Where it fits
- A ChatGPT-powered support bot can access account details or draft customer actions.
- A custom GPT-like app uses hidden instructions, files, or tool calls.
- A team is switching model versions and needs to retest prior attack cases.
Operational steps
- Create a test profile for the ChatGPT app surface and allowed behavior.
- Run attack cases for instruction override, hidden prompt extraction, tool coercion, and data leakage.
- Compare results across model versions or providers when routing changes.
- Use the report to adjust prompts, guardrails, authorization checks, and logging policy.
Common risks
- The model refuses obvious attacks but leaks instructions through translation or summarization prompts.
- Tool calls trust model output without server-side authorization.
- A file or web page supplies malicious instructions that outrank the intended policy.
How PromptGuard Scan fits the workflow
PromptGuard Scan supports ChatGPT security testing as part of a broader LLM security workflow, with model-agnostic scan packs and CI-friendly evidence.