Security skill for detecting prompt injection attacks - regex + LLM layers
- Python 100%
- New Layer 2: Local prompt-guard service (localhost:8080) - Uses protectai/deberta-v3-base-prompt-injection-v2 - ~50ms latency, ~99% accuracy, multilingual - CLI flags: --local, --skip-regex, --threshold - Environment: PROMPT_GUARD_URL, PROMPT_GUARD_THRESHOLD Architecture now: Regex (~80%) → Local DeBERTa (~99%) → LLM API (fallback) Tested with English and German injection attempts. |
||
|---|---|---|
| references | ||
| scripts | ||
| SKILL.md | ||