Attack Catalog
16 attack categories aligned to OWASP LLM Top 10 and MITRE ATLAS
Tachyonic ships 186 attack patterns across 16 categories — 144 from the open attack taxonomy plus 42 built-in offensive modules. Each attack is mapped to the OWASP LLM Top 10 2025 and MITRE ATLAS.
Categories
Prompt Injection (LLM01)
Direct prompt injection attacks that attempt to override system instructions, extract hidden context, or alter model behavior through crafted user input.
- 20 attack definitions (PI-001 to PI-020)
- Tests: instruction override, role hijacking, delimiter attacks, encoding bypass
System Prompt Extraction (LLM01)
Attempts to leak the system prompt or hidden instructions configured by the developer.
- 12 attack definitions (SPL-001 to SPL-012)
- Tests: direct extraction, indirect leakage, format manipulation
Jailbreak (LLM01)
Guardrail bypass techniques that attempt to make the model ignore safety constraints.
- 22 attack definitions (JB-001 to JB-022)
- Tests: DAN prompts, character roleplay, encoding tricks, multi-language bypass, hypothetical framing
Indirect Injection (LLM01)
Attacks that inject malicious instructions through external data sources (RAG documents, tool responses, web content) rather than direct user input.
- Requires RAG-enabled targets
Tool Abuse (LLM06)
Manipulation of tool/function calling to achieve unintended actions.
- 12 attack definitions (EA-001 to EA-012)
- Tests: parameter manipulation, tool chaining, path traversal via tools, SQL injection via tools, SSRF via API tools
Multi-Turn Manipulation (LLM01)
Attacks that build context across multiple conversation turns to gradually shift model behavior.
- 8 attack definitions (MT-001 to MT-008)
- Tests: context poisoning, incremental jailbreak, trust building
Vision Injection (LLM01)
Attacks targeting vision-enabled models through crafted images.
- 12 attack definitions (VI-001 to VI-012)
- Tests: text-in-image injection, steganographic payloads, adversarial patches
- Requires vision-capable target
Sensitive Disclosure (LLM02)
Attempts to extract sensitive information, PII, credentials, or internal data from model responses.
- 10 attack definitions (SID-001 to SID-010)
- Tests: social engineering, data exfiltration, credential harvesting
Supply Chain (LLM05)
Attacks targeting the model's dependency chain, including tool schemas, plugins, and external integrations.
- 8 attack definitions (SC-001 to SC-008)
- Tests: tool schema poisoning, dependency confusion, plugin backdoor
Improper Output (LLM09)
Testing for unsafe output that could be used for downstream attacks (XSS, code injection, etc.).
- 8 attack definitions (IOH-001 to IOH-008)
- Tests: code generation safety, HTML/JS injection, command injection in generated code
Unbounded Consumption (LLM10)
Resource exhaustion attacks that attempt to consume excessive compute, tokens, or time.
- 2 attack definitions (UC-001 to UC-002)
- Tests: recursive expansion, infinite loop prompts
Permission Escalation (LLM06)
Attacks that attempt to escalate privileges within agent systems, bypassing authorization controls.
- Tests: admin impersonation, token privilege escalation, cross-tenant data access, role confusion
Multi-Agent Injection (LLM06)
Attacks targeting multi-agent architectures where multiple AI agents collaborate.
- Tests: orchestrator override, agent impersonation, inter-agent prompt injection
Misinformation (LLM09)
Attempts to generate convincing false information or manipulate factual outputs.
Vector Embedding (LLM08)
Manipulation of vector embeddings and semantic search systems.
- 8 attack definitions (VE-001 to VE-008)
Video Injection (LLM01)
Attacks targeting video-processing models.
Severity Levels
| Level | Weight | Description |
|---|---|---|
| Critical | 5.0 | Remote code execution, credential leakage, full system compromise |
| High | 3.0 | Data exfiltration, privilege escalation, safety bypass |
| Medium | 2.0 | Information disclosure, partial guardrail bypass |
| Low | 1.0 | Minor information leakage, low-impact behavior change |
| Info | 0.5 | Informational findings, no direct security impact |
Resistance Score
After a scan, Tachyonic calculates a resistance score from 0-100:
| Score | Rating |
|---|---|
| 90-100 | Excellent |
| 75-89 | Good |
| 50-74 | Fair |
| 25-49 | Poor |
| 0-24 | Critical |
The score is severity-weighted — a single critical finding has more impact than multiple low findings.
Custom Heuristics
Attack definitions are YAML files. Set HEURISTICS_PATH to load a custom attack library:
export HEURISTICS_PATH=/path/to/your/attacks
tachyonic scan --target ... --include-builtinSee the attack schema for the YAML format.