Incident Response for AI Systems

AI-Specific IR Considerations

Traditional incident response frameworks (NIST SP 800-61, SANS) apply, but AI incidents have unique characteristics:

  • Attribution is harder. A prompt injection attack looks like a normal user query.
  • Blast radius is unclear. If a model is compromised via poisoning, every output since the last known-good checkpoint is suspect.
  • Evidence is ephemeral. Conversation logs may not capture the full context. Model state isn't easily snapshot-able.
  • Remediation is slow. You can't patch a model the way you patch software. Retraining takes weeks and costs millions.

AI Incident Categories

CategoryExampleSeverity
Data leakage via AIModel outputs PII, credentials, or proprietary dataCritical
Prompt injection in productionAttacker hijacks AI assistant behaviorHigh
Model compromisePoisoned model deployed, backdoor activatedCritical
Shadow AI data exposureEmployee uploads sensitive data to unauthorized AI toolHigh
Hallucination with impactAI provides false information leading to business decisionMedium-High
AI-powered social engineeringDeepfake or AI-generated phishing targeting employeesHigh
API abuse / extractionAnomalous query patterns indicating model theftMedium

Response Playbook

Immediate (0-4 hours)

  1. Confirm the incident — is this a real AI-specific issue or a traditional security incident?
  2. Contain — disable the affected AI endpoint, revoke API keys, block the source
  3. Preserve evidence — export conversation logs, model version, system prompt, RAG state
  4. Notify stakeholders — CISO, legal, privacy team, affected business owners

Short-term (4-48 hours)

  1. Determine scope — how many users affected? What data exposed?
  2. Root cause analysis — was it injection, poisoning, misconfiguration, or insider?
  3. Remediate — patch system prompt, update filters, rollback model if needed
  4. Communicate — internal notification, customer notification if data exposed

Long-term (1-4 weeks)

  1. Post-incident review — what failed and why?
  2. Update controls — new filters, monitoring rules, access restrictions
  3. Red team validation — test that the fix actually works
  4. Policy updates — revise AI governance based on lessons learned
  5. Regulatory reporting — if required (GDPR breach notification, etc.)

Tabletop Exercise Scenarios

Run these quarterly with your IR team:

  1. Scenario: Customer reports the chatbot revealed another customer's account details
  2. Scenario: Security researcher publishes a blog post with your extracted system prompt and API keys
  3. Scenario: Internal monitoring detects a fine-tuned model was deployed with a backdoor
  4. Scenario: An employee's AI-generated phishing email compromises a VIP target
  5. Scenario: Your AI vendor (OpenAI/Anthropic) reports a data breach affecting your API usage