🧯

Incident Postmortem Report

Capture incident facts quickly, then turn them into clear root cause analysis and follow-up actions.

Incident Postmortem Report
You’ll be asked to sign in first.
softwaresaasincident-reviewpostmortemreliability
Use this after service incidents when the team needs a clean record of what happened, impact, and what changes next. Voice-first prompt suggestion: "Create a postmortem for incident INC-2026-0215. Checkout latency spiked after deploy. Severity SEV-2. Incident ran from 18:04 to 19:12 UTC. Customer impact: 18% delayed checkouts and 4% failures. Timeline: alert, incident start, rollback, cache tuning, recovery. Root cause: retry policy increased downstream load. Improvements: better pre-release load testing and tighter alerts. Action items with owners and due dates included."

Form preview

Overview
Incident Title*
Text
[not set]
Incident ID*
Text
[not set]
Severity*
Text
SEV-1SEV-2SEV-3SEV-4
SEV-1
SEV-2
SEV-3
SEV-4
Incident Window*
Text
Start and end time
[not set]
Summary*
Text Area
Customer Impact*
Text Area
Analysis
Timeline
List
TimeEventOwner
+ Add+ Add+ Add
Root Cause*
Text Area
Contributing Factors
Text Area
Follow-up
What Went Well
Text Area
What To Improve
Text Area
Action Items
List
ActionOwnerPriorityDue DateStatus
+ Add+ Add+ Add+ Add+ Add
EXAMPLE REPORTSample data — shows what you’ll get after recording

Example filled report

Overview
Incident Title
Checkout API latency spike during evening traffic
Incident ID
INC-2026-0215
Severity
SEV-2
Incident Window
2026-02-15 18:04 UTC to 2026-02-15 19:12 UTC
Summary
Latency increased sharply after a deployment, causing retries and intermittent checkout failures until rollback and cache tuning stabilized response time.
Customer Impact
Around 18% of checkout attempts were delayed and about 4% failed during the peak hour.
Analysis
Timeline
TimeEventOwner
18:04Latency alert triggeredOn-call SRE
18:11Incident declared and war room startedIncident commander
18:29Rollback to previous releaseBackend engineer
18:47Cache configuration tunedPlatform engineer
19:12Metrics normalized and incident resolvedOn-call SRE
Root Cause
A new retry policy increased downstream load under peak traffic, which pushed cache misses and elevated response times.
Contributing Factors
Alert threshold was too loose for evening load and dashboard lacked queue depth visibility.
Follow-up
What Went Well
Fast incident declaration and clear ownership reduced time to stabilization.
What To Improve
Run load tests for retry policy changes and tighten latency alert thresholds.
Action Items
ActionOwnerPriorityDue DateStatus
Add queue depth panel to checkout dashboardPlatform teamHigh2026-02-20Open
Create pre-release load test gate for retry changesBackend teamHigh2026-02-22Open
Incident Postmortem Report | Voiz Report Templates