Elite vs. Low DORA Performers: What Separates Them in 2026
For six years, the DORA report divided the world into four performance tiers: elite, high, medium, and low. The 2025 report replaced that with seven archetypes — a recognition that the original buckets were masking important nuance. The top tier still exists, and the gap between it and the bottom tier is wider than ever. Here is what the data shows about what actually separates them.
What this article covers
The seven 2025 DORA archetypes, what elite performers do differently in practice, AI adoption patterns across performance tiers, deployment frequency benchmarks by company size, and the organizational practices that correlate most strongly with elite performance.
From Four Tiers to Seven Archetypes
The original four-tier model — elite, high, medium, low — was designed to communicate a simple idea to a broad audience: there is a performance distribution, and you can roughly locate your organization within it. It did that job well for years.
Its limitation became apparent as the research accumulated more data. Teams at the same "medium" level had wildly different profiles: one team might have excellent deployment frequency and terrible MTTR; another might have the reverse. Lumping them into the same bucket obscured what actually needed to change.
The 2025 DORA archetypes represent clusters identified through latent class analysis — statistical groupings of real teams based on their actual metric combinations. The seven archetypes are:
- Accelerators — high throughput, high stability. The classic elite performers.
- Stable Shippers — moderate frequency, very low CFR. Risk-averse teams with strong quality gates.
- Fast Movers — high frequency, elevated CFR. Moving fast and occasionally breaking things.
- Reliability Focused — low frequency, elite MTTR. SRE-driven organizations optimizing for incident response.
- Batch Releasers — infrequent deploys, moderate stability. Sprint-based release cycles with manual QA gates.
- Struggling Shippers — low frequency, high CFR, long MTTR. Classic low performers.
- Reactive Teams — moderate frequency but extremely high rework rate. Deployment pipeline exists, quality gates do not.
The Accelerator Archetype: What Elite Actually Looks Like
Accelerators — the 2025 equivalent of the elite tier — represent approximately 18% of teams surveyed in the 2025 report, down from 26% in 2024. The decrease is partly explained by the new rework rate metric raising the bar for elite classification.
| Metric | Accelerators (Elite) | Struggling Shippers |
|---|---|---|
| Deployment Frequency | Multiple times/day | 1–3× per month |
| Lead Time | < 1 hour | 1–4 weeks |
| Change Failure Rate | < 5% | > 20% |
| MTTR | < 1 hour | > 24 hours |
| Rework Rate | < 5% | > 25% |
The gap between Accelerators and Struggling Shippers on deployment frequency is roughly 60×. This is not a gap that can be closed by hiring better engineers or investing in better tooling alone. It reflects a fundamental difference in how the delivery pipeline is architected and how risk is managed.
What Accelerators Do Differently: The Five Practices
1. Trunk-based development with feature flags
Accelerators almost universally practice trunk-based development — all engineers commit directly to main (or a very short-lived branch with a maximum lifespan of one day). This eliminates the merge conflicts and integration overhead that slow batch-release teams down. Feature flags allow code to be deployed to production without being activated — decoupling the deployment decision from the release decision.
The 2025 DORA data found trunk-based development in 89% of Accelerator teams versus 23% of Struggling Shippers.
2. Automated deployment pipelines with quality gates
Every commit to main triggers an automated pipeline that runs tests, builds the artifact, and deploys to production if all gates pass. The key word is "gates" — not just tests, but coverage thresholds, lint checks, security scans, and increasingly, deploy risk score thresholds. Accelerators do not manually approve deployments for routine changes; the pipeline is the gatekeeper.
3. Comprehensive observability before deployment
Accelerators invest heavily in pre-deploy validation: synthetic tests, canary deploys that route 1–5% of traffic to the new version before full rollout, and automated rollback triggers based on error rate or latency SLO breach. They can deploy frequently because they can recover quickly — and recovery is automated, not manual.
4. Small, focused changes
The median PR size at Accelerator teams is 84 lines changed, compared to 340 lines for Struggling Shippers. Smaller changes are easier to review, easier to test, and dramatically easier to roll back when something goes wrong. This practice is cultural as much as technical — it requires engineering managers and tech leads to actively discourage large PRs during planning, not just review.
5. Pre-deploy risk scoring
This is the newest addition to the Accelerator playbook. The 2025 report found that 71% of Accelerator teams use some form of automated pre-deploy risk assessment — a score or classification that evaluates the risk of a pending change before it merges. This signal is used to route high-risk changes to mandatory review queues and to block deployments of changes above a risk threshold during high-traffic periods.
AI Adoption Patterns Across the Archetypes
AI coding assistant adoption in 2025 was near-universal — over 90% of teams surveyed reported at least some AI-assisted development. But adoption patterns differed significantly by archetype in ways that are illuminating.
| Archetype | AI Adoption | AI Review Process |
|---|---|---|
| Accelerators | High (42% AI-assisted) | Differentiated review — AI PRs get extra scrutiny |
| Stable Shippers | Low–Medium (18%) | Conservative adoption, strict code review |
| Fast Movers | Very High (61%) | Minimal differentiation — driving up CFR |
| Reactive Teams | High (49%) | No differentiation — highest rework rate |
| Struggling Shippers | Medium (31%) | Inconsistent — not the primary bottleneck |
The pattern is clear: Accelerators adopt AI tools at high rates but differentiate their review process for AI-assisted PRs. Fast Movers adopt at even higher rates but treat all PRs the same — and pay for it in CFR and rework rate. The tool is not the variable. The review process is.
Deployment Frequency Benchmarks by Company Size
One of the most important adjustments to the DORA benchmark conversation in 2025 is the explicit acknowledgment that deployment frequency targets should be calibrated to company size and stage. The raw numbers:
| Company Size | Accelerator Target | Median for Size |
|---|---|---|
| Seed / Early (<20 eng) | 5–15× per day | 2–4× per day |
| Series A/B (20–100 eng) | 10–30× per day | 1× per day |
| Growth (100–500 eng) | 30–100× per day | 2–5× per day |
| Enterprise (>500 eng) | 100+ per day (across services) | Weekly–monthly |
The enterprise numbers look extreme until you account for the fact that large organizations typically operate dozens of microservices, each with its own deployment pipeline. The deployment frequency benchmark applies per service, not per organization. A 500-engineer company with 50 services deploying twice per week per service is at 100 deployments per week — squarely in the medium tier. An Accelerator at that scale would have most services deploying daily or multiple times daily.
What Moving from Low to High Performance Actually Requires
The DORA research has consistently found that technical practices alone are not enough to produce elite performance. The organizations that move from low to high archetype over 12–18 months typically make three non-technical changes alongside the technical ones:
Leadership that treats engineering metrics as organizational health data, not scorecards. Teams that are measured on DORA metrics as KPIs to optimize tend to game them. Teams where leadership uses DORA data to identify systemic blockers and invest in removing them tend to improve.
Psychological safety to report incidents honestly. CFR and MTTR data is only accurate if engineers report incidents completely. In low-safety cultures, incidents get triaged as "not a deployment issue" to protect individuals or teams. Accurate DORA data requires an environment where the honest number is the safe number.
Platform investment that makes the right thing easy. Elite teams make small, frequent, reviewed changes not because they have stronger discipline but because their platform makes it easy. The investment required is in developer experience: faster CI, one-click rollback, automated risk scoring, and deployment environments that mirror production.
See which archetype your team matches
Koalr automatically calculates all five DORA metrics and shows your trend over time. Connect GitHub and your incident tool in under 5 minutes.