Metrics for Software: Choosing the Right KPIs for Your Stage
At Shiksha Infotech, we were a 12-person team and I tried to implement the same metrics framework I'd read about from companies 100x our size. It was absurd — we were tracking deployment frequency when we deployed once a month. At UshaOm with 27 engineers, a subset of those metrics started making sense. At Salesken with 40+ engineers, all four DORA metrics became essential. The right metrics depend entirely on where you are.
Not every company needs every metric. A 3-person startup at seed stage tracking 30 metrics wastes time better spent shipping. A 100-person company at Series B needs metrics a startup didn't.
Company stage, business model, and team size determine which metrics matter. This guide walks through how to choose the right metrics for where you are.
The Problem: Metric Paralysis
Engineering leaders often inherit "best practices" metrics without questioning whether they fit their stage. The result:
- Seed stage startups spending time on dashboards instead of shipping
- Series A companies lacking visibility because they skipped foundational metrics
- Mature companies tracking too many metrics, unclear which matter
- Teams gaming metrics instead of shipping value
This happens because nobody talks about metrics in stage-appropriate ways. Everyone quotes "elite teams deploy 10x per day" without mentioning that's only achievable with infrastructure 90% of startups don't have.
Part 1: Seed Stage—Focus on Ship Speed and Burn Efficiency
Constraints at seed stage:
- 2-10 engineers
- Limited runway (12-18 months)
- Customer feedback is everything
- Infrastructure is "whatever works"
- No established processes
Goal: Ship features fast. Validate product-market fit. Keep burn rate sustainable.
Metrics that matter:
1. Deployment Frequency (Daily or Weekly)
Why: At seed, shipping is your competitive advantage. You want to get features to customers and collect feedback.
Target: At least weekly. Daily is better, but not required.
How to measure: Count deployments in your Git history or deployment system.
Why not others: Deployment frequency is the only shipping speed metric that matters at seed. Don't worry about lead time or cycle time yet—your process probably isn't repeatable enough to measure meaningfully.
Real example: A Series A company admitted they were a "weekly deployment company" for years. At seed, they shipped twice weekly. They asked: "Why did shipping slow down?" Investigation revealed:
- Team grew from 2 to 6 engineers
- No code review process initially, added after first incident
- Manual testing replaced by tests in CI
- Staging environment required to match production (added safety)
None of these are bad—they're necessary maturity. But they require acknowledging that you can't compare seed-stage metrics to Series B metrics.
2. Feature Adoption (Weeks to First Use)
Why: It doesn't matter how fast you ship if customers don't use it.
Target: Features should see meaningful usage (<50% of users) within 1 week. If a feature launches and sits unused for a month, you were wrong about what customers wanted.
How to measure: Add analytics to each feature. Track unique users and frequency of use in the first 2 weeks.
Real example: A productivity SaaS built an "offline mode" feature. Shipped it, celebrated the launch. 6 weeks later: 2% adoption. They were shocked. Investigation revealed:
- Feature was buried in settings (didn't know it existed)
- Customers were already solving offline workflows with workarounds
- They built for perceived demand, not actual demand
They killed the feature, reallocated the engineer, shipped 3 other features that customers had explicitly requested. Lesson: at seed, ask customers what they want before building.
3. Burn Rate Efficiency
Why: You have 12-18 months of runway. You need to know if you'll run out of money before you reach product-market fit.
Formula: Monthly cash burn / Revenue (or take annual burn / projected annual revenue)
Target: Depends on your metrics, but typically:
- <6 months of runway remaining = too high burn
- 12-18 months of runway = comfortable
-
24 months of runway = hiring aggressively to grow faster
How to measure: Accounting team provides monthly cash burn. Divide by monthly recurring revenue (MRR) if you have revenue, or use runway remaining.
Why this matters: Burn rate efficiency answers the question: "Can we reach profitability or Series A before running out of money?" If not, you need to increase revenue or decrease burn.
Real example: A developer tools company burning $300K/month with $0 revenue had 18 months of runway. They asked: "Can we reach Series A before running out of money?" The math: if they raised in month 18, they'd have a 0-month buffer. They made hard choices:
- Reduced burn to $200K/month (company-wide hiring freeze)
- Focused on enterprise sales (higher revenue per customer)
- Cut features that didn't impact sales
They didn't hit profitability but reached $50K MRR in month 15, made a strong Series A case.
4. Engineering Velocity (Features Per Sprint)
Why: At seed, you're iterating fast. You need to know if you're speeding up, slowing down, or plateauing.
Formula: Features completed per sprint (or per month)
Target: Consistent week-to-week. Trend is more important than absolute number.
How to measure: Count completed features or story points per sprint. Use same units weekly.
**Why not: Velocity in a traditional sense is unreliable at seed (too much variation). Instead, track "features shipped per sprint" and look at trends, not numbers.
Real example: A team shipped 5, 6, 3, 8, 4, 7 features per month (no pattern). They felt unproductive. But they shipped 33 features in 6 months. A competitor shipped 45 features (better execution, more resources). But the seed-stage company won because their features were more aligned with customer needs. Moral: consistency matters more than speed. Trends matter more than single-month numbers.
Part 2: Series A—Scale Team and Process
Changes from seed:
- 10-30 engineers
- Clear product-market fit
- Revenue/customers validating the model
- Need to scale without breaking things
Goal: Maintain shipping speed while improving quality. Build foundation for scale.
New metrics to add:
1. Deployment Frequency (Maintain Daily+)
Why: Series A teams should match seed-stage shipping speed, then improve it. No regression allowed.
Target: Daily or more.
What often happens: Shipping slows as team grows (more code review, testing, staging). Actively prevent this.
Real example: A company shipped 10x per day at seed. By Series A (6 engineers → 15 engineers), they dropped to 3x per day. They asked: "Why?" Turns out:
- Code review time went from 1 hour to 6 hours (people were busier)
- More testing required for larger surface area
- Approval processes for infrastructure changes
They solved it by:
- Enforcing "code review within 2 hours" SLA
- Building test automation for the new surface area
- Creating a "deployment approval checklist" (fast, objective)
Back to 10x per day.
2. Cycle Time (< 1 Week for Most Work)
Why: As team grows, work can get stuck. Tracking cycle time reveals bottlenecks early.
Formula: Time from work starting to shipping.
Target: Median cycle time <1 week for typical work. Some work will be longer (architecture changes), but 80% of work should be <7 days.
How to measure: Use project management tool to track time in each status. Calculate weekly or bi-weekly.
Why matters: Long cycle time reveals blockers:
- Waiting on approval
- Waiting on other teams
- Unclear requirements
- Technical complexity
Real example: A team had 3-week median cycle time. Investigation showed:
- 2 weeks in "waiting for QA" (QA was a bottleneck)
- 1 week actual development
Solution: Automated testing + developers responsible for test coverage. QA shifted to exploratory testing (uncovering edge cases, not regression testing). Cycle time dropped to 5 days.
3. Quality: Change Failure Rate
Why: As you ship more, you need to maintain quality. Change failure rate reveals if you're sacrificing stability for speed.
Formula: (Deployments requiring rollback / Total deployments) × 100
Target: <30%. Elite is <15%.
How to measure: Tag failed deployments in incident system.
Why matters: High failure rate creates emergency work, burns engineers out, damages customer trust.
Real example: A company shipping daily had 35% failure rate (1 in 3 deployments needed rollback). They investigated. Most failures were:
- Database migrations that broke backward compatibility (fixed with blue-green migrations)
- Uncaught exceptions in edge cases (fixed with better error handling)
- Configuration issues (fixed with config validation before deploy)
After fixes: 12% failure rate. Now they ship confidently.
4. Technical Debt Ratio (10-20% of Capacity)
Why: As you scale, technical debt accumulates. You need a systematic way to pay it down.
Formula: Story points of tech debt / Total story points × 100
Target: 10-20% of sprint capacity dedicated to tech debt.
How to measure: Tag issues as "Technical Debt" in backlog. Sum their story points.
Why matters: Too little tech debt work = complexity explodes. Too much = you're not shipping features.
Real example: A company had 0% tech debt allocation. Their codebase became so fragile that every feature took 2x longer (spending time working around existing problems). They allocated 20% of capacity to tech debt (refactoring, testing, architecture). This felt like "lost productivity" initially. But 6 months later, feature velocity actually increased (less time fighting complexity).
Part 3: Series B+—Optimize for Scale and Outcomes
Changes from Series A:
- 30-100+ engineers
- Multiple teams (frontend, backend, devops, etc.)
- International customers with higher uptime requirements
- Hiring and onboarding complexity
Goal: Maintain shipping velocity, improve quality and reliability, connect engineering to business outcomes.
New metrics to add:
1. Engineering ROI
Why: Engineering budget at Series B is significant. Board/investors want to know if it's paying off.
Formula: Annual revenue attributed to engineering / Total annual engineering spend
Target: Varies by industry, but 10:1 (every dollar of engineering generates $10 of revenue) is healthy for SaaS.
How to measure:
- Track revenue of customers using major features
- Subtract customer acquisition cost, compare to feature development cost
- Include cost savings from infrastructure/platform work
Real example: A company spent $5M annually on engineering. They analyzed ROI:
- Core product features: $60M revenue, $3.5M cost = 17:1 ROI
- Integrations: $15M revenue, $1M cost = 15:1 ROI
- Infrastructure/reliability: Enables above, $500K cost
- Internal tools: $0 revenue, $200K cost (but saves $1M annually in dev time)
ROI calculation: ($60M + $15M + $1M savings) / $5M = 15.2:1
They used this to justify hiring and investment in high-ROI areas.
2. Customer Satisfaction (NPS, CSAT)
Why: As you scale, customer satisfaction can drop (more customers, more failure modes, more bugs). Track it.
Formula: Net Promoter Score = % Promoters (9-10 rating) - % Detractors (0-6)
Target: SaaS NPS: 30-50 is healthy | 50+ is excellent
How to measure: Send quarterly NPS surveys. Calculate promoters minus detractors.
Real example: An infrastructure company's NPS dropped from 45 to 32 over 2 quarters. Detractor feedback: "Deployment takes too long and they don't communicate status." Engineering priorities:
- Reduced deployment time from 2 hours to 30 minutes
- Added status page with real-time deployment info
- Automated incident communication
NPS recovered to 48 in next quarter.
3. Reliability/Uptime
Why: Enterprise customers pay for reliability. As you scale, uptime becomes a revenue lever.
Formula: (Total time available - Total downtime) / Total time available × 100
Target: 99%+ is standard. 99.9% is high-performance. 99.99% requires significant investment.
How to measure: Monitoring system (Datadog, New Relic, custom). Track incident duration and frequency.
Real example: A company targeting enterprise had 99.2% uptime (28 minutes of downtime per month). They started losing deals to competitors with 99.9% uptime. They invested in:
- Redundancy (multi-region deployment)
- Better monitoring (catch issues before customers notice)
- Automated failover
- Chaos engineering (find failure modes before customers hit them)
Within 6 months: 99.95% uptime. This became a sales differentiator.
4. Developer Experience Score
Why: As you scale, onboarding and developer experience matter. Happy developers are more productive and stay longer.
Formula: Composite of: build time, test speed, onboarding time, documentation quality, tool satisfaction
Target: Varies, but track trends. Improving year-over-year is good.
How to measure:
- Quarterly developer survey (1-10 scale on each dimension)
- Track objectively: build time (from CI logs), onboarding time (from HR data)
Real example: A company with high turnover (engineers leaving after 18 months) investigated. Common complaint: "Setup is painful, feedback loops are slow, documentation is terrible." They invested in:
- Improved onboarding (interactive setup, scripts, 1-week buddy)
- Build acceleration (cache, parallelization, reduce to 3 minutes)
- Documentation (living docs, video walkthroughs)
- Better local dev experience
Next cohort of hires reported much better experience. Retention improved.
The Metrics Selection Matrix
This matrix helps choose which metrics matter for your stage:
| Metric | Seed | Series A | Series B+ |
|---|---|---|---|
| Deployment Frequency | Essential | Essential | Essential |
| Lead Time | Optional | Important | Essential |
| Change Failure Rate | Monitor | Essential | Essential |
| MTTR | Monitor | Important | Essential |
| Cycle Time | Monitor | Important | Essential |
| Technical Debt | Monitor | Important | Essential |
| Feature Adoption | Essential | Important | Important |
| Burn Rate Efficiency | Essential | Monitor | N/A |
| Engineering ROI | N/A | Optional | Essential |
| NPS/Customer Satisfaction | Nice-to-have | Important | Essential |
| Uptime/Reliability | Monitor | Important | Essential |
| Developer Experience | Nice-to-have | Important | Important |
| Velocity | Important | Important | Optional* |
| WIP | Optional | Important | Important |
*At Series B+, absolute velocity matters less than predictability.
The Golden Rule: Start with 3-5 Metrics
Don't try to track everything. Pick 3-5 metrics for your stage:
Seed: deployment frequency, feature adoption, burn rate efficiency Series A: deployment frequency, cycle time, change failure rate Series B+: deployment frequency, engineering ROI, uptime, NPS
Once you've improved these, add more. Once you've solved one problem, shift focus to the next.
How AI Agents Change Metrics by Stage
Manually tracking metrics at different stages is tedious:
- Seed stage: "Did we ship something this week?"
- Series A: "Where are cycle time blockers?"
- Series B+: "Which teams have the highest engineering ROI?"
Glue's AI agents automatically:
- Seed: Monitor deployments, alert if shipping slows
- Series A: Flag cycle time increases, identify bottlenecks (review, testing, approval)
- Series B+: Correlate team metrics with business outcomes, forecast capacity planning
Rather than manually running queries, your team gets proactive insights tailored to your stage.
Final Note: Metrics Evolve with Your Company
The metrics you track should evolve as your company does. Seed metrics are irrelevant at Series B. Series A metrics can become noise at Series B+ if you stop revisiting.
Every 6 months, ask:
- Are we still focused on the right problem?
- Do our metrics reflect that?
- Which metrics do we track but not act on?
- Which decisions are we making blind?
Metrics are guides, not scripture. Use them to see reality clearly. When reality changes, your metrics should too.
Track what matters for your stage. Glue's AI agents monitor the right metrics for where you are, surface insights automatically, and help your team focus on unblocking work instead of running dashboards. See how other engineering leaders at your stage are using AI agents to scale without sacrificing speed or quality.
Related Reading
- Coding Metrics That Actually Matter
- Engineering Metrics Examples: 20+ Key Metrics Your Team Should Track
- Engineering Team Metrics: The Complete Framework
- DORA Metrics: The Complete Guide for Engineering Leaders
- Sprint Velocity: The Misunderstood Metric
- Software Productivity: What It Really Means and How to Measure It