Skip to main content

Calculated Metrics

Calculated metrics enable organizations to define and apply customized, weighted evaluation criteria across individuals, teams and projects. The only true limitations is what you can imagine based on your available data.

Deanna Smith
Updated by Deanna Smith

Objective and fully customizable intelligence

Calculated metrics enable organizations to define and apply customized, weighted evaluation criteria across individuals, teams and projects. The only true limitations is what you can imagine based on your available data.

By translating qualitative performance dimensions into structured, weighted scoring models, these metrics provide transparent, data-backed insights aligned with your priorities.

Combined with alerts, not only can you track performance, but you can also receive notifications and take immediate action on any significant changes.


Case Study: Developer Scoring

Measuring team and individual performance of developers has always been a complex task and is open not only to subjective input and criteria, but also gamification as it is difficult for leadership to focus simultaneously on all relevant areas and weigh them appropriately. This results not only in data that lacks objectivity, but also frequently ends with developers focusing their efforts in the one or two easily measurable areas often used for this purpose and consequently not doing as well in other areas.

Instead of relying on fragmented metrics or subjective review cycles, Developer Scoring allows leaders to operationalize what 'great performance' means in their organization and measure it consistently.

How it Works:

Step 1: Define Criteria Framework

Organizations configure weighted categories and sub-criteria aligned to their engineering philosophy and strategic goals.

Step 2: Aggregate Performance Data

Scoring integrates relevant delivery, quality, planning and collaboration data based on your defined criteria.

Step 3: Weighted Scoring Calculation

Each metric is normalized and calculated according to its assigned weight, producing a composite score.

Step 4: Insights andΒ  Reporting

Leaders can view performance at individual, team and organizational levels with full transparency into contributing factors.

Sample Scoring Model:

Category

Weight

Sample Sub-Criteria

Quality

30%

Failed build steps, reopened bugs, PR review cycles, time in status

Planning

20%

Planning accuracy, planning delivery

Velocity

30%

Issue cycle time, PR pickup time, Closed issue complexity

Teamwork

20%

Production issues solved, customer issues solved, PR reviews

Aggregated Developer Score
Planning & Teamwork Scores

Quality and Planning scores

How did we do?

Core Concepts & Terms

Contact