A retrospective board that stays grounded in Jira visibility
Retrospective comparison, not surveillance
Project page / SeeCodes
Team Competition
Visible contributors
3
Jira visibility enforced
Visible solved tasks
11
Done tasks in scope
Relative effort produced
764
Composite effort units in scope
Monthly baseline
100
Typical visible completed task
AI-generated retrospective board
“Effort” is a productivity-oriented composite of active minutes, changed LOC, files changed, and architecture / logic / UI-specification signals.
Raw effort = active minutes × 1.8 + LOC changed × 0.18 + files changed × 3, plus bonuses for architecture (+24), logic (+14), and UI / specification (+10). A relative score of 100 means roughly “typical for this month.”
Competition Board
Sort by relative effort, solved tasks, logic, architecture, UI/spec, or contributor name.
Avery Chen
5 solved • 7 contributed tasks
Relative effort 312 • Active minutes 184 • LOC changed 1210 • Files changed 47 • Avg solved-task effort 128
Priya Shah
4 solved • 6 contributed tasks
Relative effort 254 • Active minutes 162 • LOC changed 990 • Files changed 39 • Avg solved-task effort 117
Jordan Miles
3 solved • 5 contributed tasks
Relative effort 198 • Active minutes 141 • LOC changed 740 • Files changed 31 • Avg solved-task effort 109
Solved Task Board
Top solved tasks in scope with relative effort and transparent raw-effort drivers.
PLAT-221
Rework auth token rotation
Assignee Avery Chen • 23 active min • 78 LOC changed • 3 files changed
PAY-84
Stabilize billing guardrails
Assignee Priya Shah • 22 active min • 78 LOC changed • 4 files changed
WEB-97
Refine logout and error messaging
Assignee Jordan Miles • 19 active min • 61 LOC changed • 3 files changed
How the page reads effectiveness from effort
Important nuance: the formula is an effort proxy
Raw effort = active minutes × 1.8
+ LOC changed × 0.18
+ files changed × 3
+ architecture signal ? 24 : 0
+ logic signal ? 14 : 0
+ UI / spec signal ? 10 : 0
Relative score ≈ raw effort / monthly typical raw effort × 100
100 ≈ "typical for this month"The strongest scientific case is for the choice of inputs and for the monthly normalization step. The exact coefficients are still product heuristics chosen so that time, churn, diffusion, and semantically heavier work all stay visible in one interpretable index.
How to read a score of 100
- 100 ≈ a typical visible completed task in the current month.
- 120 means roughly 20% above that month’s visible baseline, not 20% 'better' than another team in another repository.
- 60 means lighter than that month’s typical visible completed task, not low value or weak performance.
Where “effectiveness” actually shows up
- Solved count shows whether effort is turning into finished work.
- Average solved-task effort shows whether someone is closing lighter or heavier completed tasks.
- Dominant-area labels show whether contribution skewed toward architecture, logic, or UI/spec work.
Active minutes × 1.8
Minutes with detected work activity give the score a focused-work component, but the page explicitly frames them as a proxy for retrospective context rather than payroll time.
Changed LOC × 0.18
Code churn captures implementation magnitude. The lower coefficient prevents raw line volume from overpowering every other signal on the board.
Files changed × 3
Cross-file changes usually imply broader reasoning, more coordination, and more places where side effects can appear, so the model gives diffusion visible weight.
Semantic bonuses
Architecture (+24), logic (+14), and UI / specification (+10) bonuses stop the model from pretending that every edit has the same blast radius or product meaning.
Why this formula shape is scientifically defensible
Research-informed structure; heuristic coefficients
Activity belongs, but never alone
Developer-productivity research consistently treats activity and flow as useful dimensions of work, while warning that activity-only measurement is easy to misuse.
LOC changed is a defensible size signal
Empirical software-engineering research shows that relative code churn is informative for change magnitude and defect risk, which makes changed LOC a reasonable ingredient in a composite score.
Files changed captures diffusion
Just-in-time defect-prediction studies repeatedly find that distributed changes across more files or subsystems are harder to reason about and more likely to be risky.
Architecture deserves extra credit
Architecture and modifiability research shows that semantic transformations and change propagation affect cost beyond raw structural counts, which supports a larger architecture bonus than a UI / spec bonus.
Research anchors behind the rationale
- SPACE framework: productivity is multidimensional, and activity metrics should not be used in isolation.
- Relative code churn research: changed LOC becomes useful when interpreted as contextual change magnitude rather than as a vanity metric.
- Just-in-time quality assurance research: files touched and change diffusion are among the most useful signals for risky software changes.
- Architecture modifiability research: semantic architectural transformations affect change cost beyond raw structural metrics alone.
- Computer-activity studies: active minutes can correlate with perceived productivity, but extremes and context still matter.
How to use the score safely
- Treat the score as a retrospective index, not as a universal truth about individual productivity.
- Compare contributors only inside similar scope, visibility, and time windows rather than across unrelated months or teams.
- Read high effort together with solved tasks, risk, and review quality; effort alone is not effectiveness.
- Never turn the score into payroll, surveillance, or a single-number performance system.
Where teams use it
- Run retrospectives with a shared view of solved work, contributor mix, and relative task effort.
- Celebrate specialists and generalists without forcing the conversation into raw ticket counts.
- Spot work that needed multiple contributors or carried unusually high relative effort.
- Find follow-up topics for pairing, knowledge sharing, or planning quality in the next sprint.
Best used for retrospectives and coaching