training · general

Why tracking the wrong metrics is quietly killing your long-term progress

Bradley Hunt ·
process metrics training log consistency habit formation behavioral consistency identity-based habits performance tracking long-term adherence training methodology goal setting intrinsic motivation adaptive training deliberate practice treadmill training outcome metrics

Your finish time is a photograph. Your training log is the film.

The photograph tells you where you ended up. It says nothing about how you got there, whether the process was sustainable, or whether you can repeat it. Most athletes spend the majority of their mental energy optimising for the photograph.

That is not a training strategy. It is a hope strategy.

The athletes who compound over years, the ones who keep showing up in their forties and fifties with genuine fitness, not nostalgia fitness, are not necessarily more talented or more motivated. They have built tracking systems that reward the work itself rather than just the outcome of the work. That distinction sounds small. The performance gap it creates over a five-year horizon is not.

Why outcome metrics fail you between races

A finish time, a power output, a race placing. These are outcome metrics. They are real, they matter, and they should be part of any serious training log. But they have a structural problem: they only update at the end of something.

If your next race is 16 weeks away, your outcome metrics are frozen for 16 weeks. You are training every day against a feedback system that cannot tell you whether what you are doing is working until you cross a finish line four months from now.

This creates a specific kind of motivational vacuum. Research in sports psychology has documented it consistently. Weinberg and Gould’s work on goal-setting and motivation shows that athletes who train exclusively toward a single outcome event experience measurable motivation decline after that event, and often during the long training blocks preceding it. The goal is too far away to provide daily feedback. The training feels disconnected from the result.

There is a second failure mode. Outcome metrics reward results that are partially outside your control. Weather, course conditions, a bad night of sleep before race day. You can train perfectly for 16 weeks and produce a slower time than your fitness deserves. If your entire feedback system is built on that number, a race-day variable just invalidated months of genuine progress.

Process metrics solve both problems.

What process metrics actually measure

A process metric measures an input, not an output. It asks: did you do the thing, and did you do it at the quality level you intended?

In business process management literature, process metrics are defined as quantifiable indicators of the efficiency, effectiveness, and consistency of a process, independent of the downstream outcome. The same logic applies directly to training. Your aerobic base is a process. Your strength development is a process. Your recovery management is a process. None of these produce a visible outcome every week, but all of them determine what outcome is possible when race day arrives.

The practical difference looks like this:

  • Outcome metric: “I ran a 42-minute 10km last month.”
  • Process metric: “I completed 94% of planned sessions in the last eight weeks, with 87% of aerobic sessions in the correct intensity zone.”

The first tells you where you were. The second tells you what your next result is likely to be.

BJ Fogg’s behavior model, developed through his research at Stanford’s Behavior Design Lab, makes a related point about motivation. Motivation is inherently variable. It spikes before a goal and crashes after it. Systems anchored to process are structurally more stable than systems anchored to outcomes because they generate feedback at the frequency of the behavior, not at the frequency of the result. Every completed session is a data point. Every week of consistent training is a signal the system can read.

Deliberate practice: the missing layer most training logs ignore

Anders Ericsson’s research on deliberate practice is one of the most cited and least applied findings in performance science. The core finding: expert performance is not primarily a function of volume. It is a function of the quality of attention brought to each repetition.

Deliberate practice has a specific structure. You set a precise execution target before the session. You perform the session with focused attention on that target. You review whether you hit it. Then you adjust.

Most training logs record volume and outcome. Almost none record execution intent. That gap is where the deliberate practice framework gets lost.

Applied to endurance training, deliberate practice looks like this:

  • Before a zone 2 run: “My target is to hold 140-148bpm for 60 minutes without drifting above 150.”
  • After the session: “I held zone for 52 of 60 minutes. Drifted in the last 8 minutes on the climb. Adjust pacing on that section next week.”
  • Before a tempo block: “Target is 4:45/km for three intervals. Focus is on even splits, not a fast first interval.”
  • After: “Splits were 4:43, 4:47, 4:52. Went out slightly fast. Adjust.”

The session becomes a skill development exercise, not just a fitness accumulation event. The process metric that captures this is execution quality: not how hard the session was, but how well you executed the intention behind it.

This is also why the process-first framework connects to long-term development rather than just short-term consistency. Volume compounds. Deliberate practice compounds faster, because each session is also improving the athlete’s ability to execute subsequent sessions well.

Why treadmill running is a process metric problem

Treadmill running is the clearest example of outcome-focused training in practice, and it is worth examining directly because it explains why so many athletes find it unrewarding.

On a treadmill, the belt controls your pace. The environment does not change. There is no wind, no gradient variation, no terrain to read. Your attention, with nothing external to engage it, defaults to the one variable that is visible: the clock. How much time is left. How far you have gone. When it will be over.

You are, structurally, focused entirely on the outcome. The process has been removed.

This is not a motivation problem. It is a design problem. The treadmill strips out almost every cue that makes outdoor running a skill-development activity. Pacing judgment, terrain adaptation, cadence adjustment on gradient, reading your own effort against a changing environment. These are the process elements that make running interesting and developmentally rich. The treadmill replaces them with a number counting down.

The fix is not to avoid treadmills entirely. They have legitimate uses: controlled interval work, heat acclimatisation, running when conditions make outdoor training unsafe. But using a treadmill well requires deliberately reintroducing process focus. Set a cadence target and track it. Work on foot strike. Run a structured interval session where the execution quality of each rep is the metric, not the total time on the belt. Give yourself something to execute, not just something to endure.

If your treadmill sessions feel like something to survive rather than something to do well, that is the process metric problem made physical.

The Sisyphus reframe

Ross Edgley’s framing of the Sisyphus principle, developed in The Art of Resilience, is useful here. The athlete who finds genuine meaning in the push, not just the summit, is structurally more durable than the one chasing a finish time. This is not motivational poster language. It is a description of how long-term athletic capacity actually compounds.

The push is where adaptation happens. The summit is where adaptation gets measured. If you only care about the summit, you will optimise for the measurement event rather than for the adaptation process. These are not the same thing, and the difference becomes visible over years.

A process-first tracking system operationalises this. It makes the push legible, not just the summit. When your log shows 47 consecutive weeks of training adherence, that number carries meaning independent of any single race result. It is a record of the adaptation that has accumulated. It is also a record of identity: this is the kind of athlete you are.

James Clear’s work on identity-based habits, drawing on Fogg and behavioral science more broadly, makes the mechanism explicit. Athletes who define themselves by the practice rather than the result show higher long-term adherence. Each process metric entry is a small vote for the identity of someone who does the work. The finish time is the consequence of enough of those votes accumulating.

Five process metrics worth tracking across any sport

These apply regardless of sport. Adjust the specifics to your context, but the categories hold.

Session completion rate. The percentage of planned sessions you actually complete over a rolling four-week window. Not whether they were perfect, just whether they happened. Based on my reading of adherence research, a completion rate above 85% over 12 weeks is a stronger predictor of long-term progress than any single session quality metric. It tells you your planning is realistic and your commitment is consistent.

Intensity zone accuracy. For endurance athletes especially, the gap between intended intensity and actual intensity is where most training errors live. If your plan calls for a zone 2 aerobic session and you drift into zone 3, that session is not what you think it is. Tracking the percentage of sessions executed at the correct intensity reveals whether your training is actually structured or just effortful. The two are not the same (the interference effect between strength and endurance work is one downstream consequence of mismanaged intensity, covered in more detail in the interference effect article).

Sleep consistency. Not just total hours, but variance. An athlete averaging seven hours with a standard deviation of 90 minutes is in a different recovery state than one averaging seven hours within a 20-minute window. That is my synthesis of the sleep science rather than a single study finding, but the practical implication is consistent across the literature: sleep consistency predicts next-day training quality more reliably than sleep duration alone. It is a process metric because it is something you actively manage, not something that happens to you.

Perceived effort quality. A simple one-to-five rating after each session: not how hard it was, but how well you executed what you intended. A session rated five for effort but two for quality is diagnostic. It tells you something about fatigue, readiness, or planning that a power file alone will not surface. Over time, the distribution of quality ratings tells you whether your training load is sustainable.

Recovery indicator trend. HRV, resting heart rate, or a subjective readiness score, tracked as a seven-day rolling average rather than a daily point. The trend matters more than any single reading. A downward trend over two weeks is a signal that load is outpacing recovery, regardless of what your finish times are doing.

Auditing your current system

Most training logs are outcome-heavy by default. GPS watches record distance and pace. Training platforms surface power and speed. Strava gives you a leaderboard. None of this is wrong, but it creates a tracking environment that defaults to outcome measurement.

A useful audit takes about 20 minutes. Pull your last eight weeks of training entries. For each session, ask: does this entry tell me what I did, or does it tell me how well I executed the intention behind what I did?

If the answer is consistently the former, your log is a record of outcomes. It is not a learning system.

The fix is not to add more data. It is to add the right data. Three fields added to every session entry will shift the balance:

  • Did I complete this session as planned? (Yes / Modified / Missed)
  • What was my execution quality? (1-5)
  • What was my recovery status going in? (Fresh / Normal / Tired / Depleted)

These three fields cost 30 seconds per session and generate the diagnostic data that outcome metrics cannot. After eight weeks, you have a dataset that can tell you whether your missed sessions cluster around specific training blocks, whether your quality ratings correlate with your recovery status, and whether your planned load matches your actual capacity.

That is the information you need to improve the process, not just measure the result.

Combining process and outcome data without letting outcomes dominate

The goal is not to stop tracking outcomes. It is to stop letting outcomes dominate the feedback system.

A practical ratio: three process metrics for every one outcome metric in your regular review. Weekly, you review session completion, intensity accuracy, and recovery trend. Monthly, you review a fitness indicator, a power test, a time trial, something that measures where your adaptation has landed. Quarterly, you race or test against an external benchmark.

This structure keeps outcome data in its correct role: periodic calibration of whether the process is producing the intended adaptation. It stops outcome data from functioning as the primary motivational signal, which it is poorly suited to be.

Pelaris is built around this principle, tracking both the inputs and the outputs of training to surface whether the process is working, not just whether the last result was good. The methodology behind it treats consistency and execution quality as first-order variables, not secondary to performance data.

The distinction matters. If your platform only surfaces pace and power, it is only telling you what happened. If it also surfaces completion rate and execution quality, it is starting to tell you why.

What a process-first training log looks like in practice

Concrete example. An intermediate runner, 35-45km per week, targeting a half marathon in 18 weeks.

Their current log: distance, pace, heart rate, weekly mileage total. Outcome-heavy.

A process-first version of the same log adds:

  • Session completion flag (planned vs. actual)
  • Zone accuracy note (target zone vs. actual zone for key sessions)
  • Morning readiness rating (1-5, logged before training)
  • Weekly execution quality average (mean of daily quality ratings)
  • Sleep variance for the week (best night vs. worst night, in minutes)

The outcome data stays. The pace and heart rate files are still there. But now the log is also a record of the process that produced those numbers. When the half marathon result comes in, the athlete can trace it back through 18 weeks of process data and understand what drove it, not just what it was.

More importantly, during those 18 weeks, the athlete has a feedback system that updates daily. The process is visible. The work is legible. The push has its own record, separate from the summit.

That is what makes training sustainable across years rather than just productive across a single race cycle.


What this means for your training

  • Outcome metrics only update at race day. If your entire feedback system depends on results, you have no signal for the 16 weeks of training that determine the result. Add process metrics to close that gap.
  • Session completion rate over a rolling four-week window is the single most predictive process metric for long-term progress. Aim above 85% consistently before optimising anything else.
  • Intensity zone accuracy matters more than volume. A session executed at the wrong intensity is not the session you planned, regardless of what the distance or time file says.
  • Track recovery as a trend, not a daily point. A seven-day rolling average of HRV or readiness tells you far more than yesterday’s number alone.
  • The practical starting point is three fields added to every session entry: completion status, execution quality (1-5), and recovery status going in. Eight weeks of that data will tell you more about your training than a year of outcome-only logging.
  • Deliberate practice means setting an execution target before each session, not just a distance or pace. Review whether you hit it. That review loop is where skill development actually happens.