When Data Abundance Meets Planning Wisdom
- liam522
- 11 hours ago
- 7 min read

I still remember when we overlaid mobile phone location data onto what we thought was a complete picture of rail demand. Traditional models told a familiar story: home to work, work to home, AM and PM peaks anchored to major employment centres. But they were incomplete.
What appeared was a set of consistent, high-volume movements that didn't line up with recognised peaks. Short-duration, off-peak trips clustered around health facilities, colleges, and logistics employment. Traditional modelling had filtered them out, too fragmented to register in a peak-focused framework. But the data showed they were reliable, repeatable, and sensitive to service gaps.
These passengers were not choosing rail because it was fastest on paper. They were choosing it when it was legible, predictable, and low risk.
Optimising for the wrong question
Planners were optimising for performance metrics that made sense to institutions, not for the decision criteria people actually used. Most planning frameworks privilege speed, capacity, and peak efficiency. Legibility barely features at all.
What the data showed was that many passengers weren't asking "what is fastest?" They were asking "will this work for someone like me, today, without stress?" A service that's two minutes slower but obvious, consistent, and forgiving will outperform a faster but brittle one for large parts of the market.
Off-peak travellers (shift workers, carers, students) were highly sensitive to missed connections, unclear platforms, and opaque ticketing rules. When those risks were removed, demand appeared. The elasticity wasn't about minutes. It was about perceived exposure to failure.
Planners were optimising outputs rather than conditions, tuning the machine to run efficiently under assumed behaviour instead of shaping environments where people could make good decisions with low cognitive load.
What we measure versus what matters
When I started seeing transport planning through the lens of confidence, my relationship with data changed. Some of the things we were measuring very accurately turned out to be only weak proxies for the decisions people were actually making.
We measure punctuality obsessively, but most passengers experience reliability as a narrative, not a statistic. One missed connection can outweigh ten on-time trips. From a confidence perspective, variance and recovery matter more than the mean. Our datasets were good at counting events, poor at capturing how those events were perceived.
The most important drivers of confidence were not visible to our core data sources at all. Confusion, hesitation, abandonment, and silent mode shift leave no trace. When someone quietly switches to a car because the system feels risky, the data records nothing. Traditional datasets are biased toward successful journeys. They under-sample failure.
The missing role: who owns meaning?
In many transport organisations, data ownership is clear but meaning is fragmented. Analytics teams produce outputs. Operations teams monitor performance. Strategy teams write objectives. What's missing is a recognised role that's accountable for turning patterns into judgement.
Big data doesn't fail because it's misunderstood. It fails because no one is authorised to say what it implies.
When post-pandemic demand didn't return to pre-pandemic peaks, the response was to reinterpret what demand meant and adjust the system. The key point isn't the data. It's who's allowed to act on its implications.
This role isn't a modeller. It's closer to a systems editor, someone who curates assumptions, highlights blind spots, and forces trade-offs into the open. Confidence is produced in the seams of the system: handoffs, rules, recovery from failure, moments of doubt. Those seams are invisible if you only optimise what's easy to count.
The automation paradox
We're building systems that are extraordinarily good at producing signals, and increasingly poor at deciding what to do with them. Automation excels at pattern detection and speed. AI systems are getting better at flagging anomalies. But none of these systems can decide which patterns matter, which can be ignored, or which should override an existing objective.
That isn't a technical gap. It's a governance gap.
The promise of big data quietly shifted the centre of gravity. Instead of asking fewer, harder questions, organisations started asking many easier ones. Real-time visibility became a proxy for understanding. But interpretation isn't about immediacy. It's about consequence.
The human interpreter's role isn't to compete with automation, but to frame it. To decide which questions are worth automating. To slow the system down when speed produces false certainty. To recognise when a statistically significant signal is behaviourally irrelevant, and when a weak signal is a leading indicator of something structural.
Organisations are investing heavily in data pipelines without investing in interpretive capacity. The result isn't insight at scale, but ambiguity at scale.
Hyper-local data and the myopia risk
Street-level footfall, dwell time, desire lines: granular data feels actionable because it's close to daily life. The risk is that planners start optimising for what shows up most clearly today, locking in existing patterns and quietly penalising those who are absent because the street doesn't work for them yet.
What prevents that is treating hyper-local data as diagnostic, not declarative. It should tell you where friction exists, not what success looks like. A drop in footfall might signal a failing scheme. It might also signal that through-movement has reduced because the street now feels safer for slower users.
Hyper-local data is most powerful when used after a strategic choice has been made, not before. If you start with the data and let it define the goal, you almost always end up optimising for throughput, convenience for existing users, or commercial yield. Street-level data is excellent at showing who's present. It's almost silent on who's excluded.
The equity trap in big data
Digital traces overwhelmingly reflect the behaviour of people who are already connected, confident, and visible to the system.
Smartphone data, app usage, card transactions, even sensor-based footfall all privilege those who move frequently, carry devices, and interact digitally without friction. The people most affected by service deserts are often the least likely to leave a usable trace. If you're not careful, the data will tell you the system is working precisely because it isn't seeing those it excludes.

The first discipline is to stop treating absence as neutrality. A blank heatmap isn't an empty place. It's an unanswered question. When a dataset shows low activity, the correct response is to ask whether the system has made participation costly, risky, or illegible for certain groups. In equity work, silence is often the signal.
The second step is to invert the burden of proof. Instead of asking disadvantaged areas to demonstrate demand, planners should ask what conditions would need to change for demand to appear. Big data is useful here, not to infer needs directly, but to identify where the system currently fails to invite use. Poor connectivity, indirect routes, unreliable information, punitive fares. These are structural features that can be diagnosed even when people are absent.
This is where qualitative evidence stops being a "nice to have." Community insight, lived experience, and on-the-ground observation aren't supplements to big data. They're counterweights. Without them, you end up optimising for those already served.
Privacy, dignity, and system design
The trade-off is usually described as privacy versus utility, as if better service is something you earn by surrendering data. That framing quietly normalises surveillance as the price of inclusion.
For people who already experience institutional scrutiny (migrants, low-income households, disabled passengers) that isn't a neutral bargain. It's a deterrent.
The way through is to stop equating personal granularity with service quality. Many of the most important planning insights don't require identifying individuals at all. They require understanding conditions: reliability at the edges of the network, transfer penalties, information gaps.
Those are system properties, not personal ones. You can design for them using aggregated, purpose-limited data without tracking who someone is. If your service only works when you know who someone is, where they've been before, and what they might do next, then the service is fragile by design.
People who most need forgiving, legible services are often those least comfortable consenting to pervasive data collection. Designing systems that require digital visibility to unlock fairness effectively asks them to trade dignity for access. That isn't an acceptable default.
What to build first
If I were advising a transport organisation just beginning to build its big data capability today, I'd tell them to build three things before they build anything else. None of them are technical.
1. A named owner of meaning. Create an explicit role that's accountable for interpretation, not reporting. This person needs permission to range across silos and to speak upstream. Their job is to turn patterns into implications, and to say when evidence is good enough to act.
2. A decision framing discipline. Every dataset should arrive with three questions already answered: what decision is this meant to inform, over what time horizon, and what would change if the signal moved. If those answers are missing, the data isn't unfinished. It's misdirected.
3. A tolerance for judgement. Interpretation involves risk. Someone has to say, we will act on this with uncertainty, and we accept the consequences. Build a norm where decisions based on partial evidence are expected, reviewed, and refined, rather than punished.
If an organisation gets these capacities right, the technology choices become easier. What matters is that someone owns meaning, decisions are framed before data arrives, and judgement is treated as a core professional skill rather than a liability.
Build that first, and big data becomes a planning asset rather than a paralysis engine.
The path forward
The shift from forecaster to interpreter is not optional. It is already happening. The question is whether we build the organisational capacity to interpret thoughtfully, or whether we drown in signals we cannot act on.
Big data promised to reduce guesswork. In practice, it has moved guesswork upstream. The choice now is whether to pretend the system is objective, or to accept that interpretation is a core capability that has to be designed, resourced, and protected.
The organisations that move forward will not be the ones with the most sophisticated dashboards or the most granular feeds. They will be the ones that made a conscious choice about which uncertainties they will tolerate, who is authorised to interpret meaning, and what permission exists to be wrong.
When data abundance meets planning wisdom, the result is not automatic insight. It is the hard work of turning patterns into judgement, and judgement into action. That work cannot be automated. It can only be owned.
_edited.jpg)





Comments