Rebuilding the Geospatial Supply Chain: A New Push for Spatial Data Quality
“Data quality is a problem we need to address.” When that statement resurfaced in recent industry commentary, it struck a familiar chord. For many organizations working in geospatial technology, the challenge of spatial data quality has been recognized for decades. Yet despite earlier attempts to codify standards and define evaluation frameworks, the issue remains unresolved—though the conditions for progress have changed dramatically.
This discussin aims to connect historical efforts with present opportunities and to outline a practical path forward for achieving meaningful improvements in spatial data quality.
Lessons from Earlier Efforts
Concerns about “fit for purpose” data are not new. In the early 1990s, shortly after the release of the Digital Chart of the World (DCW), practitioners began confronting the consequences of using datasets outside their original design context. The DCW, developed primarily for navigation, was freely available and quickly adopted for applications ranging from hydrology analysis to planning.
As adoption expanded, it became clear that a dataset optimized for one operational context could not automatically satisfy other analytical needs. These experiences led to the development of ISO 19113 and ISO 19114 under ISO/TC 211. These standards articulated principles and procedures for evaluating geographic information, formally acknowledging that datasets are increasingly shared and repurposed.
However, by the mid-1990s, momentum slowed. The primary obstacle was technological: large-scale quantitative assessment of spatial data was computationally impractical. Most quality evaluations remained qualitative, limiting the ability to rigorously assess logical consistency or positional accuracy. Without quantitative tools, the standards lacked operational traction.
Why the Issue Is More Urgent Today
Since that period, the geospatial standards ecosystem has matured significantly. The Open Geospatial Consortium (OGC) has overseen the development of Web Feature Service (WFS), Web Map Service (WMS), and Geography Markup Language (GML). These standards have made distributed datasets widely discoverable and interoperable.
At the same time, service-oriented architectures, semantic web frameworks, and ontology languages have emerged. Yet the Semantic Web Rules Language (SWRL) remains insufficiently aligned with spatial reasoning requirements. As geospatial data proliferate across web registries and portals, quality concerns intensify.
Public and private sectors have invested billions in spatial datasets—many of which lack the positional accuracy demanded by modern GPS-driven applications. The risk is not merely academic. When unreliable data underpin routing, flood modeling, infrastructure planning, or emergency response, the consequences can be serious.
A Framework for Moving Forward
Meaningful progress requires addressing the “supply chain” of spatial data—from creation through distribution and reuse.
Step 1: Define Domain Requirements
Rather than attempting to reinvent standards from scratch, existing thematic frameworks such as those used in DCW or INSPIRE Annex I provide a starting point. The first task is to clarify what users require when aggregating datasets across administrative boundaries. For applications such as transportation routing, flood assessment, or development planning, explicit quality thresholds must be articulated.
This step reframes quality as contextual: fit for what purpose?
Step 2: Establish Practical Quality Measures
Quality descriptors must be manageable and embedded within metadata. Overly complex schemas risk overwhelming data producers. A concise set of performance indicators—analogous to key business metrics—can provide clarity without excessive burden.
Such measures should capture logical consistency, completeness, positional accuracy, temporal validity, and thematic reliability. Crucially, they must be supported by an automated rules expression language capable of evaluating datasets quantitatively. Unlike in 1995, computational power and validation tools are now sufficient to enable large-scale automated assessments.
If a fully mature rules language cannot be developed immediately, interim approaches could incorporate structured peer review models. User communities could contribute pseudo-quantitative ratings, similar to established online evaluation systems, to provide rapid insight into dataset reliability.
Automation is central. Rather than defaulting to caveat emptor, users should be able to evaluate whether data meet their requirements before committing to acquisition or integration.
Step 3: Extend and Repair the Existing Supply Chain
The vast inventory of existing public sector geospatial data cannot be discarded. In the European Union alone, public geographic information assets represent enormous financial investment. The challenge is not replacement, but remediation.
Automated validation rules can serve dual purposes: measuring quality and identifying correctable deficiencies. Where feasible, rule-based processes can enhance datasets, improving their fitness for purpose. If these enhancements generate measurable value, cost-recovery mechanisms can be justified.
The objective is to evolve the supply chain so that legacy datasets become usable in modern contexts such as cross-border routing, urban planning, and emergency management.
Toward Industry Alignment
The geospatial industry stands at a pivotal moment. Data availability has increased dramatically due to open standards and web services. Public visibility has expanded, raising expectations of reliability. The technical barriers that once prevented quantitative assessment have diminished.
Achieving sustained progress requires coordinated engagement through initiatives such as those within the OGC. Standards must evolve from conceptual principles to operational tools that automate measurement, reporting, and improvement of spatial data quality.
The urgency is clear. As spatial data underpin more critical decisions across government, infrastructure, and emergency response, ensuring that these datasets are demonstrably fit for purpose becomes not only a technical imperative, but a societal one.















