Copyright © 2013 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C liability, trademark and document use rules apply.
Many national, regional and local governments, as well as other organisations in- and outside of the public sector, collect numeric data and aggregate this data into statistics. There is a need to publish these statistics in a standardised, machine-readable way on the web, so that they can be freely integrated and reused in consuming applications.
In this document, the W3C Government Linked Data Working Group presents use cases and requirements supporting a recommendation of the RDF Data Cube Vocabulary [QB-2013]. The group obtained use cases from existing deployments of and experiences with an earlier version of the data cube vocabulary [QB-2010]. The group also describes a set of requirements derived from the use cases and to be considered in the recommendation.
This section describes the status of this document at the time of its publication. Other documents may supersede this document. A list of current W3C publications and the latest revision of this technical report can be found in the W3C technical reports index at http://www.w3.org/TR/.
This document is an editorial update to an Editor's Draft of the "Use Cases and Requirements for the Data Cube Vocabulary" developed by the W3C Government Linked Data Working Group.
This document was published by the Government Linked Data Working Group as a Working Group Note. If you wish to make comments regarding this document, please send them to public-gld-comments@w3.org (subscribe, archives). All comments are welcome.
Publication as a Working Group Note does not imply endorsement by the W3C Membership. This is a draft document and may be updated, replaced or obsoleted by other documents at any time. It is inappropriate to cite this document as other than work in progress.
This document was produced by a group operating under the 5 February 2004 W3C Patent Policy . W3C maintains a public list of any patent disclosures made in connection with the deliverables of the group; that page also includes instructions for disclosing a patent. An individual who has actual knowledge of a patent which the individual believes contains Essential Claim(s) must disclose the information in accordance with section 6 of the W3C Patent Policy .
The rest of this document is structured as follows. We will first give a short introduction of the specificities of modelling statistics. Then, we will describe use cases that have been derived from existing deployments or feedback to the earlier data cube vocabulary version. In particular, we describe possible benefits and challenges of use cases. Afterwards, we will describe concrete requirements that were derived from those use cases and that have been taken into account for the specification.
We use the name data cube vocabulary throughout the document when referring to the vocabulary.
In the following, we describe the challenge of an RDF vocabulary for publishing statistics as Linked Data.
Describing statistics - collected and aggregated numeric data - is challenging for the following reasons:
The Statistical Data and Metadata eXchange [SDMX] - the ISO standard for exchanging and sharing of statistical data and metadata among organisations - uses "multidimensional model" that caters for the specificity of modelling statistics. It allows to describe statistics as observations. Observations exhibit values (Measures) that depend on dimensions (Members of Dimensions).
Since the SDMX standard has proven applicable in many contexts, the vocabulary adopts the multidimensional model that underlies SDMX and will be compatible to SDMX.
Statistics is the study of the collection, organisation, analysis, and interpretation of data. Statistics comprise statistical data.
The basic structure of statistical data is a multidimensional table (also called a data cube) [SDMX], i.e., a set of observed values organized along a group of dimensions, together with associated metadata. If aggregated we refer to statistical data as "macro-data" whereas if not, we refer to "micro-data".
Statistical data can be collected in a dataset , typically published and maintained by an organisation [SDMX]. The dataset contains metadata, e.g., about the time of collection and publication or about the maintaining and publishing organisation.
Source data is data from datastores such as RDBs or spreadsheets that acts as a source for the Linked Data publishing process.
Metadata about statistics defines the data structure and give contextual information about the statistics.
A format is machine-readable if it is amenable to automated processing by a machine, as opposed to presentation to a human user.
A publisher is a person or organisation that exposes source data as Linked Data on the Web.
A consumer is a person or agent that uses Linked Data from the Web.
A registry collects metadata about statistical data in a registration fashion.
This section presents scenarios that are enabled by the existence of a standard vocabulary for the representation of statistics as Linked Data.
(Use case taken from SDMX Web Dissemination Use Case [SDMX 2.1])
Since we have adopted the multidimensional model that underlies SDMX, we also adopt the "Web Dissemination Use Case" which is the prime use case for SDMX since it is an increasing popular use of SDMX and enables organisations to build a self-updating dissemination system.
The Web Dissemination Use Case contains three actors, a structural metadata web service (registry) that collects metadata about statistical data in a registration fashion, a data web service (publisher) that publishes statistical data and its metadata as registered in the structural metadata web service, and a data consumption application (consumer) that first discovers data from the registry, then queries data from the corresponding publisher of selected data, and then visualises the data.
In the following, we illustrate the processes from this use case in a flow diagram by SDMX and describe what activities are enabled in this use case by having statistics described in a machine-readable format.
Benefits:
Requirements:
The SDMX Web Dissemination Use Case can be concretised by several sub-use cases, detailed in the following sections.
(This use case has been summarised from Ian Dickinson et al. [COINS])
More and more organisations want to publish statistics on the web, for reasons such as increasing transparency and trust. Although in the ideal case, published data can be understood by both humans and machines, data often is simply published as CSV, PDF, XSL etc., lacking elaborate metadata, which makes free usage and analysis difficult.
Therefore, the goal in this use case is to use a machine-readable and application-independent description of common statistics with use of open standards, to foster usage and innovation on the published data.
In the "COINS as Linked Data" project [COINS], the Combined Online Information System (COINS) shall be published using a standard Linked Data vocabulary.
Via the Combined Online Information System (COINS), HM Treasury, the principal custodian of financial data for the UK government, releases previously restricted financial information about government spendings.
The COINS data has a hypercube structure. It describes financial transactions using seven independent dimensions (time, data-type, department etc.) and one dependent measure (value). Also, it allows thirty-three attributes that may further describe each transaction.
COINS is an example of one of the more complex statistical datasets being publishing via data.gov.uk.
Part of the complexity of COINS arises from the nature of the data being released.
The published COINS datasets cover expenditure related to five different years (2005–06 to 2009–10). The actual COINS database at HM Treasury is updated daily. In principle at least, multiple snapshots of the COINS data could be released through the year.
According to the COINS as Linked Data project, the reason for publishing COINS as Linked Data are threefold:
The COINS use case leads to the following challenges:
Requirements::
(Part of this use case has been contributed by Rinke Hoekstra. See CEDA_R and Data2Semantics for more information.)
Not only in government, there is a need to publish considerable amounts of statistical data to be consumed in various (also unexpected) application scenarios. Typically, Microsoft Excel sheets are made available for download. Those excel sheets contain single spreadsheets with several multidimensional data tables, having a name and notes, as well as column values, row values, and cell values.
Benefits:
Challenges in this use case:
Existing work:
Requirements:
(Use case has been taken from [QB4OLAP] and from discussions at publishing-statistical-data mailing list)
It often comes up in statistical data that you have some kind of 'overall' figure, which is then broken down into parts.
Example (in pseudo-turtle RDF):
ex:obs1 sdmx:refArea ; sdmx:refPeriod "2011"; ex:population "60" . ex:obs2 sdmx:refArea ; sdmx:refPeriod "2011"; ex:population "50" . ex:obs3 sdmx:refArea ; sdmx:refPeriod "2011"; ex:population "5" . ex:obs4 sdmx:refArea ; sdmx:refPeriod "2011"; ex:population "3" . ex:obs5 sdmx:refArea ; sdmx:refPeriod "2011"; ex:population "2" .
We are looking for the best way (in the context of the RDF/Data
Cube/SDMX approach) to express that the values for the
England/Scotland/Wales/ Northern Ireland ought to add up to the value
for the UK and constitute a more detailed breakdown of the overall UK
figure? Since we might also have population figures for France,
Germany, EU27, it is not as simple as just taking a
qb:Slice
where you fix the time period and the measure.
Similarly, Etcheverry and Vaisman [QB4OLAP] present the use case to publish household data from StatsWales and Open Data Communities.
This multidimensional data contains for each fact a time dimension with one level Year and a location dimension with levels Unitary Authority, Government Office Region, Country, and ALL.
As unit, units of 1000 households is used.
In this use case, one wants to publish not only a dataset on the bottom most level, i.e. what are the number of households at each Unitary Authority in each year, but also a dataset on more aggregated levels.
For instance, in order to publish a dataset with the number of households at each Government Office Region per year, one needs to aggregate the measure of each fact having the same Government Office Region using the SUM function.
Importantly, one would like to maintain the relationship between the resulting datasets, i.e., the levels and aggregation functions.
Again, this use case does not simply need a selection (or "dice" in OLAP context) where one fixes the time period dimension.
Requirements:
(Use case has been provided by Epimorphics Ltd, in their UK Bathing Water Quality deployment)
As part of their work with data.gov.uk and the UK Location Programme Epimorphics Ltd have been working to pilot the publication of both current and historic bathing water quality information from the UK Environment Agency as Linked Data.
The UK has a number of areas, typically beaches, that are designated as bathing waters where people routinely enter the water. The Environment Agency monitors and reports on the quality of the water at these bathing waters.
The Environement Agency's data can be thought of as structured in 3 groups:
The most important dimensions of the data are bathing water, sampling point, and compliance classification.
Challenges:
Existing Work:
Requirements:
(This use case has been taken from Eurostat Linked Data Wrapper and Linked Statistics Eurostat Data, both deployments for publishing Eurostat SDMX as Linked Data using the draft version of the data cube vocabulary)
As mentioned already, the ISO standard for exchanging and sharing statistical data and metadata among organisations is Statistical Data and Metadata eXchange [SDMX]. Since this standard has proven applicable in many contexts, we adopt the multidimensional model that underlies SDMX and intend the standard vocabulary to be compatible to SDMX.
Therefore, in this use case we intend to explain the benefit and challenges of publishing SDMX data as Linked Data. As one of the main adopters of SDMX, Eurostat publishes large amounts of European statistics coming from a data warehouse as SDMX and other formats on the web. Eurostat also provides an interface to browse and explore the datasets. However, linking such multidimensional data to related data sets and concepts would require downloading of interesting datasets and manual integration.The goal here is to improve integration with other datasets; Eurostat data should be published on the web in a machine-readable format, possible to be linked with other datasets, and possible to be freeley consumed by applications. Both Eurostat Linked Data Wrapper and Linked Statistics Eurostat Data intend to publish Eurostat SDMX data as 5-star Linked Open Data. Eurostat data is partly published as SDMX, partly as tabular data (TSV, similar to CSV). Eurostat provides a TOC of published datasets as well as a feed of modified and new datasets. Eurostat provides a list of used codelists, i.e., range of permitted dimension values. Any Eurostat dataset contains a varying set of dimensions (e.g., date, geo, obs_status, sex, unit) as well as measures (generic value, content is specified by dataset, e.g., GDP per capita in PPS, Total population, Employment rate by sex).
Benefits:
Challenges:
Non-requirements:
Requirements:
(This use case has mainly been taken from the COINS project [COINS])
In several applications, relationships between statistical data need to be represented.
The goal of this use case is to describe provenance, transformations, and versioning around statistical data, so that the history of statistics published on the web becomes clear. This may also relate to the issue of having relationships between datasets published.
For instance, the COINS project [COINS] has at least four perspectives on what they mean by "COINS" data: the abstract notion of "all of COINS", the data for a particular year, the version of the data for a particular year released on a given date, and the constituent graphs which hold both the authoritative data translated from HMT's own sources. Also, additional supplementary information which they derive from the data, for example by cross-linking to other datasets.
Another specific use case is that the Welsh Assembly government publishes a variety of population datasets broken down in different ways. For many uses then population broken down by some category (e.g. ethnicity) is expressed as a percentage. Separate datasets give the actual counts per category and aggregate counts. In such cases it is common to talk about the denominator (often DENOM) which is the aggregate count against which the percentages can be interpreted.
Another example for representing relationships between statistical data are transformations on datasets, e.g., addition of derived measures, conversion of units, aggregations, OLAP operations, and enrichment of statistical data. A concrete example is given by Freitas et al. [COGS] and illustrated in the following figure.
Here, numbers from a sustainability report have been created by a number of transformations to statistical data. Different numbers (e.g., 600 for year 2009 and 503 for year 2010) might have been created differently, leading to different reliabilities to compare both numbers.
Benefits:
Making transparent the transformation a dataset has been exposed to. Increases trust in the data.
Challenges:
qb:DataSet
(e.g. ex:populationCount
and ex:populationPercent
)?
Existing Work:
Requirements:
(Use case taken from SMART research project)
Data that is published on the Web is typically visualized by transforming it manually into CSV or Excel and then creating a visualization on top of these formats using Excel, Tableau, RapidMiner, Rattle, Weka etc.
This use case shall demonstrate how statistical data published on the web can be with few effort visualized inside a webpage, without using commercial or highly-complex tools.
An example scenario is environmental research done within the SMART research project. Here, statistics about environmental aspects (e.g., measurements about the climate in the Lower Jordan Valley) shall be visualized for scientists and decision makers. Statistics should also be possible to be integrated and displayed together. The data is available as XML files on the web. On a separate website, specific parts of the data shall be queried and visualized in simple charts, e.g., line diagrams.
Challenges of this use case are:
Requirements:
(Use case taken from Google Public Data Explorer (GPDE))
Google Public Data Explorer (GPDE) provides an easy possibility to visualize and explore statistical data. Data needs to be in the Dataset Publishing Language (DSPL) to be uploaded to the data explorer. A DSPL dataset is a bundle that contains an XML file, the schema, and a set of CSV files, the actual data. Google provides a tutorial to create a DSPL dataset from your data, e.g., in CSV. This requires a good understanding of XML, as well as a good understanding of the data that shall be visualized and explored.
In this use case, the goal is to take statistical data published on the web and to transform it into DSPL for visualization and exploration with as few effort as possible.
For instance, Eurostat data about Unemployment rate downloaded from the web as shown in the following figure:
Benefits:
Challenges of this use case are:
Non-requirements:
Requirements:
(Use case taken from Financial Information Observation System (FIOS))
Online Analytical Processing (OLAP) [OLAP] is an analysis method on multidimensional data. It is an explorative analysis methode that allows users to interactively view the data on different angles (rotate, select) or granularities (drill-down, roll-up), and filter it for specific information (slice, dice).
OLAP systems that first use ETL pipelines to Extract-Load-Transform relevant data for efficient storage and queries in a data warehouse and then allows interfaces to issue OLAP queries on the data are commonly used in industry to analyse statistical data on a regular basis.
The goal in this use case is to allow analysis of published statistical data with common OLAP systems [OLAP4LD]
For that a multidimensional model of the data needs to be generated. A multidimensional model consists of facts summarised in data cubes. Facts exhibit measures depending on members of dimensions. Members of dimensions can be further structured along hierarchies of levels.
An example scenario of this use case is the Financial Information Observation System (FIOS) [FIOS], where XBRL data provided by the SEC on the web is to be re-published as Linked Data and made possible to explore and analyse by stakeholders in a web-based OLAP client Saiku.
The following figure shows an example of using FIOS. Here, for three different companies, cost of goods sold as disclosed in XBRL documents are analysed. As cell values either the number of disclosures or - if only one available - the actual number in USD is given:
Benefits:
Challenges:
Requirements:
(Use case motivated by Data Catalog vocabulary)
After statistics have been published as Linked Data, the question remains how to communicate the publication and let users discover the statistics. There are catalogs to register datasets, e.g., CKAN, datacite.org, da|ra, and Pangea. Those catalogs require specific configurations to register statistical data.
The goal of this use case is to demonstrate how to expose and distribute statistics after publication. For instance, to allow automatic registration of statistical data in such catalogs, for finding and evaluating datasets. To solve this issue, it should be possible to transform the published statistical data into formats that can be used by data catalogs.
A concrete use case is the structured collection of RDF Data Cube Vocabulary datasets in the PlanetData Wiki. This list is supposed to describe statistical datasets on a higher level - for easy discovery and selection - and to provide a useful overview of RDF Data Cube deployments in the Linked Data cloud.
Unanticipated Uses:
Existing Work:
Requirements:
The use cases presented in the previous section give rise to the following requirements for a standard representation of statistics. Requirements are cross-linked with the use cases that motivate them.
The draft version of the vocabulary builds upon SDMX Standards Version 2.0. A newer version of SDMX, SDMX Standards, Version 2.1, is available.
The requirement is to at least build upon Version 2.0, if specific use cases derived from Version 2.1 become available, the working group may consider building upon Version 2.1.
Background information:
Required by:
There should be a consensus on the issue of flattening or abbreviating data; one suggestion is to author data without the duplication, but have the data publication tools "flatten" the compact representation into standalone observations during the publication process.
Background information:
Required by:
First, hierarchical code lists may be supported via SKOS [SKOS]. Allow for cross-location and cross-time analysis of statistical datasets.
Second, one can think of non-SKOS hierarchical code lists. E.g., if
simple
skos:narrower
/
skos:broader
relationships are not sufficient or if a vocabulary uses specific
hierarchical properties, e.g.,
geo:containedIn
.
Also, the use of hierarchy levels needs to be clarified. It has been
suggested, to allow
skos:Collections
as value of
qb:codeList
.
Background information:
Required by:
An number of organisations, particularly in the Climate and Meteorological area already have some commitment to the OGC "Observations and Measurements" (O&M) logical data model, also published as ISO 19156. Are there any statements about compatibility and interoperability between O&M and Data Cube that can be made to give guidance to such organisations?
Background information:
Required by:
Background information:
Required by:
Background information:
Required by:
Background information:
Required by:
Background information:
Required by:
Clarify the relationship between DCAT and QB.
Background information:
Required by:
We thank John Erickson, Rinke Hoekstra, Bernadette Hyland, Aftab Iqbal, Dave Reynolds, Biplav Srivastava, Villazón-Terrazas for feedback and input.