Air Quality Data Network
< Back to | Workshops | Air Quality Data Network
What is the AQ data network?
Air quality data from mandated and scientific observations and numerical model results have long been gathered, analyzed and distributed through various channels. Data providers and their "customers" have established working procedures allowing them to exchange the data they need to "get their job done". However, as could be demonstrated in the case of individual air pollution events, it is rather difficult to obtain additional data from other sources when needed. Historically, the existing networks have developed regionally and there are as many different procedures to provide and exchange data as there are data networks. With interoperability as a guiding principle, the GEO AQ CoP tries to enable connections across regions and institutional borders. When fully implemented, a truly interoperable data network will make it much easier for the users of AQ data to find relevant data sets, to obtain the portion of these data they are interested in, and to analyze different data sets in a coherent fashion.
What is needed to achieve the goal of interoperability?
The most fundamental change towards interoperable data exchange is the replacement of 1:1 provider-user relationships with user-driven machine-to-machine connections. Just as when visiting foreign countries, this requires some agreement about the "language" in which information is exchanged. In the data network, these are the "protocols" (e.g. WCS, WFS, WMS, SOS, etc.) and the data formats (e.g. netcdf, csv, etc.). A client application and a data server must support the same protocol so that it is possible to obtain information about the data that are stored on the server and to formulate specific requests for data extraction. Once the data arrive, the client must be able to handle the data format in which the content is packaged.
Beyond the basic data access described above, there are other important technical and non-technical elements without which interoperability cannot be achieved. In order to find relevant data sets (the first step in each analysis), data catalogues must be established and agreement must be sought with respect to the metadata facets which describe the actual data sets. Related to this is the definition of a common vocabulary (for example species names, platform or instrument names, etc.). Furthermore, individual data providers may have specific requirements concerning the traceability of data, descriptions of ownership, tracing of data users or data access restrictions. There are different ways to define a "network topology" consisting of client and catalogue applications, data servers, mediators, web processing tools, etc.. Some agreement must be reached how this topology may look like in the AQ world in order to achieve the goal of integrating data and to avoid the creation of a new splinter network.
In August 2011, a meeting was organized upon invitation by the GEO AQ CoP in order to take stock of the existing AQ data network and its interoperability status (see | Solta workshop 2011) and to discuss the challenges and potential solutions on the technical and networking level. From the set-up of this workshop, a number of specific topics were identified. The links below lead to summaries of the workshop discussions and serve as a basis for further elaboration which shall proceed in this Wiki.
Interoperable AQ network status
Network Topology (ToBeDone)
|
|
|
Technical Issues
Implementing a fully interoperable data network requires agreements about a number of issues on the technical level. This includes for example the selection of [| OGC] standard(s) for geographical information exchange, the packaging of the data content (data format), and the metadata description of individual data sets and data catalogues. Eventually, all of these issues must be realized in computer code, and such code needs to be maintained and adapted to changing requirements.
The GEO AQ CoP is providing an open source community WCS server software which has demonstrated its functionality over the past years in the [DataFed] and [FZ Jülich] environments. At the Solta 2011 this code was adopted as reference implementation for the AQ data network, and several participants pledged to install the software at their data nodes or to contribute to the further development of the server functionality.
The sections below briefly summarize the main technical issues related to functioning of the AQ data network and the community WCS server tool. The section headings link to more detailed discussions about each of these issues.
Standards and Conventions to enable the AQ Data Network
- [| OGC] Web Coverage Service (WCS)
- [| Network Common Data Format (netCDF)] data model
- [| Climate and Forecasting (CF)] metadata conventions
- [| ISO 19115] metadata requirements
WCS Server Software
Network Users