Abstract

The Ocean Observatories Initiative (OOI) consists of seven research sites with over 800 instruments collecting ocean, seafloor, and meteorological data in the world's oceans, extending from the Irminger Sea to the Southern Ocean. The scale with which data are produced from the instruments and their platforms presents key challenges, including how to (1) build data and information management infrastructure that associates measurements from multiple instruments for concurrent observations, (2) develop data delivery mechanisms that meet a variety of needs, (3) ensure timely release of data, and (4) formulate capabilities to provide stable, longterm support for research and societal needs. As a strategy for maintenance of the sustained, large-scale, and variety of scientific observations collected, the OOI Cyberinfrastructure (CI) developed an end-to-end system designed to store, query, process, and disseminate the compiled information. These resources include raw instrument values and derived data products, metadata associated with instrument and platform deployments (e.g., deployment dates, water depths, instrument manufacturer, etc.), calibration coefficients, data provenance and descriptors of computational algorithms and transformation functions, and other related outputs. As with many data and information management systems, the need to monitor and improve upon the performance of the various interrelated components of the CI is an integral part in establishing the success of the system. For this reason, the data evaluation team at Rutgers initiated an end-to-end system data quality audit that ensures the system can accurately and completely deliver data within the required specifications. This was applied to a subset of representative platforms for all instrument types. Thus, based on specific system failures that can prevent production of quality data products, we prepared a set of tests and built tools that function as a troubleshooting method for system repair and enhancement. The method was used to report on the system's ability to: (a) Respond to data queries (b) Provide links to data product files (c) Produce all relevant data products (d) Produce correct provenance information (e) Produce quality science data products (f) Parse data with the correct specifications and the correct number of particles (g) Calculate different data levels with correct dimensions and units

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call