Figure 1 shows the number of observatories from which we keep data in analogue and digital forms.
Optical recording on photo paper was originally used for most analogue recording.The digital recording of the data observed by modern electronic magnetometers started to increase from around 1980, and in 1992, finally overtook analogue recording. In 2000, the number of analogue stations had decreased to less than 10% of the total, and now all data are provided in digital form. In mid-1990s, the Internet and World Wide Web became popular, with WDC - Geomagnetism, Kyoto starting its web service in 1995.
The WDCs for Geomagnetism have been exchanging among themselves the data collected at each data centre for 60 years. During the analogue data era, it took money and manpower to collect data from distant observatories and copy them onto microfilms; and the big data centres such as WDC-A in Boulder (now World Data Service for Geophysics) or WDC-B in Moscow (now WDC - Solar-Terrestrial Physics, Moscow) mainly collected the data and distributed them to the other smaller data centres. After shifting to the digital data and Internet era, the situation changed. Collecting data via the Internet is much easier than collecting photo papers from distant stations, and international collaboration is also much easier than before.
Nowadays, more than half of geomagnetic data are provided through an international consortium, INTERMAGNET (WDS Network Member). The transition from analogue to digital recording thus also changed the main player in the provision of geomagnetic data services.
The longest time series of geomagnetic data are certainly the ones acquired by magnetic observatories (Fig. 1), some of which reach a century of uninterrupted measurements.
There are currently about 200 open magnetic observatories worldwide. In each of them, absolute vector observations of the Earth's magnetic field are recorded accurately and continuously, with a time resolution of one minute or less, over a long period of time. Magnetic observatory data are 'primary data' that are extensively used in the derivation of data products ('secondary data') such as: International Geomagnetic Reference Field models, geomagnetic indices, space weather applications…
Figure 1. Paris declination series: annual means of declination corrected and adjusted to actual French National Magnetic Observatory - CLF (Mandea and LeMouël, 2016).
Since the beginning of the 1960’s (the birth of the World Data Centre system established in 1957 provided archives for the observational data resulting from the 'International Geophysical Year'), magnetic observatories data have been mostly publicly available (Fig. 2). Getting access to a network of stations is much more interesting than having access to just one isolated observatory.
Figure 2. Location of magnetic observatories (all periods) having at least one datum ingested into the Geomagnetism Data Portal of WDC – Geomagnetism, Edinburgh [WDS Regular Member].
The cooperative spirit within the geomagnetic community thus knows a fairly long-standing history that has had to cope with the successive technological revolutions regarding data recording (e.g., analogic to numeric; Fig. 3), but also regarding the way data are made available (from yearly books, via isolated recording supports, up to connected data repositories). In this regard, the community had practices based on fair-play and goodwill recognition of data sources/providers. Such practices worked, and would have worked for many more decades without new challenges to meet the changing requirements of users and stakeholders.
Indeed, in our increasingly connected world, it is evermore important to closely follow evolution regarding data management. Some aspects were previously not sufficiently taken into account, such as the discovery, citation, and reuse of the geomagnetic data. Nowadays, it appears no longer possible to keep sources of data for only 'informed people', and the existing licensing conditions for distribution of geomagnetic data and data products are (in part) not adequately elaborated to address this change and need to be improved.
IAGA has thus agreed to set up Task Forces on the abovementioned aspects, with a consensus already found when it comes to the aims of data/ data-product licensing and Digital Object Identifier (DOI) minting to:
– Provide recognition and acknowledgement. – Enable creation of new data products from primary data (e.g., geomagnetic indices) or in combination with other data sources (e.g., global models of geomagnetic field). – Prevent the change and/or appropriation of data by a third party. – Enable reuse of data in a reproducible way. – Supply metadata that enable unique identification of a dataset, as well as providing relevant information to the user. – Use machine-readable and widely used licenses. – Enable easy online access to research data for discovery.
The work is in progress such that it meets the state-of-art when it comes to applying licenses and minting DOI for geomagnetic data and data products, with the goal to ensure the availability into the 21st century of the tremendous efforts achieved by generations of observers in geomagnetism throughout the world.
At the end of October last year, the members of the International Council for Science and the International Social Science Councilvoted overwhelming for a merger of the two organizations. The new organization—which will be called the International Science Council and which should come into being in summer 2018—will serve as a single, global voice for science and will help to provide the evidence base for, and coordinate action on, issues of public concern.
The importance of data for enabling science and for providing the necessary evidence base for action was necessarily both a major concern and discussion point during the meeting at which this historic vote was taken. Access to high-quality data from multiple disciplines is needed to be able to understand and address the complex issues facing our global society. New pluridisciplinary approaches to analyzing and modelling data will be required. And the data upon which decision-making and management of our planet rests must be open access, freely available, and subject to public scrutiny.
So far, so good. However, recognition of the importance of free and open access to data is only the beginning. The new 'voice of science' in the 21st century will need to champion the infrastructure required to ensure free and open access to data. Data stewardship cannot be achieved through pious statements or international accords, it requires the existence of data stewards—organizations that are funded and supported to provide professional support for data archiving, data analysis, and data sharing.
The mission of the World Data System is, of course, to provide an umbrella for data stewards worldwide and to champion new and better ways of ensuring the continuance of our data infrastructure. But there is still a long way to go to ensure both the continued funding for the many organizations that are part of this landscape and that these organizations continue to adopt and promote best data practices.
Too much of the data compilation is currently being done by individual scientists or science teams on short-term funding; too much of the work of data stewardship is currently being done pro bono. Neither of these situations is sustainable. Thus, we must hope that the new International Science Council will make the practical issues of data stewardship in the 21st century a major focus of its work. And then we really will have something to celebrate next summer!
Whether your research is performed in a lab, in the field, or at the office, and with a large or small team, it inevitably involves research information, or data. These data are valuable, and deserve to be properly managed. Over the last few years, the notion that good data management is an important part of scientific practice has increasingly found widespread acceptance.
Data management planning is the structured way of thinking about the research data you are going to collect. What type of research data will the research project produce? What format will you use? How will you store them and how can they be accessed? By thinking about these questions at an early stage and documenting your answers you will avert future problems as a researcher.
One of the ways to think about the data collecting process is by using a format: a Data Management Plan (DMP). These formats come in a variety of shapes and sizes, depending on the research discipline, requirements from the research funder, and local initiatives.
A DMP can be a separate document. It helps the researcher identify and list the risks with regard to management of research data during the entire research process. Because not everything is known from the outset, it is recommended to treat the DMP as a 'living document', which can be revised and detailed periodically.
Research funders nowadays often require that a DMP is included in the project proposal. Typically, the research proposal either contains a data section or a separate DMP is incorporated as an annex. For research funders, the reason behind it is to promote open access to research data: in their opinion, research data produced in the context of a publicly funded research project should be freely made available for reuse and verification. Recent cases of data manipulation and fraud emphasize the importance of access to the original data.
By increasing awareness of research data management across the board—from funders, to researchers, to support staff—we can ensure that research data are handled properly, both in the present and future. More data will be available for reuse, and more data will be reused. In the end, reuse of any form will help contribute to the solutions of today's grand challenges.
The Essentials 4 Data Support is an introductory course for those who provide support to researchers in storing, managing, archiving and sharing their research data (data support staff). With this course, Research Data Netherlands aims to contribute to the professional development of, and coordination among, data support staff. The course covers the basic building blocks of the discipline and revolves around online material. Research Data Netherlands has placed the online learning materials at everyone's disposal free of charge, based on the idea of open access and sharing knowledge. You are free to take the online course anytime at your own initiative. http://datasupport.researchdata.nl/en/about-the-course/
The ICSU World Data System (ICSU-WDS) and the Data Seal of Approval have recently collaborated on the alignment of their respective sets of criteria for certification as a Trusted Digital Repository, and is in process of establishing a joint certification authority—the CoreTrustSeal—to manage the certification process associated with it. This activity contributes to a significant future focus on the trust that can be placed in elements of a distributed global research infrastructure, and the increased automation of its verification. However, it is the tip of the iceberg.
The WDS Knowledge Network defines many of the components of research activity for which there is some form of trusted service or infrastructure component required: ranging from the obvious need to reliably refer to research outputs, researchers, institutions, artefacts, projects, and the like, though the more complex aspects of trusted repositories, registries, vocabulary, and ontology services, to the assigning of levels of maturity, sustainability, or quality to these.
The trust that is required for research infrastructure to function properly is somewhat different to the trust that can be placed in the content that is curated by the research infrastructure—although one has to recognize that the two aspects are interrelated and, in some instances, inseparable. Furthermore, the trust that can be placed in content should ideally also distinguish between the significance and usability of that content, and its quality. These facets are not necessarily the same, but again are conflated to some extent in discussions about fitness-for-use, quality metrics, and the like.
Let’s work though these distinctions at the hand of some examples.
There are arguments that this stream of self-correcting progress is broken, especially in some disciplines, and this is strongly related to the criteria for trust. Such criteria are largely stated informally and implemented with varying degrees of diligence in research institutions, and are mostly delegated to peer review to determine if the result is trustworthy. Peer review purports to determine originality (not easily automated, and essentially linked to end-user value), quality (certainly possible to automate) and validity (can be partly automated).
One could—and in my view, should—argue that processes can be verified objectively and preferably automatically, and that our aim should be to certify their veracity using measurable criteria. Such validity and quality criteria could be extended to feasibility of reproduction, access to supporting datasets, and the like. References to widely used protocols and methods, standards, samples, and research pattern—increasingly linked to persistent identifiers—also increase the verifiable level of trust in the process.
Vocabulary (name) services play an increasingly important role in research infrastructures for a variety of reasons. Firstly, vocabularies and name services are critical to the realization of the semantic web and Linked Open Data: in essence, reducing ambiguity by referring precisely to a concept, entity, relationship, and/or characteristic of either. Secondly, these services are used to enhance the experiences of users and the value of knowledge by navigating the relationships that exist among them, which is conceptually captured in the WDS Knowledge Network and is increasingly implemented, for example, in projects such as Scholix. Again, one should not confuse the acceptability of the vocabulary or service content (e.g., whether all taxonomists in the world agree that a taxon is correct), and the quality of the service provided by the infrastructure component. For the first case, there may never be agreement (especially with taxonomists!); but, for the latter, it is a relatively simple matter to determine what constitutes a well-defined, standardized vocabulary or name service, and community efforts are underway to document and define these criteria. In addition to such operational requirements, one should include the need for sustainability and continued access into the reasonable future.
In general, one can distinguish—for all of the elements of the WDS Knowledge Network—a clear separation between judgements about value (significance, originality, inclusiveness, consensus, etc.) and the quality of the process (sustainability, standards compliance, reproducibility, and similar concerns). And, extrapolating this into the future, I suspect that we need to get ready for the following:
Significant broadening of services and infrastructure that cover all aspects of the WDS Knowledge Network, as well as a parallel rise in the need for certification of these services and infrastructure. Already, there is a perceived need for the certification of repositories of open source code and of vocabulary services, to name but two.
Increased automation of the certification of processes that is in tune with an expected, rapid upturn in artificial intelligence and machine learning. This will be needed because I have no doubt that the scientific method will be increasingly automated within the next decade or so. We are already overwhelmed by volumes of data and numbers of publications, and science cannot scale any further as it is limited by human capacity.
On the basis of the above, and with science increasingly reliant on trust in a wider context, ICSU-WDS should start focussing on defining trust criteria beyond data repositories and services, and on how to automate its assessment: this being the only really scalable solution to a problem of rapidly growing scope.
1 There is a parallel focus on review and consolidation or synthesis based on existing knowledge.
A Blog post by Lindsey M. Harriman (SGT, Inc. Contractor to USGS EROS Center/LP DAAC) and Alex de Sherbinin (WDS Scientific Committee member)
Socioeconomic and Earth Sciences researchers in search of pertinent data can now reap the benefits of a recent collaboration between two Regular Members of the ICSU World Data System.
Today, our planet supports about 7.6 billion people, with a projected increase to nearly 10 billion by 2050, and more than 11 billion by 2100. These 7.6 billion people are using land and water resources to meet their basic needs. As the population increases, their use of, and their impact on, Earth’s resources is going to change. Researchers who study the dynamics between such human–land interactions and their changes over time will look at a range of variables, such as surface temperature, vegetation health, forest cover extent, and change in land cover and habitat, as well as impacts of natural disasters, and climate trends and extremes.
Research questions that often ask about such dynamics include:
What is the proximity between populated areas and fire occurrences over time?
What is the correlation between the increase of population and land surface temperature in urban areas?
How has population affected land-cover change and vegetation growth over time in urban sprawl areas?
How will land-cover changes affect flood and drought risk around rural and urban settlements?
To answer these types of questions, researchers need to integrate census data with Earth observation data, including data collected by NASA’s Earth Science Division Operating Missions. Recently, two NASA Distributed Active Archive Centers (DAACs)—the Land Processes DAAC (LP DAAC; WDS Regular Member) and the Socioeconomic Data and Applications Center (SEDAC; WDS Regular Member)—collaborated to make that integration much easier. LP DAAC and SEDAC worked together to provide access to georeferenced population data alongside land remote sensing data in the Application for Extracting and Exploring Analysis Ready Samples (AppEEARS). SEDAC’s Gridded Population of the World version 4 (GPWv4) aggregates census data from around the world into a globally consistent grid with 30 arc-second resolution (1 kilometer at the equator) for population density and counts. Soon researchers will also have access to age and sex distribution grids. LP DAAC disseminates land remote sensing data collected by several NASA missions—including from the popular Moderate Resolution Imaging Spectroradiometer (MODIS) sensor onboard Terra and Aqua—and provides access to a selection of these datasets through AppEEARS.
Figure 1. Daily land surface temperature in Kelvin (K) and population trend, 2010–2017 for rural and urban points in North Carolina (based on MODIS MOD11A1 daily 1-km data and GPWv4, UN-Adjusted) (a) Farm northwest of Nashville, North Carolina, USA. The red pin represents the location 36°N, 78°W. Image: Google Maps. Time series plots: output from AppEEARS. (b) Suburban area of Charlotte, North Carolina, USA, experiencing rapid population growth. The red pin represents the approximate location 35°N, 81°W). Image: Google Maps. Time series plots: output from AppEEARS.
Figure 1 provides examples of time series plots of population growth and daily land surface temperature using the Point Sample function in AppEEARS. Users can interact with these visualizations within the application and also download the data values in comma separated value format.
Additionally, LP DAAC has collaborated with a third DAAC, the National Snow and Ice Data Center DAAC (NSIDC DAAC; WDS Regular Member), to provide MODIS snow-cover data from its archive for access through AppEEARS as an additional variable describing land dimension. SEDAC, LP DAAC, and NSIDC DAAC are all part of NASA’s Earth Observing System Data and Information System, and through their collaborations, AppEEARS now provides access to more than 100 data products from the three data centers in a single place, at no cost to the user. Many possible combinations of data can be extracted from AppEEARS for use in analyses of the dynamics between populations and ecosystems over time.
AppEEARS also provides benefits during the data preparation process. When performing a sample request, users drastically reduce the amount of data they ultimately need to download to perform their analysis. AppEEARS enables users to subset data based on geographic and temporal parameters, as well as by specific data layer. Since users can reformat the data and reproject within the application, the amount of post-processing required is reduced. Furthermore, AppEEARS not only provides data values, but also quality data values and their descriptions, when applicable. Lastly, users can visualize plots of the data values (point sample) or summary statistics (area samples) from the sample request within the application.
The collaboration around AppEEARS represents an initial step away from the idea that users need to download large amounts of data for local filtering, processing, integration, and analysis, and moves towards a model where analysis-ready data can be more immediately accessed. Coordinated tools and application development on the substantial holdings of all 12 DAACs is an important strategic direction for NASA’s Earth Science Data and Information System Project (WDS Network Member).