The National Science Foundation Ice Core Facility (NSF-ICF, fka NICL) is in the process of building a new facility including freezer and scientist support space. The facility is being designed to minimize environmental impacts while maximizing ice core curation and science support. In preparation for the new facility, we are updating research equipment and integrating ice core data collection and processing by assigning International Generic Sample Numbers (IGSN) to advance the “FAIR”ness and establish clear provenance of samples, fostering the next generation of linked research data products. The NSF-ICF team, in collaboration with the US ice core science community, has established a metadata schema for the assignment of IGSNs to ice cores and samples. In addition, in close coordination with the US ice core community, we are adding equipment modules that expand traditional sets of physical property, visual stratigraphy, and electrical conductance ice core measurements. One such module is an ice core hyperspectral imaging (HSI) system. Adapted for the cold laboratory settings, the SPECIM SisuSCS HSI system can collect up to 224 bands using a continuous line-scanning mode in the visible and near-infrared (VNIR) 400-1000 nm spectral region. A modular system design allows expansion of spectral properties in the future. The second module is an updated multitrack electrical conductance system. These new data will guide real time optimization of sampling for planned analyses during ice core processing, especially for ice with deformed or highly compressed layering. The aim is to facilitate the collection of robust, long-term, and FAIR data archives for every future ice core section processed at NSF-ICF. The NSF-ICF is fully funded by the National Science Foundation and operated by the U.S. Geological Survey.
more »
« less
GeospaceLAB: Python package for managing and visualizing data in space physics
In the space physics community, processing and combining observational and modeling data from various sources is a demanding task because they often have different formats and use different coordinate systems. The Python package GeospaceLAB has been developed to provide a unified, standardized framework to process data. The package is composed of six core modules, including DataHub as the data manager, Visualization for generating publication quality figures, Express for higher-level interfaces of DataHub and Visualization , SpaceCoordinateSystem for coordinate system transformations, Toolbox for various utilities, and Configuration for preferences. The core modules form a standardized framework for downloading, storing, post-processing and visualizing data in space physics. The object-oriented design makes the core modules of GeospaceLAB easy to modify and extend. So far, GeospaceLAB can process more than twenty kinds of data products from nine databases, and the number will increase in the future. The data sources include, e.g., measurements by EISCAT incoherent scatter radars, DMSP, SWARM, and Grace satellites, OMNI solar wind data, and GITM simulations. In addition, the package provides an interface for the users to add their own data products. Hence, researchers can easily collect, combine, and view multiple kinds of data for their work using GeospaceLAB. Combining data from different sources will lead to a better understanding of the physics of the studied phenomena and may lead to new discoveries. GeospaceLAB is an open source software, which is hosted on GitHub. We welcome everyone in the community to contribute to its future development.
more »
« less
- PAR ID:
- 10386431
- Date Published:
- Journal Name:
- Frontiers in Astronomy and Space Sciences
- Volume:
- 9
- ISSN:
- 2296-987X
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
AbstractÐPrivacy of data as well as providing anonymization of data for various kinds of analysis have been addressed in the context of tabular transactional data which was mainstream. With the advent of the Internet and social networks, there is an emphasis on using different kinds of graphs for modeling and analysis. In addition to single graphs, the use of MultiLayer Networks (or MLNs) for modeling and analysis is becoming popular for complex data having multiple types of entities and relationships. They provide a better understanding of data as well as flexibility and efficiency of analysis. In this article, we understand the provenance of data privacy and some of the thinking on extending it to graph data models. We will focus on the issues of data privacy for models that are different from traditional data models and discuss alternatives. We will also consider privacy from a visualization perspective as we have developed a community Dashboard for MLN generation, analysis, and visualization based on our research.more » « less
-
Abstract Atmospheric aerosol and chemistry modules are key elements in Earth system models (ESMs), as they predict air pollutant concentrations and properties that can impact human health, weather, and climate. The current uncertainty in climate projections is partly due to the inaccurate representation of aerosol direct and indirect forcing. Aerosol/chemistry parameterizations used within ESMs and other atmospheric models span large structural and parameter uncertainties that are difficult to assess independently of their host models. Moreover, there is a strong need for a standardized interface between aerosol/chemistry modules and the host model to facilitate portability of aerosol/chemistry parameterizations from one model to another, allowing not only a comparison between different parameterizations within the same modeling framework, but also quantifying the impact of different model frameworks on aerosol/chemistry predictions. To address this need, we have initiated a new community effort to coordinate the construction of a Generalized Aerosol/Chemistry Interface (GIANT) for use across weather and climate models. We aim to organize a series of community workshops and hackathons to design and build GIANT, which will serve as the interface between a range of aerosol/chemistry modules and the physics and dynamics components of atmospheric host models. GIANT will leverage ongoing efforts at the U.S. modeling centers focused on building next-generation ESMs and the international AeroCom initiative to implement this common aerosol/chemistry interface. GIANT will create transformative opportunities for scientists and students to conduct innovative research to better characterize structural and parametric uncertainties in aerosol/chemistry modules, and to develop a common set of aerosol/chemistry parameterizations.more » « less
-
Abstract The National Ecological Observatory Network (NEON) provides over 180 distinct data products from 81 sites (47 terrestrial and 34 freshwater aquatic sites) within the United States and Puerto Rico. These data products include both field and remote sensing data collected using standardized protocols and sampling schema, with centralized quality assurance and quality control (QA/QC) provided by NEON staff. Such breadth of data creates opportunities for the research community to extend basic and applied research while also extending the impact and reach of NEON data through the creation of derived data products—higher level data products derived by the user community from NEON data. Derived data products are curated, documented, reproducibly‐generated datasets created by applying various processing steps to one or more lower level data products—including interpolation, extrapolation, integration, statistical analysis, modeling, or transformations. Derived data products directly benefit the research community and increase the impact of NEON data by broadening the size and diversity of the user base, decreasing the time and effort needed for working with NEON data, providing primary research foci through the development via the derivation process, and helping users address multidisciplinary questions. Creating derived data products also promotes personal career advancement to those involved through publications, citations, and future grant proposals. However, the creation of derived data products is a nontrivial task. Here we provide an overview of the process of creating derived data products while outlining the advantages, challenges, and major considerations.more » « less
-
ABSTRACT Cosmological simulations serve as invaluable tools for understanding the Universe. However, the technical complexity and substantial computational resources required to generate such simulations often limit their accessibility within the broader research community. Notable exceptions exist, but most are not suited for simultaneously studying the physics of galaxy formation and cosmic reionization during the first billion years of cosmic history. This is especially relevant now that a fleet of advanced observatories (e.g. James Webb Space Telescope, Nancy Grace Roman Space Telescope, SPHEREx, ELT, SKA) will soon provide an holistic picture of this defining epoch. To bridge this gap, we publicly release all simulation outputs and post-processing products generated within the thesan simulation project at www.thesan-project.com. This project focuses on the z ≥ 5.5 Universe, combining a radiation-hydrodynamics solver (arepo-rt), a well-tested galaxy formation model (IllustrisTNG) and cosmic dust physics to provide a comprehensive view of the Epoch of Reionization. The thesan suite includes 16 distinct simulations, each varying in volume, resolution, and underlying physical models. This paper outlines the unique features of these new simulations, the production and detailed format of the wide range of derived data products, and the process for data retrieval. Finally, as a case study, we compare our simulation data with a number of recent observations from the James Webb Space Telescope, affirming the accuracy and applicability of thesan. The examples also serve as prototypes for how to utilize the released data set to perform comparisons between predictions and observations.more » « less
An official website of the United States government

