In high energy physics (HEP), analysis metadata comes in many forms—from theoretical cross-sections, to calibration corrections, to details about file processing. Correctly applying metadata is a crucial and often time-consuming step in an analysis, but designing analysis metadata systems has historically received little direct attention. Among other considerations, an ideal metadata tool should be easy to use by new analysers, should scale to large data volumes and diverse processing paradigms, and should enable future analysis reinterpretation. This document, which is the product of community discussions organised by the HEP Software Foundation, categorises types of metadata by scope and format and gives examples of current metadata solutions. Important design considerations for metadata systems, including sociological factors, analysis preservation efforts, and technical factors, are discussed. A list of best practices and technical requirements for future analysis metadata systems is presented. These best practices could guide the development of a future cross-experimental effort for analysis metadata tools.
more » « less- NSF-PAR ID:
- 10375325
- Author(s) / Creator(s):
- ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; more »
- Publisher / Repository:
- Springer Science + Business Media
- Date Published:
- Journal Name:
- Computing and Software for Big Science
- Volume:
- 6
- Issue:
- 1
- ISSN:
- 2510-2036
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
null (Ed.)Although engineering graduates are well prepared in the technical aspects of engineering, it is widely acknowledged that there is a need for a greater understanding of the socio-economic contexts in which they will practice their profession. The National Academy of Engineering (NAE) reinforces the critical role that engineers should play in addressing both problems and opportunities that are technical, social, economic, and political in nature in solving the grand challenges. This paper provides an overview of a nascent effort to address this educational need. Through a National Science Foundation (NSF) funded program, a team of researchers at West Virginia University has launched a Holistic Engineering Project Experience (HEPE). This undergraduate course provides the opportunity for engineering students to work with social science students from the fields of economics and strategic communication on complex and open-ended transportation engineering problems. This course involves cross-disciplinary teams working under diverse constraints of real-world social considerations, such as economic impacts, public policy concerns, and public perception and outreach factors, considering the future autonomous transportation systems. The goal of the HEPE platform is for engineering students to have an opportunity to build non-technical—but highly in-demand—professional skills that promote collaboration with others involved in the socio-economic context of engineering matters. Conversely, the HEPE approach provides an opportunity for non-engineering students to become exposed to key concepts and practices in engineering. This paper outlines the initial implementation of the HEPE program, by placing the effort in context of broader trends in education, by outlining the overall purposes of the program, discussing the course design and structure, reviewing the learning experience and outcomes assessment process, and providing preliminary results of a baseline survey that gauges students interests and attitudes towards collaborative and interdisciplinary learning.more » « less
-
Abstract The long-term sustainability of the high-energy physics (HEP) research software ecosystem is essential to the field. With new facilities and upgrades coming online throughout the 2020s, this will only become increasingly important. Meeting the sustainability challenge requires a workforce with a combination of HEP domain knowledge and advanced software skills. The required software skills fall into three broad groups. The first is fundamental and generic software engineering (e.g., Unix, version control, C++, and continuous integration). The second is knowledge of domain-specific HEP packages and practices (e.g., the ROOT data format and analysis framework). The third is more advanced knowledge involving specialized techniques, including parallel programming, machine learning and data science tools, and techniques to maintain software projects at all scales. This paper discusses the collective software training program in HEP led by the HEP Software Foundation (HSF) and the Institute for Research and Innovation in Software in HEP (IRIS-HEP). The program equips participants with an array of software skills that serve as ingredients for the solution of HEP computing challenges. Beyond serving the community by ensuring that members are able to pursue research goals, the program serves individuals by providing intellectual capital and transferable skills important to careers in the realm of software and computing, inside or outside HEP.
-
This paper discusses three aspects of nonlinear dynamic analysis (NDA) practices that are important for evaluating the seismic performance of geotechnical structures affected by liquefaction or cyclic softening: (1) selection and calibration of constitutive models, (2) comparison of NDA results using two or more constitutive models, and (3) documentation. The ability of the selected constitutive models and calibration protocols to approximate the loading responses important to the system being analyzed is one of several technical factors affecting the quality of results from an NDA. Comparisons of single element simulations against empirical data for a broad range of loading conditions are essential for evaluating this factor. Critical comparisons of NDAs using two or more constitutive models are valuable for evaluating modeling uncertainty for specific systems and for identifying modeling limitations that need improvement. The utility of an NDA study depends on the documentation being sufficiently thorough to facilitate effective reviews, advance best practices, and support future reexaminations of a system's seismic performance.more » « less
-
null (Ed.)The systemic challenges of the COVID-19 pandemic require cross-disciplinary collaboration in a global and timely fashion. Such collaboration needs open research practices and the sharing of research outputs, such as data and code, thereby facilitating research and research reproducibility and timely collaboration beyond borders. The Research Data Alliance COVID-19 Working Group recently published a set of recommendations and guidelines on data sharing and related best practices for COVID-19 research. These guidelines include recommendations for researchers, policymakers, funders, publishers and infrastructure providers from the perspective of different domains (Clinical Medicine, Omics, Epidemiology, Social Sciences, Community Participation, Indigenous Peoples, Research Software, Legal and Ethical Considerations). Several overarching themes have emerged from this document such as the need to balance the creation of data adherent to FAIR principles (findable, accessible, interoperable and reusable), with the need for quick data release; the use of trustworthy research data repositories; the use of well-annotated data with meaningful metadata; and practices of documenting methods and software. The resulting document marks an unprecedented cross-disciplinary, cross-sectoral, and cross-jurisdictional effort authored by over 160 experts from around the globe. This letter summarises key points of the Recommendations and Guidelines, highlights the relevant findings, shines a spotlight on the process, and suggests how these developments can be leveraged by the wider scientific community.more » « less
-
Internet of Samples (iSamples): Toward an interdisciplinary cyberinfrastructure for material samplesnull (Ed.)Abstract Sampling the natural world and built environment underpins much of science, yet systems for managing material samples and associated (meta)data are fragmented across institutional catalogs, practices for identification, and discipline-specific (meta)data standards. The Internet of Samples (iSamples) is a standards-based collaboration to uniquely, consistently, and conveniently identify material samples, record core metadata about them, and link them to other samples, data, and research products. iSamples extends existing resources and best practices in data stewardship to render a cross-domain cyberinfrastructure that enables transdisciplinary research, discovery, and reuse of material samples in 21st century natural science.more » « less