skip to main content


Title: Aspects of scaling and scalability for flow-based sampling of lattice QCD
Abstract

Recent applications of machine-learned normalizing flows to sampling in lattice field theory suggest that such methods may be able to mitigate critical slowing down and topological freezing. However, these demonstrations have been at the scale of toy models, and it remains to be determined whether they can be applied to state-of-the-art lattice quantum chromodynamics calculations. Assessing the viability of sampling algorithms for lattice field theory at scale has traditionally been accomplished using simple cost scaling laws, but as we discuss in this work, their utility is limited for flow-based approaches. We conclude that flow-based approaches to sampling are better thought of as a broad family of algorithms with different scaling properties, and that scalability must be assessed experimentally.

 
more » « less
Award ID(s):
2141336
NSF-PAR ID:
10472660
Author(s) / Creator(s):
; ; ; ; ; ; ; ; ; ; ; ;
Publisher / Repository:
Springer Science + Business Media
Date Published:
Journal Name:
The European Physical Journal A
Volume:
59
Issue:
11
ISSN:
1434-601X
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. In this work, we introduce a scalable and efficient GPU-accelerated methodology for volumetric particle advection and finite-time Lyapunov exponent (FTLE) calculation, focusing on the analysis of Lagrangian coherent structures (LCS) in large-scale direct numerical simulation (DNS) datasets across incompressible, supersonic, and hypersonic flow regimes. LCS play a significant role in turbulent boundary layer analysis, and our proposed methodology offers valuable insights into their behavior in various flow conditions. Our novel owning-cell locator method enables efficient constant-time cell search, and the algorithm draws inspiration from classical search algorithms and modern multi-level approaches in numerical linear algebra. The proposed method is implemented for both multi-core CPUs and Nvidia GPUs, demonstrating strong scaling up to 32,768 CPU cores and up to 62 Nvidia V100 GPUs. By decoupling particle advection from other problems, we achieve modularity and extensibility, resulting in consistent parallel efficiency across different architectures. Our methodology was applied to calculate and visualize the FTLE on four turbulent boundary layers at different Reynolds and Mach numbers, revealing that coherent structures grow more isotropic proportional to the Mach number, and their inclination angle varies along the streamwise direction. We also observed increased anisotropy and FTLE organization at lower Reynolds numbers, with structures retaining coherency along both spanwise and streamwise directions. Additionally, we demonstrated the impact of lower temporal frequency sampling by upscaling with an efficient linear upsampler, preserving general trends with only 10% of the required storage. In summary, we present a particle search scheme for particle advection workloads in the context of visualizing LCS via FTLE that exhibits strong scaling performance and efficiency at scale. Our proposed algorithm is applicable across various domains, requiring efficient search algorithms in large, structured domains. While this article focuses on the methodology and its application to LCS, an in-depth study of the physics and compressibility effects in LCS candidates will be explored in a future publication.

     
    more » « less
  2. In this work, we introduce a scalable and efficient GPU-accelerated methodology for volumetric particle advection and finite-time Lyapunov exponent (FTLE) calculation, focusing on the analysis of Lagrangian Coherent Structures (LCS) in large-scale Direct Numerical Simulation (DNS) datasets across incompressible, supersonic, and hypersonic flow regimes. LCS play a significant role in turbulent boundary layer analysis, and our proposed methodology offers valuable insights into their behavior in various flow conditions. Our novel owning-cell locator method enables efficient, constant-time cell search, and the algorithm draws inspiration from classical search algorithms and modern multi-level approaches in numerical linear algebra. The proposed method is implemented for both multi-core CPUs and Nvidia GPUs, demonstrating strong scaling up to 32,768 CPU cores and up to 62 Nvidia V100 GPUs. By decoupling particle advection from other problems, we achieve modularity and extensibility, resulting in consistent parallel efficiency across different architectures. Our methodology was applied to calculate and visualize the FTLE on four turbulent boundary layers at different Reynolds and Mach numbers, revealing that coherent structures grow more isotropic proportional to the Mach number, and their inclination angle varies along the streamwise direction. We also observed increased anisotropy and FTLE organization at lower Reynolds numbers, with structures retaining coherency along both spanwise and streamwise directions. Additionally, we demonstrated the impact of lower temporal frequency sampling by upscaling with an efficient linear upsampler, preserving general trends with only 10% of the required storage. In summary, we present a particle search scheme for particle advection workloads in the context of visualizing LCS via FTLE that exhibits strong scaling performance and efficiency at scale. Our proposed algorithm is applicable across various domains requiring efficient search algorithms in large structured domains. While this manuscript focuses on the methodology and its application to LCS, an in-depth study of the physics and compressibility effects in LCS candidates will be explored in a future publication. 
    more » « less
  3. Abstract

    Non‐native species are now common in community assemblages, but the influence of multiple introductions on ecosystem functioning remains poorly understood. In highly invaded systems, one promising approach is to use functional traits to scale measured individuals’ effects on ecosystem function up to the community level. This approach assumes that functional traits provide a common currency among species to relate individuals to ecosystem functioning.

    The goals of this study were to (i) test whether the relationship between body size and ecosystem functioning (per capita nutrient recycling) was best described by general or species‐specific scaling models; (ii) relate community structure (total biomass, average body size, non‐native dominance) to aggregated, community‐level nutrient recycling rates and ratios; and (iii) determine whether conclusions regarding the relationships between community structure and aggregate ecosystem functioning differed between species‐specific and general scaling approaches.

    By combining experimental incubations and field surveys, we compare consumer‐mediated nutrient recycling of fish communities along a non‐native dominance gradient in the Verde River watershed of central Arizona,USA. Data from ˜340 field‐sampled freshwater fish demonstrated support for general allometric relationships predicted by the metabolic theory of ecology (NH4‐N scaling coefficient = 0.72 [0.64–0.80];PO4‐P = 0.67 [0.47–0.86]). However, the best‐fit models for N and P included species‐specific random effects for both allometric slopes and intercepts.

    According to species‐specific models, stream fish communities recycled 1–12 mmolNH4‐N/hr (median = 2.8 mmol/hr) and 0.02–0.74 mmolPO4‐P/hr (median = 0.07 mmol/hr) at N:P ratios between 13.3 and 83.5 (median = 28.8). General models generated similar estimates forNH4‐N recycling but less accurate estimates forPO4‐P. Stochastic simulations that incorporated error around allometric parameter estimates led to qualitatively similar but larger differences between general and species‐specific results.

    Community structure influenced aggregate nutrient recycling, but specific conclusions depended on the scaling approach. Total biomass explained much of the among‐community variation in aggregateNH4‐N andPO4‐P for both model types, whereas non‐native dominance alone best predicted variation in aggregate N:P. Surprisingly, species‐specific and general models both reached significant yet quantitatively opposing conclusions regarding the relationship between N:P supply and non‐native dominance.

    Study results indicate that shifting fish community structure can substantially alter ecosystem functioning in this river system. However, some inferred relationships between community structure and aggregate nutrient recycling varied depending on whether general or species‐specific scaling approaches were taken. Although trait‐based approaches to link environmental change, community structure and ecosystem function hold much promise, it will be important to consider when species‐specific versus general models are necessary to scale from individuals to ecosystems.

     
    more » « less
  4. Abstract

    Headwater catchments are the fundamental units that connect the land to the ocean. Hydrological flow and biogeochemical processes are intricately coupled, yet their respective sciences have progressed without much integration. Reaction kinetic theories that prescribe rate dependence on environmental variables (e.g., temperature and water content) have advanced substantially, mostly in well‐mixed reactors, columns, and warming experiments without considering the characteristics of hydrological flow at the catchment scale. These theories have shown significant divergence from observations in natural systems. On the other hand, hydrological theories, including transit time theory, have progressed substantially yet have not been incorporated into understanding reactions at the catchment scale. Here we advocate for the development of integrated hydro‐biogeochemical theories across gradients of climate, vegetation, and geology conditions. The lack of such theories presents barriers for understanding mechanisms and forecasting the future of the Critical Zone under human‐ and climate‐induced perturbations. Although integration has started and co‐located measurements are well under way, tremendous challenges remain. In particular, even in this era of “big data,” we are still limited by data and will need to (1) intensify measurements beyond river channels and characterize the vertical connectivity and broadly the shallow and deep subsurface; (2) expand to older water dating beyond the time scales reflected in stable water isotopes; (3) combine the use of reactive solutes, nonreactive tracers, and isotopes; and (4) augment measurements in environments that are undergoing rapid changes. To develop integrated theories, it is essential to (1) engage models at all stages to develop model‐informed data collection strategies and to maximize data usage; (2) adopt a “simple but not simplistic,” or fit‐for‐purpose approach to include essential processes in process‐based models; (3) blend the use of process‐based and data‐driven models in the framework of “theory‐guided data science.” Within the framework of hypothesis testing, model‐data fusion can advance integrated theories that mechanistically link catchments' internal structures and external drivers to their functioning. It can not only advance the field of hydro‐biogeochemistry, but also enable hind‐ and fore‐casting and serve the society at large. Broadly, future education will need to cultivate thinkers at the intersections of traditional disciplines with hollistic approaches for understanding interacting processes in complex earth systems.

    This article is categorized under:

    Engineering Water > Methods

     
    more » « less
  5. Abstract

    The prevalence and intensity of parasites in wild hosts varies across space and is a key determinant of infection risk in humans, domestic animals and threatened wildlife. Because the immune system serves as the primary barrier to infection, replication and transmission following exposure, we here consider the environmental drivers of immunity. Spatial variation in parasite pressure, abiotic and biotic conditions, and anthropogenic factors can all shape immunity across spatial scales. Identifying the most important spatial drivers of immunity could help pre‐empt infectious disease risks, especially in the context of how large‐scale factors such as urbanization affect defence by changing environmental conditions.

    We provide a synthesis of how to apply macroecological approaches to the study of ecoimmunology (i.e. macroimmunology). We first review spatial factors that could generate spatial variation in defence, highlighting the need for large‐scale studies that can differentiate competing environmental predictors of immunity and detailing contexts where this approach might be favoured over small‐scale experimental studies. We next conduct a systematic review of the literature to assess the frequency of spatial studies and to classify them according to taxa, immune measures, spatial replication and extent, and statistical methods.

    We review 210 ecoimmunology studies sampling multiple host populations. We show that whereas spatial approaches are relatively common, spatial replication is generally low and unlikely to provide sufficient environmental variation or power to differentiate competing spatial hypotheses. We also highlight statistical biases in macroimmunology, in that few studies characterize and account for spatial dependence statistically, potentially affecting inferences for the relationships between environmental conditions and immune defence.

    We use these findings to describe tools from geostatistics and spatial modelling that can improve inference about the associations between environmental and immunological variation. In particular, we emphasize exploratory tools that can guide spatial sampling and highlight the need for greater use of mixed‐effects models that account for spatial variability while also allowing researchers to account for both individual‐ and habitat‐level covariates.

    We finally discuss future research priorities for macroimmunology, including focusing on latitudinal gradients, range expansions and urbanization as being especially amenable to large‐scale spatial approaches. Methodologically, we highlight critical opportunities posed by assessing spatial variation in host tolerance, using metagenomics to quantify spatial variation in parasite pressure, coupling large‐scale field studies with small‐scale field experiments and longitudinal approaches, and applying statistical tools from macroecology and meta‐analysis to identify generalizable spatial patterns. Such work will facilitate scaling ecoimmunology from individual‐ to habitat‐level insights about the drivers of immune defence and help predict where environmental change may most alter infectious disease risk.

     
    more » « less