skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Towards a data-driven model of hadronization using normalizing flows
We introduce a model of hadronization based on invertible neural networks that faithfully reproduces a simplified version of the Lund string model for meson hadronization. Additionally, we introduce a new training method for normalizing flows, termed MAGIC, that improves the agreement between simulated and experimental distributions of high-level (macroscopic) observables by adjusting single-emission (microscopic) dynamics. Our results constitute an important step toward realizing a machine-learning based model of hadronization that utilizes experimental data during training. Finally, we demonstrate how a Bayesian extension to this normalizing-flow architecture can be used to provide analysis of statistical and modeling uncertainties on the generated observable distributions.  more » « less
Award ID(s):
2103889
PAR ID:
10615030
Author(s) / Creator(s):
; ; ; ; ; ; ;
Publisher / Repository:
SciPost
Date Published:
Journal Name:
SciPost Physics
Volume:
17
Issue:
2
ISSN:
2542-4653
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. We introduce a novel method for extracting a fragmentation model directly from experimental data without requiring an explicit parametric form, called Histories and Observables for Monte-Carlo Event Reweighting (HOMER), consisting of three steps: the training of a classifier between simulation and data, the inference of single fragmentation weights, and the calculation of the weight for the full hadronization chain. We illustrate the use of HOMER on a simplified hadronization problem, aq\bar{q} q q string fragmenting into pions, and extract a modified Lund string fragmentation functionf(z) f ( z ) . We then demonstrate the use of HOMER on three types of experimental data: (i) binned distributions of high-level observables, (ii) unbinned event-by-event distributions of these observables, and (iii) full particle cloud information. After demonstrating thatf(z) f ( z ) can be extracted from data (the inverse of hadronization), we also show that, at least in this limited setup, the fidelity of the extractedf(z) f ( z ) suffers only limited loss when moving from (i) to (ii) to (iii). Public code is available at https://gitlab.com/uchep/mlhad. 
    more » « less
  2. Abstract Monitoring machine health and product quality enables predictive maintenance that optimizes repairs to minimize factory downtime. Data-driven intelligent manufacturing often relies on probabilistic techniques with intractable distributions. For example, generative models of data distributions can balance fault classes with synthetic data, and sampling the posterior distribution of hidden model parameters enables prognosis of degradation trends. Normalizing flows can address these problems while avoiding the training instability or long inference times of other generative Deep Learning (DL) models like Generative Adversarial Networks (GAN), Variational Autoencoders (VAE), and diffusion networks. To evaluate normalizing flows for manufacturing, experiments are conducted to synthesize surface defect images from an imbalanced data set and estimate parameters of a tool wear degradation model from limited observations. Results show that normalizing flows are an effective, multi-purpose DL architecture for solving these problems in manufacturing. Future work should explore normalizing flows for more complex degradation models and develop a framework for likelihood-based anomaly detection. Code is available at https://github.com/uky-aism/flows-for-manufacturing. 
    more » « less
  3. Normalizing flows provide an elegant approach to generative modeling that allows for efficient sampling and exact density evaluation of unknown data distributions. However, current techniques have significant limitations in their expressivity when the data distribution is supported on a lowdimensional manifold or has a non-trivial topology. We introduce a novel statistical framework for learning a mixture of local normalizing flows as “chart maps” over the data manifold. Our framework augments the expressivity of recent approaches while preserving the signature property of normalizing flows, that they admit exact density evaluation. We learn a suitable atlas of charts for the data manifold via a vector quantized autoencoder (VQ-AE) and the distributions over them using a conditional flow. We validate experimentally that our probabilistic framework enables existing approaches to better model data distributions over complex manifolds. 
    more » « less
  4. Bouamor, Houda; Pino, Juan; Bali, Kalia (Ed.)
    This paper presents FlowSUM, a normalizing flows-based variational encoder-decoder framework for Transformer-based summarization. Our approach tackles two primary challenges in variational summarization: insufficient semantic information in latent representations and posterior collapse during training. To address these challenges, we employ normalizing flows to enable flexible latent posterior modeling, and we propose a controlled alternate aggressive training (CAAT) strategy with an improved gate mechanism. Experimental results show that FlowSUM significantly enhances the quality of generated summaries and unleashes the potential for knowledge distillation with minimal impact on inference time. Furthermore, we investigate the issue of posterior collapse in normalizing flows and analyze how the summary quality is affected by the training strategy, gate initialization, and the type and number of normalizing flows used, offering valuable insights for future research. 
    more » « less
  5. Abstract Understanding the ages of stars is crucial for unraveling the formation history and evolution of our Galaxy. Traditional methods for estimating stellar ages from spectroscopic data often struggle with providing appropriate uncertainty estimations and are severely constrained by the parameter space. In this work, we introduce a new approach using normalizing flows—a type of deep generative model—to estimate stellar ages for evolved stars with improved accuracy and robust uncertainty characterization. The model is trained on stellar masses for evolved stars derived from asteroseismology and predicts the relationship between the carbon and nitrogen abundances of a given star and its age. Unlike standard neural network techniques, normalizing flows enable the recovery of full likelihood distributions for individual stellar ages, offering a richer and more informative perspective on uncertainties. Our method yields age estimations for 378,720 evolved stars and achieves a typical absolute age uncertainty of approximately 2 Gyr. By intrinsically accounting for the coverage and density of the training data, our model ensures that the resulting uncertainties reflect both the inherent noise in the data and the completeness of the sampled parameter space. Applying this method to data from the fifth-generation Sloan Digital Sky Survey Milky Way Mapper, we have produced the largest stellar age catalog for evolved stars to date. 
    more » « less