skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Fast Fourier Transform Reductions for Bayesian Network Inference
Bayesian Networks are useful for analyzing the properties of systems with large populations of interacting agents (e.g., in social modeling applications and distributed service applications). These networks typically have large functions (CPTs), making exact inference intractable. However, often these models have additive symmetry. In this paper we show how summation-based CPTs, especially in the presence of symmetry, can be computed efficiently through the usage of the Fast Fourier Transform (FFT). In particular, we propose an efficient method using the FFT for reducing the size of Conditional Probability Tables (CPTs) in Bayesian Networks with summation-based causal independence (CI). We show how to apply it directly towards the acceleration of Bucket Elimination, and we subsequently provide experimental results demonstrating the computational speedup provided by our method.  more » « less
Award ID(s):
2008516
PAR ID:
10376093
Author(s) / Creator(s):
; ;
Editor(s):
Camps-Valls, Gustau; Ruiz, Francisco J.; Valera, Isabel
Date Published:
Journal Name:
AI & Statistics, PMLR
Volume:
151
Page Range / eLocation ID:
6445-6458
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. The dominant use of Convolutional Neural Networks (CNNs) in several image and video analysis tasks necessitates a careful re-evaluation of the underlying software libraries for computing them for large-scale image and video databases. We focus our attention on developing methods that can be applied to large image databases or videos of large image sizes. We develop a method that maximizes throughput through the use of vector-based memory I/O and optimized 2D FFT libraries that run on all available physical cores. We also show how to decompose arbitrarily large images into smaller, optimal blocks that can be effectively processed through the use of overlap-and- add. Our approach outperforms Tensorflow for 5x5 kernels and significantly outperforms Tensorflow for 11x11 kernels. 
    more » « less
  2. Dasgupta, Sanjoy; Mandt, Stephan; Li, Yingzhen (Ed.)
    Spatial evolutionary games are used to model large systems of interacting agents. In earlier work, a method was developed using Bayesian Networks to approximate the population dynamics in these games. One advantage of that approach is that one can smoothly adjust the size of the network to get more accurate approximations. However, scaling the method up can be intractable if the number of strategies in the evolutionary game increases. In this paper, we propose a new method for computing more accurate approximations by using surrogate Bayesian Networks. Instead of doing inference on larger networks directly, we do it on a much smaller surrogate network extended with parameters that exploit the symmetry inherent to the domain. We learn the parameters on the surrogate network using KL-divergence as the loss function. We illustrate the value of this method empirically through a comparison on several evolutionary games. 
    more » « less
  3. Dasgupta, Sanjoy; Mandt, Stephan; Li, Yingzhen (Ed.)
    Spatial evolutionary games are used to model large systems of interacting agents. In earlier work, a method was developed using Bayesian Networks to approximate the population dynamics in these games. One advantage of that approach is that one can smoothly adjust the size of the network to get more accurate approximations. However, scaling the method up can be intractable if the number of strategies in the evolutionary game increases. In this paper, we propose a new method for computing more accurate approximations by using surrogate Bayesian Networks. Instead of doing inference on larger networks directly, we do it on a much smaller surrogate network extended with parameters that exploit the symmetry inherent to the domain. We learn the parameters on the surrogate network using KL-divergence as the loss function. We illustrate the value of this method empirically through a comparison on several evolutionary games. 
    more » « less
  4. ABSTRACT In 2012, a regional risk assessment was published that applied Bayesian networks (BN) to the structure of the relative risk model. The original structure of the relative risk model (RRM) was published in the late 1990s and developed during the next decade. The RRM coupled with a Monte Carlo analysis was applied to calculating risk to a number of sites and a variety of questions. The sites included watersheds, terrestrial systems, and marine environments and included stressors such as nonindigenous species, effluents, pesticides, nutrients, and management options. However, it became apparent that there were limits to the original approach. In 2009, the relative risk model was transitioned into the structure of a BN. Bayesian networks had several clear advantages. First, BNs innately incorporated categories and, as in the case of the relative risk model, ranks to describe systems. Second, interactions between multiple stressors can be combined using several pathways and the conditional probability tables (CPT) to calculate outcomes. Entropy analysis was the method used to document model sensitivity. As with the RRM, the method has now been applied to a wide series of sites and questions, from forestry management, to invasive species, to disease, the interaction of ecological and human health endpoints, the flows of large rivers, and now the efficacy and risks of synthetic biology. The application of both methods have pointed to the incompleteness of the fields of environmental chemistry, toxicology, and risk assessment. The low frequency of exposure‐response experiments and proper analysis have limited the available outputs for building appropriate CPTs. Interactions between multiple chemicals, landscape characteristics, population dynamics and community structure have been poorly characterized even for critical environments. A better strategy might have been to first look at the requirements of modern risk assessment approaches and then set research priorities.Integr Environ Assess Manag2021;17:79–94. © 2020 SETAC 
    more » « less
  5. We propose a stochastic variational inference algorithm for training large-scale Bayesian networks, where noisy-OR conditional distributions are used to capture higher-order relationships. One application is to the learning of hierarchical topic models for text data. While previous work has focused on two-layer networks popular in applications like medical diagnosis, we develop scalable algorithms for deep networks that capture a multi-level hierarchy of interactions. Our key innovation is a family of constrained variational bounds that only explicitly optimize posterior probabilities for the sub-graph of topics most related to the sparse observations in a given document. These constrained bounds have comparable accuracy but dramatically reduced computational cost. Using stochastic gradient updates based on our variational bounds, we learn noisy-OR Bayesian networks orders of magnitude faster than was possible with prior Monte Carlo learning algorithms, and provide a new tool for understanding large-scale binary data. 
    more » « less