skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Bayesian Inference: The Comprehensive Approach to Analyzing Single-Molecule Experiments
Biophysics experiments performed at single-molecule resolution provide exceptional insight into the structural details and dynamic behavior of biological systems. However, extracting this information from the corresponding experimental data unequivocally requires applying a biophysical model. In this review, we discuss how to use probability theory to apply these models to single-molecule data. Many current single-molecule data analysis methods apply parts of probability theory, sometimes unknowingly, and thus miss out on the full set of benefits provided by this self-consistent framework. The full application of probability theory involves a process called Bayesian inference that fully accounts for the uncertainties inherent to single-molecule experiments. Additionally, using Bayesian inference provides a scientifically rigorous method of incorporating information from multiple experiments into a single analysis and finding the best biophysical model for an experiment without the risk of overfitting the data. These benefits make the Bayesian approach ideal for analyzing any type of single-molecule experiment.  more » « less
Award ID(s):
2004016
PAR ID:
10465421
Author(s) / Creator(s):
; ;
Date Published:
Journal Name:
Annual Review of Biophysics
Volume:
50
Issue:
1
ISSN:
1936-122X
Page Range / eLocation ID:
191 to 208
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Developing a mechanistic understanding of protein dynamics and conformational changes at polymer interfaces is critical for a range of processes including industrial protein separations. Salting out is one example of a procedure that is ubiquitous in protein separations yet is optimized empirically because there is no mechanistic description of the underlying interactions that would allow predictive modeling. Here, we investigate peak narrowing in a model transferrin–nylon system under salting out conditions using a combination of single-molecule tracking and ensemble separations. Distinct surface transport modes and protein conformational changes at the negatively charged nylon interface are quantified as a function of salt concentration. Single-molecule kinetics relate macroscale improvements in chromatographic peak broadening with microscale distributions of surface interaction mechanisms such as continuous-time random walks and simple adsorption–desorption. Monte Carlo simulations underpinned by the stochastic theory of chromatography are performed using kinetic data extracted from single-molecule observations. Simulations agree with experiment, revealing a decrease in peak broadening as the salt concentration increases. The results suggest that chemical modifications to membranes that decrease the probability of surface random walks could reduce peak broadening in full-scale protein separations. More broadly, this work represents a proof of concept for combining single-molecule experiments and a mechanistic theory to improve costly and time-consuming empirical methods of optimization. 
    more » « less
  2. McLellan, Myles F (Ed.)
    In most U.S. jurisdictions, prosecutors are not required to clearly establish a reasonable basis for guilt prior to offering defendants plea deals. We apply Bayesian analyses, which are uniquely suited to illuminate the impact of prior probability of guilt on the informativeness of a particular outcome (i.e., a guilty plea), to demonstrate the risks of plea offers that precede evidence. Our primary prediction was that lower prior probabilities of guilt would coincide with a significantly higher risk for false guilty pleas. We incorporated data from Wilford, Sutherland into a Bayesian analysis allowing us to model the expected diagnosticity of plea acceptance across the full range of prior probability of guilt. Our analysis indicated that, as predicted, when plea offers are accepted at lower prior probabilities of guilt, the probability that a plea is actually false is significantly higher than when prior probabilities of guilt are higher. In other words, there is a trade-off between prior probability of guilt and information gain. For instance, in our analysis, when prior probability of guilt was 50%, posterior probability of guilt (after a plea) was 77.8%; when prior probability of guilt was 80%, posterior probability of guilt was 93.3%. Our results clearly indicate the importance of ensuring that there is a reasonable basis for guilt before a plea deal is extended. In the absence of shared discovery, no such reasonable basis can be established. Further, these results illustrate the additional insights gained from applying a Bayesian approach to plea-decision contexts. 
    more » « less
  3. X-ray radiography is a technique frequently used to diagnose convergent high-energy-density (HED) systems, such as inertial confinement fusion implosion, and provides unique information that is not available through self-emission measurements. We investigate the scope and limits of that information using a radiography simulation combined with the Bayesian inference workflow. The accuracy of density reconstruction from simulated radiographs of spherical implosions driven with 27 kJ laser energy is assessed, including the increase or decrease in accuracy due to the addition of Lagrangian marker layers, Poisson noise, and improved prior information. This work is the first to present the full uncertainty distributions inferred from radiography analysis in HED systems and demonstrates the importance of constructing the full posterior probability density, as opposed to a point estimate, due to the modal structure of the likelihood surface introduced by typical experimental noise sources. This general methodology can be used both for robust analysis of radiographic data and for an improved design of radiography experiments by modeling the full experimental system. 
    more » « less
  4. Abstract Causal reasoning is a fundamental cognitive ability that enables individuals to learn about the complex interactions in the world around them. However, the mechanisms that underpin causal reasoning are not well understood. For example, it remains unresolved whether children's causal inferences are best explained by Bayesian inference or associative learning. The two experiments and computational models reported here were designed to examine whether 5‐ and 6‐year‐olds will retrospectively reevaluate objects—that is, adjust their beliefs about the causal status of some objects presented at an earlier point in time based on the observed causal status of other objects presented at a later point in time—when asked to reason about 3 and 4 objects and under varying degrees of information processing demands. Additionally, the experiments and models were designed to determine whether children's retrospective reevaluations were best explained by associative learning, Bayesian inference, or some combination of both. The results indicated that participants retrospectively reevaluated causal inferences under minimal information‐processing demands (Experiment 1) but failed to do so under greater information processing demands (Experiment 2) and that their performance was better captured by an associative learning mechanism, with less support for descriptions that rely on Bayesian inference. Research HighlightsFive‐ and 6‐year‐old children engage in retrospective reevaluation under minimal information‐processing demands (Experiment 1).Five‐ and 6‐year‐old children do not engage in retrospective reevaluation under more extensive information‐processing demands (Experiment 2).Across both experiments, children's retrospective reevaluations were better explained by a simple associative learning model, with only minimal support for a simple Bayesian model.These data contribute to our understanding of the cognitive mechanisms by which children make causal judgements. 
    more » « less
  5. Abstract Multivariate spatially oriented data sets are prevalent in the environmental and physical sciences. Scientists seek to jointly model multiple variables, each indexed by a spatial location, to capture any underlying spatial association for each variable and associations among the different dependent variables. Multivariate latent spatial process models have proved effective in driving statistical inference and rendering better predictive inference at arbitrary locations for the spatial process. High‐dimensional multivariate spatial data, which are the theme of this article, refer to data sets where the number of spatial locations and the number of spatially dependent variables is very large. The field has witnessed substantial developments in scalable models for univariate spatial processes, but such methods for multivariate spatial processes, especially when the number of outcomes are moderately large, are limited in comparison. Here, we extend scalable modeling strategies for a single process to multivariate processes. We pursue Bayesian inference, which is attractive for full uncertainty quantification of the latent spatial process. Our approach exploits distribution theory for the matrix‐normal distribution, which we use to construct scalable versions of a hierarchical linear model of coregionalization (LMC) and spatial factor models that deliver inference over a high‐dimensional parameter space including the latent spatial process. We illustrate the computational and inferential benefits of our algorithms over competing methods using simulation studies and an analysis of a massive vegetation index data set. 
    more » « less