Biological systems contain a large number of molecules that have diverse interactions. A fruitful path to understanding these systems is to represent them with interaction networks, and then describe flow processes in the network with a dynamic model. Boolean modeling, the simplest discrete dynamic modeling framework for biological networks, has proven its value in recapitulating experimental results and making predictions. A first step and major roadblock to the widespread use of Boolean networks in biology is the laborious network inference and construction process. Here we present a streamlined network inference method that combines the discovery of a parsimonious network structure and the identification of Boolean functions that determine the dynamics of the system. This inference method is based on a causal logic analysis method that associates a logic type (sufficient or necessary) to node-pair relationships (whether promoting or inhibitory). We use the causal logic framework to assimilate indirect information obtained from perturbation experiments and infer relationships that have not yet been documented experimentally. We apply this inference method to a well-studied process of hormone signaling in plants, the signaling underlying abscisic acid (ABA)—induced stomatal closure. Applying the causal logic inference method significantly reduces the manual work typically required for network and Boolean model construction. The inferred model agrees with the manually curated model. We also test this method by re-inferring a network representing epithelial to mesenchymal transition based on a subset of the information that was initially used to construct the model. We find that the inference method performs well for various likely scenarios of inference input information. We conclude that our method is an effective approach toward inference of biological networks and can become an efficient step in the iterative process between experiments and computations.
more »
« less
Toward an improved understanding of causation in the ecological sciences
Abstract Society increasingly demands accurate predictions of complex ecosystem processes under novel conditions to address environmental challenges. However, obtaining the process‐level knowledge required to do so does not necessarily align with the burgeoning use in ecology of correlative model selection criteria, such as Akaike information criterion. These criteria select models based on their ability to reproduce outcomes, not on their ability to accurately represent causal effects. Causal understanding does not require matching outcomes, but rather involves identifying model forms and parameter values that accurately describe processes. We contend that researchers can arrive at incorrect conclusions about cause‐and‐effect relationships by relying on information criteria. We illustrate via a specific example that inference extending beyond prediction into causality can be seriously misled by information‐theoretic evidence. Finally, we identify a solution space to bridge the gap between the correlative inference provided by model selection criteria and a process‐based understanding of ecological systems.
more »
« less
- Award ID(s):
- 1426891
- PAR ID:
- 10372383
- Publisher / Repository:
- Wiley Blackwell (John Wiley & Sons)
- Date Published:
- Journal Name:
- Frontiers in Ecology and the Environment
- Volume:
- 20
- Issue:
- 8
- ISSN:
- 1540-9295
- Format(s):
- Medium: X Size: p. 474-480
- Size(s):
- p. 474-480
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Abstract Causal reasoning is a fundamental cognitive ability that enables individuals to learn about the complex interactions in the world around them. However, the mechanisms that underpin causal reasoning are not well understood. For example, it remains unresolved whether children's causal inferences are best explained by Bayesian inference or associative learning. The two experiments and computational models reported here were designed to examine whether 5‐ and 6‐year‐olds will retrospectively reevaluate objects—that is, adjust their beliefs about the causal status of some objects presented at an earlier point in time based on the observed causal status of other objects presented at a later point in time—when asked to reason about 3 and 4 objects and under varying degrees of information processing demands. Additionally, the experiments and models were designed to determine whether children's retrospective reevaluations were best explained by associative learning, Bayesian inference, or some combination of both. The results indicated that participants retrospectively reevaluated causal inferences under minimal information‐processing demands (Experiment 1) but failed to do so under greater information processing demands (Experiment 2) and that their performance was better captured by an associative learning mechanism, with less support for descriptions that rely on Bayesian inference. Research HighlightsFive‐ and 6‐year‐old children engage in retrospective reevaluation under minimal information‐processing demands (Experiment 1).Five‐ and 6‐year‐old children do not engage in retrospective reevaluation under more extensive information‐processing demands (Experiment 2).Across both experiments, children's retrospective reevaluations were better explained by a simple associative learning model, with only minimal support for a simple Bayesian model.These data contribute to our understanding of the cognitive mechanisms by which children make causal judgements.more » « less
-
Abstract Species distribution models (SDMs) have become increasingly popular for making ecological inferences, as well as predictions to inform conservation and management. In predictive modeling, practitioners often use correlative SDMs that only evaluate a single spatial scale and do not account for differences in life stages. These modeling decisions may limit the performance of SDMs beyond the study region or sampling period. Given the increasing desire to develop transferable SDMs, a robust framework is necessary that can account for known challenges of model transferability. Here, we propose a comparative framework to develop transferable SDMs, which was tested using satellite telemetry data from green turtles (Chelonia mydas). This framework is characterized by a set of steps comparing among different models based on (1) model algorithm (e.g., generalized linear model vs. Gaussian process regression) and formulation (e.g., correlative model vs. hybrid model), (2) spatial scale, and (3) accounting for life stage. SDMs were fitted as resource selection functions and trained on data from the Gulf of Mexico with bathymetric depth, net primary productivity, and sea surface temperature as covariates. Independent validation datasets from Brazil and Qatar were used to assess model transferability. A correlative SDM using a hierarchical Gaussian process regression (HGPR) algorithm exhibited greater transferability than a hybrid SDM using HGPR, as well as correlative and hybrid forms of hierarchical generalized linear models. Additionally, models that evaluated habitat selection at the finest spatial scale and that did not account for life stage proved to be the most transferable in this study. The comparative framework presented here may be applied to a variety of species, ecological datasets (e.g., presence‐only, presence‐absence, mark‐recapture), and modeling frameworks (e.g., resource selection functions, step selection functions, occupancy models) to generate transferable predictions of species–habitat associations. We expect that SDM predictions resulting from this comparative framework will be more informative management tools and may be used to more accurately assess climate change impacts on a wide array of taxa.more » « less
-
Abstract Causal inference practitioners have increasingly adopted machine learning techniques with the aim of producing principled uncertainty quantification for causal effects while minimizing the risk of model misspecification. Bayesian nonparametric approaches have attracted attention as well, both for their flexibility and their promise of providing natural uncertainty quantification. Priors on high‐dimensional or nonparametric spaces, however, can often unintentionally encode prior information that is at odds with substantive knowledge in causal inference—specifically, the regularization required for high‐dimensional Bayesian models to work can indirectly imply that the magnitude of the confounding is negligible. In this paper, we explain this problem and provide tools for (i) verifying that the prior distribution does not encode an inductive bias away from confounded models and (ii) verifying that the posterior distribution contains sufficient information to overcome this issue if it exists. We provide a proof‐of‐concept on simulated data from a high‐dimensional probit‐ridge regression model, and illustrate on a Bayesian nonparametric decision tree ensemble applied to a large medical expenditure survey.more » « less
-
Summary Point processes are probabilistic tools for modelling event data. While there exists a fast-growing literature on the relationships between point processes, how such relationships connect to causal effects remains unexplored. In the presence of unmeasured confounders, parameters from point process models do not necessarily have causal interpretations. We propose an instrumental variable method for causal inference with point process treatment and outcome. We define causal quantities based on potential outcomes and establish nonparametric identification results with a binary instrumental variable. We extend the traditional Wald estimation to deal with point process treatment and outcome, showing that it should be performed after a Fourier transform of the intention-to-treat effects on the treatment and outcome, and thus takes the form of deconvolution. We refer to this approach as generalized Wald estimation and propose an estimation strategy based on well-established deconvolution methods.more » « less
An official website of the United States government
