skip to main content


Title: Conditional Inferential Models: Combining Information for Prior-Free Probabilistic Inference
Summary

The inferential model (IM) framework provides valid prior-free probabilistic inference by focusing on predicting unobserved auxiliary variables. But, efficient IM-based inference can be challenging when the auxiliary variable is of higher dimension than the parameter. Here we show that features of the auxiliary variable are often fully observed and, in such cases, a simultaneous dimension reduction and information aggregation can be achieved by conditioning. This proposed conditioning strategy leads to efficient IM inference and casts new light on Fisher's notions of sufficiency, conditioning and also Bayesian inference. A differential-equation-driven selection of a conditional association is developed, and validity of the conditional IM is proved under some conditions. For problems that do not admit a conditional IM of the standard form, we propose a more flexible class of conditional IMs based on localization. Examples of local conditional IMs in a bivariate normal model and a normal variance components model are also given.

 
more » « less
NSF-PAR ID:
10397809
Author(s) / Creator(s):
;
Publisher / Repository:
Oxford University Press
Date Published:
Journal Name:
Journal of the Royal Statistical Society Series B: Statistical Methodology
Volume:
77
Issue:
1
ISSN:
1369-7412
Page Range / eLocation ID:
p. 195-217
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Summary

    Ground motion selection is a crucial step in probabilistic seismic performance assessment of structural systems. Particularly, identifying ground motion records compatible with a specific hazard level has been the major focus of past studies. We propose a multivariate return period (MRP)‐based record selection methodology to further improve the multivariate hazard consistency over a vector of intensity measures (IMs). Unlike the traditional univariate return period anchored on a scalar IM, MRP generalizes the return period concept by accommodating the joint rate of exceedance of a vector of IMs, thereby providing more holistic seismic hazard characterization. By leveraging MRP in linking the seismic hazard to a vector IMs, the proposed MRP‐based ground motion selection methodology for the first time offers a mathematically rigorous yet practical solution for multivariate hazard consistency in ground motion selection. The merit of the MRP‐based ground motion selection is demonstrated by comparing the resulting target spectra and seismic demand estimates for several case‐study structures with other state‐of‐the‐art ground motion selection alternatives. From the results, the MRP‐based ground motion selection employing Kendall's distribution function turns out to be a promising alternative, offering favorable new features including (a) moderate target spectra intensity and moderately low target spectra standard deviation; (b) superior convergence with the increase of the conditioning IM dimension and ability to approximate higher‐dimensional multivariate hazard consistency with lower dimensional conditioning IMs; and (c) capability to realistically capture the multimodal spectral shape owing to the incorporation of multivariate Gaussian mixture distribution in generating target spectra.

     
    more » « less
  2. Summary

    Intensity measure (IM) selection is a critical step in probabilistic seismic risk assessment (PSRA). In many past studies, the efficiency of an IM, which quantifies its explanatory power within a probabilistic seismic demand model, has been a predominant IM evaluation criterion. However, because PSRA requires convolution of the conditional demand model and the probabilistic seismic hazard, IM selection solely based on efficiency ignores the influence of uncertainty contribution particularly from the ground motion prediction equations (GMPEs), which is also dependent on the IMs, and may lead to biased IM selection. In the present study, the concept of joint entropy from information theory is introduced to quantify the uncertainty of unconditional multivariate seismic demands for general multiresponse structural systems, offering a holistic consideration of different sources of uncertainties. We propose a novel entropy‐based IM evaluation criterion that can serve as the basis for IM selection and develop a practical framework to facilitate the implementation of the entropy‐based IM selection in site‐specific PSRA. Based on two case‐study highway bridges, the merit of the proposed IM selection methodology is demonstrated, and the influence of IM selection on demand entropy and loss estimation in PSRA is evaluated. From the results, the proposed entropy‐based IM selection manifests improved capability in considering different sources of uncertainties and is found to deliver much more consistent and confident seismic risk estimates.

     
    more » « less
  3. Abstract

    Imputation is a popular technique for handling item nonresponse. Parametric imputation is based on a parametric model for imputation and is not robust against the failure of the imputation model. Nonparametric imputation is fully robust but is not applicable when the dimension of covariates is large due to the curse of dimensionality. Semiparametric imputation is another robust imputation based on a flexible model where the number of model parameters can increase with the sample size. In this paper, we propose a new semiparametric imputation based on a more flexible model assumption than the Gaussian mixture model. In the proposed mixture model, we assume a conditional Gaussian model for the study variable given the auxiliary variables, but the marginal distribution of the auxiliary variables is not necessarily Gaussian. The proposed mixture model is more flexible and achieves a better approximation than the Gaussian mixture models. The proposed method is applicable to high‐dimensional covariate problem by including a penalty function in the conditional log‐likelihood function. The proposed method is applied to the 2017 Korean Household Income and Expenditure Survey conducted by Statistics Korea.

     
    more » « less
  4. An influential paper by Kleibergen (2005, Econometrica 73, 1103–1123) introduces Lagrange multiplier (LM) and conditional likelihood ratio-like (CLR) tests for nonlinear moment condition models. These procedures aim to have good size performance even when the parameters are unidentified or poorly identified. However, the asymptotic size and similarity (in a uniform sense) of these procedures have not been determined in the literature. This paper does so. This paper shows that the LM test has correct asymptotic size and is asymptotically similar for a suitably chosen parameter space of null distributions. It shows that the CLR tests also have these properties when the dimension p of the unknown parameter θ equals 1. When p ≥ 2, however, the asymptotic size properties are found to depend on how the conditioning statistic, upon which the CLR tests depend, is weighted. Two weighting methods have been suggested in the literature. The paper shows that the CLR tests are guaranteed to have correct asymptotic size when p ≥ 2 when the weighting is based on an estimator of the variance of the sample moments, i.e., moment-variance weighting, combined with the Robin and Smith (2000, Econometric Theory 16, 151–175) rank statistic. The paper also determines a formula for the asymptotic size of the CLR test when the weighting is based on an estimator of the variance of the sample Jacobian. However, the results of the paper do not guarantee correct asymptotic size when p ≥ 2 with the Jacobian-variance weighting, combined with the Robin and Smith (2000, Econometric Theory 16, 151–175) rank statistic, because two key sample quantities are not necessarily asymptotically independent under some identification scenarios. Analogous results for confidence sets are provided. Even for the special case of a linear instrumental variable regression model with two or more right-hand side endogenous variables, the results of the paper are new to the literature. 
    more » « less
  5. Motivated by modern applications in which one constructs graphical models based on a very large number of features, this paper introduces a new class of cluster-based graphical models, in which variable clustering is applied as an initial step for reducing the dimension of the feature space. We employ model assisted clustering, in which the clusters contain features that are similar to the same unobserved latent variable. Two different cluster-based Gaussian graphical models are considered: the latent variable graph, corresponding to the graphical model associated with the unobserved latent variables, and the cluster-average graph, corresponding to the vector of features averaged over clusters. Our study reveals that likelihood based inference for the latent graph, not analyzed previously, is analytically intractable. Our main contribution is the development and analysis of alternative estimation and inference strategies, for the precision matrix of an unobservable latent vector Z. We replace the likelihood of the data by an appropriate class of empirical risk functions, that can be specialized to the latent graphical model and to the simpler, but under-analyzed, cluster-average graphical model. The estimators thus derived can be used for inference on the graph structure, for instance on edge strength or pattern recovery. Inference is based on the asymptotic limits of the entry-wise estimates of the precision matrices associated with the conditional independence graphs under consideration. While taking the uncertainty induced by the clustering step into account, we establish Berry-Esseen central limit theorems for the proposed estimators. It is noteworthy that, although the clusters are estimated adaptively from the data, the central limit theorems regarding the entries of the estimated graphs are proved under the same conditions one would use if the clusters were known in advance. As an illustration of the usage of these newly developed inferential tools, we show that they can be reliably used for recovery of the sparsity pattern of the graphs we study, under FDR control, which is verified via simulation studies and an fMRI data analysis. These experimental results confirm the theoretically established difference between the two graph structures. Furthermore, the data analysis suggests that the latent variable graph, corresponding to the unobserved cluster centers, can help provide more insight into the understanding of the brain connectivity networks relative to the simpler, average-based, graph. 
    more » « less