Statistical prediction plays an important role in many decision processes, such as university budgeting (depending on the number of students who will enroll), capital budgeting (depending on the remaining lifetime of a fleet of systems), the needed amount of cash reserves for warranty expenses (depending on the number of warranty returns), and whether a product recall is needed (depending on the number of potentially life-threatening product failures). In statistical inference, likelihood ratios have a long history of use for decision making relating to model parameters (e.g., in evidence-based medicine and forensics). We propose a general prediction method, based on a likelihood ratio (LR) involving both the data and a future random variable. This general approach provides a way to identify prediction interval methods that have excellent statistical properties. For example, if a prediction method can be based on a pivotal quantity, our LR-based method will often identify it. For applications where a pivotal quantity does not exist, the LR-based method provides a procedure with good coverage properties for both continuous or discrete-data prediction applications.
more » « less- Award ID(s):
- 2015390
- PAR ID:
- 10470879
- Publisher / Repository:
- Institute for Operations Research and the Management Sciences (INFORMS)
- Date Published:
- Journal Name:
- INFORMS Journal on Data Science
- Volume:
- 1
- Issue:
- 1
- ISSN:
- 2694-4022
- Page Range / eLocation ID:
- 63 to 80
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
Abstract Previous work with simulations of oceanographic high-frequency (HF) radars has identified possible improvements when using maximum likelihood estimation (MLE) for direction of arrival; however, methods for determining the number of emitters (here defined as spatially distinct patches of the ocean surface) have not realized these improvements. Here we describe and evaluate the use of the likelihood ratio (LR) for emitter detection, demonstrating its application to oceanographic HF radar data. The combined detection–estimation methods MLE-LR are compared with multiple signal classification method (MUSIC) and MUSIC parameters for SeaSonde HF radars, along with a method developed for 8-channel systems known as MUSIC-Highest. Results show that the use of MLE-LR produces similar accuracy, in terms of the RMS difference and correlation coefficients squared, as previous methods. We demonstrate that improved accuracy can be obtained for both methods, at the cost of fewer velocity observations and decreased spatial coverage. For SeaSondes, accuracy improvements are obtained with less commonly used parameter sets. The MLE-LR is shown to be able to resolve simultaneous closely spaced emitters, which has the potential to improve observations obtained by HF radars operating in complex current environments. Significance Statement We identify and test a method based on the likelihood ratio (LR) for determining the number of signal sources in observations subject to direction finding with maximum likelihood estimation (MLE). Direction-finding methods are used in broad-ranging applications that include radar, sonar, and wireless communication. Previous work suggests accuracy improvements when using MLE, but suitable methods for determining the number of simultaneous signal sources are not well known. Our work shows that the LR, when combined with MLE, performs at least as well as alternative methods when applied to oceanographic high-frequency (HF) radars. In some situations, MLE and LR obtain superior resolution, where resolution is defined as the ability to distinguish closely spaced signal sources.more » « less
-
Metamodeling has been a topic of longstanding interest in stochastic simulation because of the usefulness of metamodels for optimization, sensitivity, and real- or near-real-time decision making. Experiment design is the foundation of classical metamodeling: an effective experiment design uncovers the spatial relationships among the design/decision variables and the simulation response; therefore, more design points, providing better coverage of space, is almost always better. However, metamodeling based on likelihood ratios (LRs) turns the design question on its head: each design point provides an unbiased prediction of the response at any other location in space, but perhaps with such inflated variance as to be counterproductive. Thus, the question becomes more which design points to employ for prediction and less where to place them. In this paper we take the first comprehensive look at LR metamodeling, categorizing both the various types of LR metamodels and the contexts in which they might be employed.more » « less
-
Efficient disassembly operation is considered a promising approach toward waste reduction and End-of-Use (EOU) product recovery. However, many kinds of uncertainty exist during the product lifecycle which make disassembly decision a complicated process. The optimum disassembly sequence may vary at different milestones depending on the purpose of disassembly (repair, maintenance, reuse and recovery), product quality conditions and external factors such as consumer preference, and the market value of EOU components. A disassembly sequence which is optimum for one purpose may not be optimum in future life cycles or other purposes. Therefore, there is a need for incorporating the requirements of the entire product life-cycle when obtaining the optimum disassembly sequence. This paper applies a fuzzy method to quantify the probability that each feasible disassembly transition will be needed during the entire product lifecycle. Further, the probability values have been used in an optimization model to find the disassembly sequence with maximum likelihood. An example of vacuum cleaner is used to show how the proposed method can be applied to quantify different users’ evaluation on the relative importance of disassembly selection criteria as well as the probability of each disassembly operation.more » « less
-
This work develops a novel set of algorithms, al- ternating Gradient Descent (GD) and minimization for MRI (altGDmin-MRI1 and altGDmin-MRI2), for accelerated dynamic MRI by assuming an approximate low-rank (LR) model on the matrix formed by the vectorized images of the sequence. The LR model itself is well-known in the MRI literature; our contribution is the novel GD-based algorithms which are much faster, memory-efficient, and ‘general’ compared with existing work; and careful use of a 3-level hierarchical LR model. By ‘general’, we mean that, with a single choice of parameters, our method provides accurate reconstructions for multiple acceler- ated dynamic MRI applications, multiple sampling rates and sampling schemes. We show that our methods outperform many of the popular existing approaches while also being faster than all of them, on average. This claim is based on comparisons on 8 different retrospectively undersampled multi-coil dynamic MRI applications, sampled using either 1D Cartesian or 2D pseudo- radial undersampling, at multiple sampling rates. Evaluations on some prospectively undersampled datasets are also provided. Our second contribution is a mini-batch subspace tracking extension that can process new measurements and return reconstructions within a short delay after they arrive. The recovery algorithm itself is also faster than its batch counterpart.more » « less
-
Machine learning (ML) methods, such as artificial neural networks (ANN), k-nearest neighbors (kNN), random forests (RF), support vector machines (SVM), and boosted decision trees (DTs), may offer stronger predictive performance than more traditional, parametric methods, such as linear regression, multiple linear regression, and logistic regression (LR), for specific mapping and modeling tasks. However, this increased performance is often accompanied by increased model complexity and decreased interpretability, resulting in critiques of their “black box” nature, which highlights the need for algorithms that can offer both strong predictive performance and interpretability. This is especially true when the global model and predictions for specific data points need to be explainable in order for the model to be of use. Explainable boosting machines (EBM), an augmentation and refinement of generalize additive models (GAMs), has been proposed as an empirical modeling method that offers both interpretable results and strong predictive performance. The trained model can be graphically summarized as a set of functions relating each predictor variable to the dependent variable along with heat maps representing interactions between selected pairs of predictor variables. In this study, we assess EBMs for predicting the likelihood or probability of slope failure occurrence based on digital terrain characteristics in four separate Major Land Resource Areas (MLRAs) in the state of West Virginia, USA and compare the results to those obtained with LR, kNN, RF, and SVM. EBM provided predictive accuracies comparable to RF and SVM and better than LR and kNN. The generated functions and visualizations for each predictor variable and included interactions between pairs of predictor variables, estimation of variable importance based on average mean absolute scores, and provided scores for each predictor variable for new predictions add interpretability, but additional work is needed to quantify how these outputs may be impacted by variable correlation, inclusion of interaction terms, and large feature spaces. Further exploration of EBM is merited for geohazard mapping and modeling in particular and spatial predictive mapping and modeling in general, especially when the value or use of the resulting predictions would be greatly enhanced by improved interpretability globally and availability of prediction explanations at each cell or aggregating unit within the mapped or modeled extent.more » « less