Growth curve models (GCMs), with their ability to directly investigate within-subject change over time and between-subject differences in change for longitudinal data, are widely used in social and behavioral sciences. While GCMs are typically studied with the normal distribution assumption, empirical data often violate the normality assumption in applications. Failure to account for the deviation from normality in data distribution may lead to unreliable model estimation and misleading statistical inferences. A robust GCM based on conditional medians was recently proposed and outperformed traditional growth curve modeling when outliers are present resulting in nonnormality. However, this robust approach was shown to perform less satisfactorily when leverage observations existed. In this work, we propose a robust double medians growth curve modeling approach (DOME GCM) to thoroughly disentangle the influence of data contamination on model estimation and inferences, where two conditional medians are employed for the distributions of the within-subject measurement errors and of random effects, respectively. Model estimation and inferences are conducted in the Bayesian framework, and Laplace distributions are used to convert the optimization problem of median estimation into a problem of obtaining the maximum likelihood estimator for a transformed model. A Monte Carlo simulation study has been conducted to evaluate the numerical performance of the proposed approach, and showed that the proposed approach yields more accurate and efficient parameter estimates when data contain outliers or leverage observations. The application of the developed robust approach is illustrated using a real dataset from the Virginia Cognitive Aging Project to study the change of memory ability.
more »
« less
Assessing the Impact of Precision Parameter Prior in Bayesian Non-parametric Growth Curve Modeling
Bayesian non-parametric (BNP) modeling has been developed and proven to be a powerful tool to analyze messy data with complex structures. Despite the increasing popularity of BNP modeling, it also faces challenges. One challenge is the estimation of the precision parameter in the Dirichlet process mixtures. In this study, we focus on a BNP growth curve model and investigate how non-informative prior, weakly informative prior, accurate informative prior, and inaccurate informative prior affect the model convergence, parameter estimation, and computation time. A simulation study has been conducted. We conclude that the non-informative prior for the precision parameter is less preferred because it yields a much lower convergence rate, and growth curve parameter estimates are not sensitive to informative priors.
more »
« less
- Award ID(s):
- 1951038
- PAR ID:
- 10225042
- Date Published:
- Journal Name:
- Frontiers in Psychology
- Volume:
- 12
- ISSN:
- 1664-1078
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Batch normalization (BN) is a popular and ubiquitous method in deep learning that has been shown to decrease training time and improve generalization performance of neural networks. Despite its success, BN is not theoretically well understood. It is not suitable for use with very small mini-batch sizes or online learning. In this paper, we propose a new method called Batch Normalization Preconditioning (BNP). Instead of applying normalization explicitly through a batch normalization layer as is done in BN, BNP applies normalization by conditioning the parameter gradients directly during training. This is designed to improve the Hessian matrix of the loss function and hence convergence during training. One benefit is that BNP is not constrained on the mini-batch size and works in the online learning setting. Furthermore, its connection to BN provides theoretical insights on how BN improves training and how BN is applied to special architectures such as convolutional neural networks. For a theoretical foundation, we also present a novel Hessian condition number based convergence theory for a locally convex but not strong-convex loss, which is applicable to networks with a scale-invariant property.more » « less
-
Growth curve models have been widely used to analyse longitudinal data in social and behavioural sciences. Although growth curve models with normality assumptions are relatively easy to estimate, practical data are rarely normal. Failing to account for non‐normal data may lead to unreliable model estimation and misleading statistical inference. In this work, we propose a robust approach for growth curve modelling using conditional medians that are less sensitive to outlying observations. Bayesian methods are applied for model estimation and inference. Based on the existing work on Bayesian quantile regression using asymmetric Laplace distributions, we use asymmetric Laplace distributions to convert the problem of estimating a median growth curve model into a problem of obtaining the maximum likelihood estimator for a transformed model. Monte Carlo simulation studies have been conducted to evaluate the numerical performance of the proposed approach with data containing outliers or leverage observations. The results show that the proposed approach yields more accurate and efficient parameter estimates than traditional growth curve modelling. We illustrate the application of our robust approach using conditional medians based on a real data set from the Virginia Cognitive Aging Project.more » « less
-
Abstract Hyporheic exchange in streams is critical to ecosystem functions such as nutrient cycling along river corridors, especially for slowly moving or small stream systems. The transient storage model (TSM) has been widely used for modeling of hyporheic exchange. TSM calibration, for hyporheic exchange, is typically used to estimate four parameters, including the mass exchange rate coefficient, the dispersion coefficient, stream cross‐sectional area, and hyporheic zone cross‐sectional area. Prior studies have raised concerns regarding the non‐uniqueness of the inverse problem for the TSM, that is, the occurrence of different parameter vectors resulting in TSM solution that reproduces the observed in‐stream tracer break through curve (BTC) with the same error. This leads to practical non‐identifiability in determining the unknown parameter vector values even when global‐optimal values exist, and the parameter optimization becomes practically non‐unique. To address this problem, we applied the simulated annealing method to calibrate the TSM to BTCs, because it is less susceptible to local minima‐induced non‐identifiability. A hypothetical (or synthetic) tracer test data set with known parameters was developed to demonstrate the capability of the simulated annealing method to find the global minimum parameter vector, and it identified the “hypothetically‐true” global minimum parameter vector even with input data that were modified with up to 10% noise without increasing the number of iterations required for convergence. The simulated annealing TSM was then calibrated using two in‐stream tracer tests conducted in East Fork Poplar Creek, Tennessee. Simulated annealing was determined to be appropriate for quantifying the TSM parameter vector because of its search capability for the global minimum parameter vector.more » « less
-
Abstract This paper addresses the parameter estimation problem for lithium-ion battery pack models comprising cells in series. This valuable information can be exploited in fault diagnostics to estimate the number of cells that are exhibiting abnormal behaviour, e.g. large resistances or small capacities. In particular, we use a Bayesian approach to estimate the parameters of a two-cell arrangement modelled using equivalent circuits. Although our modeling framework has been extensively reported in the literature, its structural identifiability properties have not been reported yet to the best of the authors’ knowledge. Moreover, most contributions in the literature tackle the estimation problem through point-wise estimates assuming Gaussian noise using e.g. least-squares methods (maximum likelihood estimation) or Kalman filters (maximum a posteriori estimation). In contrast, we apply methods that are suitable for nonlinear and non-Gaussian estimation problems and estimate the full posterior probability distribution of the parameters. We study how the model structure, available measurements and prior knowledge of the model parameters impact the underlying posterior probability distribution that is recovered for the parameters. For two cells in series, a bimodal distribution is obtained whose modes are centered around the real values of the parameters for each cell. Therefore, bounds on the model parameters for a battery pack can be derived.more » « less