Road injuries are rated among the top 10 causes of death by the World Health Organization, and the only one that is not a disease. The total economic cost of motor vehicle crashes in the United States was estimated to be $242 billion a year. This study examines multiple factors of accidents simultaneously with a goal of generating an interpretable model that can predict the occurrence of an accident given road conditions and driver behavior. The study compared 4 machine learning and deep learning modeling techniques on a dataset of 7707 trips collected by the Second Strategic Highway Research Program. A gradient boosted model was found to be most accurate and interpretable in accident prediction. This modeling technique also allows us to rank the feature importance of the factors in the model. The study finds that driver behavior, pre-incident maneuvers and secondary task duration are the most important variables in the predictive model. Using these conclusions will allow us to perform more work to infer these accident causes directly from vehicle sensor data in the future.
more »
« less
Designing More Informative Multiple-Driver Experiments
For decades, multiple-driver/stressor research has examined interactions among drivers that will undergo large changes in the future: temperature, pH, nutrients, oxygen, pathogens, and more. However, the most commonly used experimental designs—present-versus-future and ANOVA—fail to contribute to general understanding or predictive power. Linking experimental design to process-based mathematical models would help us predict how ecosystems will behave in novel environmental conditions. We review a range of experimental designs and assess the best experimental path toward a predictive ecology. Full factorial response surface, fractional factorial, quadratic response surface, custom, space-filling, and especially optimal and sequential/adaptive designs can help us achieve more valuable scientific goals. Experiments using these designs are challenging to perform with long-lived organisms or at the community and ecosystem levels. But they remain our most promising path toward linking experiments and theory in multiple-driver research and making accurate, useful predictions.
more »
« less
- Award ID(s):
- 1840868
- PAR ID:
- 10534770
- Publisher / Repository:
- Annual Reviews
- Date Published:
- Journal Name:
- Annual Review of Marine Science
- Volume:
- 16
- Issue:
- 1
- ISSN:
- 1941-1405
- Page Range / eLocation ID:
- 513 to 536
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Abstract Understanding treatment effect heterogeneity has become an increasingly popular task in various fields, as it helps design personalized advertisements in e-commerce or targeted treatment in biomedical studies. However, most of the existing work in this research area focused on either analysing observational data based on strong causal assumptions or conducting post hoc analyses of randomized controlled trial data, and there has been limited effort dedicated to the design of randomized experiments specifically for uncovering treatment effect heterogeneity. In the manuscript, we develop a framework for designing and analysing response adaptive experiments toward better learning treatment effect heterogeneity. Concretely, we provide response adaptive experimental design frameworks that sequentially revise the data collection mechanism according to the accrued evidence during the experiment. Such design strategies allow for the identification of subgroups with the largest treatment effects with enhanced statistical efficiency. The proposed frameworks not only unify adaptive enrichment designs and response-adaptive randomization designs but also complement A/B test designs in e-commerce and randomized trial designs in clinical settings. We demonstrate the merit of our design with theoretical justifications and in simulation studies with synthetic e-commerce and clinical trial data.more » « less
-
We are rapidly approaching a future in which cancer patient digital twins will reach their potential to predict cancer prevention, diagnosis, and treatment in individual patients. This will be realized based on advances in high performance computing, computational modeling, and an expanding repertoire of observational data across multiple scales and modalities. In 2020, the US National Cancer Institute, and the US Department of Energy, through a trans-disciplinary research community at the intersection of advanced computing and cancer research, initiated team science collaborative projects to explore the development and implementation of predictive Cancer Patient Digital Twins. Several diverse pilot projects were launched to provide key insights into important features of this emerging landscape and to determine the requirements for the development and adoption of cancer patient digital twins. Projects included exploring approaches to using a large cohort of digital twins to perform deep phenotyping and plan treatments at the individual level, prototyping self-learning digital twin platforms, using adaptive digital twin approaches to monitor treatment response and resistance, developing methods to integrate and fuse data and observations across multiple scales, and personalizing treatment based on cancer type. Collectively these efforts have yielded increased insights into the opportunities and challenges facing cancer patient digital twin approaches and helped define a path forward. Given the rapidly growing interest in patient digital twins, this manuscript provides a valuable early progress report of several CPDT pilot projects commenced in common, their overall aims, early progress, lessons learned and future directions that will increasingly involve the broader research community.more » « less
-
Abstract Adaptive plasticity in thermal tolerance traits may buffer organisms against changing temperatures, making such responses of particular interest in the face of global climate change. Although population variation is integral to the evolvability of this trait, many studies inferring proxies of physiological vulnerability from thermal tolerance traits extrapolate data from one or a few populations to represent the species. Estimates of physiological vulnerability can be further complicated by methodological effects associated with experimental design. We evaluated how populations varied in their acclimation capacity (i.e., the magnitude of plasticity) for critical thermal maximum (CTmax) in two species of tailed frogs (Ascaphidae), cold‐stream specialists. We used the estimates of acclimation capacity to infer physiological vulnerability to future warming. We performed CTmax experiments on tadpoles from 14 populations using a fully factorial experimental design of two holding temperatures (8 and 15°C) and two experimental starting temperatures (8 and 15°C). This design allowed us to investigate the acute effects of transferring organisms from one holding temperature to a different experimental starting temperature, as well as fully acclimated responses by using the same holding and starting temperature. We found that most populations exhibited beneficial acclimation, where CTmax was higher in tadpoles held at a warmer temperature, but populations varied markedly in the magnitude of the response and the inferred physiological vulnerability to future warming. We also found that the response of transferring organisms to different starting temperatures varied substantially among populations, although accounting for acute effects did not greatly alter estimates of physiological vulnerability at the species level or for most populations. These results underscore the importance of sampling widely among populations when inferring physiological vulnerability, as population variation in acclimation capacity and thermal sensitivity may be critical when assessing vulnerability to future warming.more » « less
-
null (Ed.)Education research has experienced a methodological renaissance over the past two decades, with a new focus on large-scale randomized experiments. This wave of experiments has made education research an even more exciting area for statisticians, unearthing many lessons and challenges in experimental design, causal inference, and statistics more broadly. Importantly, educational research and practice almost always occur in a multilevel setting, which makes the statistics relevant to other fields with this structure, including social policy, health services research, and clinical trials in medicine. In this article we first briefly review the history that led to this new era in education research and describe the design features that dominate the modern large-scale educational experiments. We then highlight some of the key statistical challenges in this area, including endogeneity of design, heterogeneity of treatment effects, noncompliance with treatment assignment, mediation, generalizability, and spillover. Though a secondary focus, we also touch on promising trial designs that answer more nuanced questions, such as the SMART design for studying dynamic treatment regimes and factorial designs for optimizing the components of an existing treatment.more » « less
An official website of the United States government

