skip to main content


Title: Visualization and ecohydrologic models: Opening the box
Abstract

Earth system models synthesize the science of interactions amongst multiple biophysical and, increasingly, human processes across a wide range of scales. Ecohydrologic models are a subset of earth system models that focus particularly on the complex interactions between ecosystem processes and the storage and flux of water. Ecohydrologic models often focus at scales where direct observations occur: plots, hillslopes, streams, and watersheds, as well as where land and resource management decisions are implemented. These models complement field‐based and data‐driven science by combining theory, empirical relationships derived from observation and new data to create virtual laboratories. Ecohydrologic models are tools that managers can use to ask “what if” questions and domain scientists can use to explore the implications of new theory or measurements. Recent decades have seen substantial advances in ecohydrologic models, building on both new domain science and advances in software engineering and data availability. The increasing sophistication of ecohydrologic models however, presents a barrier to their widespread use and credibility. Their complexity, often encoding 100s of relationships, means that they are effectively “black boxes,” at least for most users, sometimes even to the teams of researchers that contribute to their design. This opacity complicates the interpretation of model results. For models to effectively advance our understanding of how plants and water interact, we must improve how we visualize not only model outputs, but also the underlying theories that are encoded within the models. In this paper, we outline a framework for increasing the usefulness of ecohydrologic models through better visualization. We outline four complementary approaches, ranging from simple best practices that leverage existing technologies, to ideas that would engage novel software engineering and cutting edge human–computer interface design. Our goal is to open the ecohydrologic model black box in ways that will engage multiple audiences, from novices to model developers, and support learning, new discovery, and environmental problem solving.

 
more » « less
Award ID(s):
2012821
NSF-PAR ID:
10452353
Author(s) / Creator(s):
 ;  
Publisher / Repository:
Wiley Blackwell (John Wiley & Sons)
Date Published:
Journal Name:
Hydrological Processes
Volume:
35
Issue:
1
ISSN:
0885-6087
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    Headwater catchments are the fundamental units that connect the land to the ocean. Hydrological flow and biogeochemical processes are intricately coupled, yet their respective sciences have progressed without much integration. Reaction kinetic theories that prescribe rate dependence on environmental variables (e.g., temperature and water content) have advanced substantially, mostly in well‐mixed reactors, columns, and warming experiments without considering the characteristics of hydrological flow at the catchment scale. These theories have shown significant divergence from observations in natural systems. On the other hand, hydrological theories, including transit time theory, have progressed substantially yet have not been incorporated into understanding reactions at the catchment scale. Here we advocate for the development of integrated hydro‐biogeochemical theories across gradients of climate, vegetation, and geology conditions. The lack of such theories presents barriers for understanding mechanisms and forecasting the future of the Critical Zone under human‐ and climate‐induced perturbations. Although integration has started and co‐located measurements are well under way, tremendous challenges remain. In particular, even in this era of “big data,” we are still limited by data and will need to (1) intensify measurements beyond river channels and characterize the vertical connectivity and broadly the shallow and deep subsurface; (2) expand to older water dating beyond the time scales reflected in stable water isotopes; (3) combine the use of reactive solutes, nonreactive tracers, and isotopes; and (4) augment measurements in environments that are undergoing rapid changes. To develop integrated theories, it is essential to (1) engage models at all stages to develop model‐informed data collection strategies and to maximize data usage; (2) adopt a “simple but not simplistic,” or fit‐for‐purpose approach to include essential processes in process‐based models; (3) blend the use of process‐based and data‐driven models in the framework of “theory‐guided data science.” Within the framework of hypothesis testing, model‐data fusion can advance integrated theories that mechanistically link catchments' internal structures and external drivers to their functioning. It can not only advance the field of hydro‐biogeochemistry, but also enable hind‐ and fore‐casting and serve the society at large. Broadly, future education will need to cultivate thinkers at the intersections of traditional disciplines with hollistic approaches for understanding interacting processes in complex earth systems.

    This article is categorized under:

    Engineering Water > Methods

     
    more » « less
  2. null (Ed.)
    The DeepLearningEpilepsyDetectionChallenge: design, implementation, andtestofanewcrowd-sourced AIchallengeecosystem Isabell Kiral*, Subhrajit Roy*, Todd Mummert*, Alan Braz*, Jason Tsay, Jianbin Tang, Umar Asif, Thomas Schaffter, Eren Mehmet, The IBM Epilepsy Consortium◊ , Joseph Picone, Iyad Obeid, Bruno De Assis Marques, Stefan Maetschke, Rania Khalaf†, Michal Rosen-Zvi† , Gustavo Stolovitzky† , Mahtab Mirmomeni† , Stefan Harrer† * These authors contributed equally to this work † Corresponding authors: rkhalaf@us.ibm.com, rosen@il.ibm.com, gustavo@us.ibm.com, mahtabm@au1.ibm.com, sharrer@au.ibm.com ◊ Members of the IBM Epilepsy Consortium are listed in the Acknowledgements section J. Picone and I. Obeid are with Temple University, USA. T. Schaffter is with Sage Bionetworks, USA. E. Mehmet is with the University of Illinois at Urbana-Champaign, USA. All other authors are with IBM Research in USA, Israel and Australia. Introduction This decade has seen an ever-growing number of scientific fields benefitting from the advances in machine learning technology and tooling. More recently, this trend reached the medical domain, with applications reaching from cancer diagnosis [1] to the development of brain-machine-interfaces [2]. While Kaggle has pioneered the crowd-sourcing of machine learning challenges to incentivise data scientists from around the world to advance algorithm and model design, the increasing complexity of problem statements demands of participants to be expert data scientists, deeply knowledgeable in at least one other scientific domain, and competent software engineers with access to large compute resources. People who match this description are few and far between, unfortunately leading to a shrinking pool of possible participants and a loss of experts dedicating their time to solving important problems. Participation is even further restricted in the context of any challenge run on confidential use cases or with sensitive data. Recently, we designed and ran a deep learning challenge to crowd-source the development of an automated labelling system for brain recordings, aiming to advance epilepsy research. A focus of this challenge, run internally in IBM, was the development of a platform that lowers the barrier of entry and therefore mitigates the risk of excluding interested parties from participating. The challenge: enabling wide participation With the goal to run a challenge that mobilises the largest possible pool of participants from IBM (global), we designed a use case around previous work in epileptic seizure prediction [3]. In this “Deep Learning Epilepsy Detection Challenge”, participants were asked to develop an automatic labelling system to reduce the time a clinician would need to diagnose patients with epilepsy. Labelled training and blind validation data for the challenge were generously provided by Temple University Hospital (TUH) [4]. TUH also devised a novel scoring metric for the detection of seizures that was used as basis for algorithm evaluation [5]. In order to provide an experience with a low barrier of entry, we designed a generalisable challenge platform under the following principles: 1. No participant should need to have in-depth knowledge of the specific domain. (i.e. no participant should need to be a neuroscientist or epileptologist.) 2. No participant should need to be an expert data scientist. 3. No participant should need more than basic programming knowledge. (i.e. no participant should need to learn how to process fringe data formats and stream data efficiently.) 4. No participant should need to provide their own computing resources. In addition to the above, our platform should further • guide participants through the entire process from sign-up to model submission, • facilitate collaboration, and • provide instant feedback to the participants through data visualisation and intermediate online leaderboards. The platform The architecture of the platform that was designed and developed is shown in Figure 1. The entire system consists of a number of interacting components. (1) A web portal serves as the entry point to challenge participation, providing challenge information, such as timelines and challenge rules, and scientific background. The portal also facilitated the formation of teams and provided participants with an intermediate leaderboard of submitted results and a final leaderboard at the end of the challenge. (2) IBM Watson Studio [6] is the umbrella term for a number of services offered by IBM. Upon creation of a user account through the web portal, an IBM Watson Studio account was automatically created for each participant that allowed users access to IBM's Data Science Experience (DSX), the analytics engine Watson Machine Learning (WML), and IBM's Cloud Object Storage (COS) [7], all of which will be described in more detail in further sections. (3) The user interface and starter kit were hosted on IBM's Data Science Experience platform (DSX) and formed the main component for designing and testing models during the challenge. DSX allows for real-time collaboration on shared notebooks between team members. A starter kit in the form of a Python notebook, supporting the popular deep learning libraries TensorFLow [8] and PyTorch [9], was provided to all teams to guide them through the challenge process. Upon instantiation, the starter kit loaded necessary python libraries and custom functions for the invisible integration with COS and WML. In dedicated spots in the notebook, participants could write custom pre-processing code, machine learning models, and post-processing algorithms. The starter kit provided instant feedback about participants' custom routines through data visualisations. Using the notebook only, teams were able to run the code on WML, making use of a compute cluster of IBM's resources. The starter kit also enabled submission of the final code to a data storage to which only the challenge team had access. (4) Watson Machine Learning provided access to shared compute resources (GPUs). Code was bundled up automatically in the starter kit and deployed to and run on WML. WML in turn had access to shared storage from which it requested recorded data and to which it stored the participant's code and trained models. (5) IBM's Cloud Object Storage held the data for this challenge. Using the starter kit, participants could investigate their results as well as data samples in order to better design custom algorithms. (6) Utility Functions were loaded into the starter kit at instantiation. This set of functions included code to pre-process data into a more common format, to optimise streaming through the use of the NutsFlow and NutsML libraries [10], and to provide seamless access to the all IBM services used. Not captured in the diagram is the final code evaluation, which was conducted in an automated way as soon as code was submitted though the starter kit, minimising the burden on the challenge organising team. Figure 1: High-level architecture of the challenge platform Measuring success The competitive phase of the "Deep Learning Epilepsy Detection Challenge" ran for 6 months. Twenty-five teams, with a total number of 87 scientists and software engineers from 14 global locations participated. All participants made use of the starter kit we provided and ran algorithms on IBM's infrastructure WML. Seven teams persisted until the end of the challenge and submitted final solutions. The best performing solutions reached seizure detection performances which allow to reduce hundred-fold the time eliptologists need to annotate continuous EEG recordings. Thus, we expect the developed algorithms to aid in the diagnosis of epilepsy by significantly shortening manual labelling time. Detailed results are currently in preparation for publication. Equally important to solving the scientific challenge, however, was to understand whether we managed to encourage participation from non-expert data scientists. Figure 2: Primary occupation as reported by challenge participants Out of the 40 participants for whom we have occupational information, 23 reported Data Science or AI as their main job description, 11 reported being a Software Engineer, and 2 people had expertise in Neuroscience. Figure 2 shows that participants had a variety of specialisations, including some that are in no way related to data science, software engineering, or neuroscience. No participant had deep knowledge and experience in data science, software engineering and neuroscience. Conclusion Given the growing complexity of data science problems and increasing dataset sizes, in order to solve these problems, it is imperative to enable collaboration between people with differences in expertise with a focus on inclusiveness and having a low barrier of entry. We designed, implemented, and tested a challenge platform to address exactly this. Using our platform, we ran a deep-learning challenge for epileptic seizure detection. 87 IBM employees from several business units including but not limited to IBM Research with a variety of skills, including sales and design, participated in this highly technical challenge. 
    more » « less
  3. Site description. This data package consists of data obtained from sampling surface soil (the 0-7.6 cm depth profile) in black mangrove (Avicennia germinans) dominated forest and black needlerush (Juncus roemerianus) saltmarsh along the Gulf of Mexico coastline in peninsular west-central Florida, USA. This location has a subtropical climate with mean daily temperatures ranging from 15.4 °C in January to 27.8 °C in August, and annual precipitation of 1336 mm. Precipitation falls as rain primarily between June and September. Tides are semi-diurnal, with 0.57 m median amplitudes during the year preceding sampling (U.S. NOAA National Ocean Service, Clearwater Beach, Florida, station 8726724). Sea-level rise is 4.0 ± 0.6 mm per year (1973-2020 trend, mean ± 95 % confidence interval, NOAA NOS Clearwater Beach station). The A. germinans mangrove zone is either adjacent to water or fringed on the seaward side by a narrow band of red mangrove (Rhizophora mangle). A near-monoculture of J. roemerianus is often adjacent to and immediately landward of the A. germinans zone. The transition from the mangrove to the J. roemerianus zone is variable in our study area. An abrupt edge between closed-canopy mangrove and J. roemerianus monoculture may extend for up to several hundred meters in some locations, while other stretches of ecotone present a gradual transition where smaller, widely spaced trees are interspersed into the herbaceous marsh. Juncus roemerianus then extends landward to a high marsh patchwork of succulent halophytes (including Salicornia bigellovi, Sesuvium sp., and Batis maritima), scattered dwarf mangrove, and salt pans, followed in turn by upland vegetation that includes Pinus sp. and Serenoa repens. Field design and sample collection. We established three study sites spaced at approximately 5 km intervals along the western coastline of the central Florida peninsula. The sites consisted of the Salt Springs (28.3298°, -82.7274°), Energy Marine Center (28.2903°, -82.7278°), and Green Key (28.2530°, -82.7496°) sites on the Gulf of Mexico coastline in Pasco County, Florida, USA. At each site, we established three plot pairs, each consisting of one saltmarsh plot and one mangrove plot. Plots were 50 m^2 in size. Plots pairs within a site were separated by 230-1070 m, and the mangrove and saltmarsh plots composing a pair were 70-170 m apart. All plot pairs consisted of directly adjacent patches of mangrove forest and J. roemerianus saltmarsh, with the mangrove forests exhibiting a closed canopy and a tree architecture (height 4-6 m, crown width 1.5-3 m). Mangrove plots were located at approximately the midpoint between the seaward edge (water-mangrove interface) and landward edge (mangrove-marsh interface) of the mangrove zone. Saltmarsh plots were located 20-25 m away from any mangrove trees and into the J. roemerianus zone (i.e., landward from the mangrove-marsh interface). Plot pairs were coarsely similar in geomorphic setting, as all were located on the Gulf of Mexico coastline, rather than within major sheltering formations like Tampa Bay, and all plot pairs fit the tide-dominated domain of the Woodroffe classification (Woodroffe, 2002, "Coasts: Form, Process and Evolution", Cambridge University Press), given their conspicuous semi-diurnal tides. There was nevertheless some geomorphic variation, as some plot pairs were directly open to the Gulf of Mexico while others sat behind keys and spits or along small tidal creeks. Our use of a plot-pair approach is intended to control for this geomorphic variation. Plot center elevations (cm above mean sea level, NAVD 88) were estimated by overlaying the plot locations determined with a global positioning system (Garmin GPS 60, Olathe, KS, USA) on a LiDAR-derived bare-earth digital elevation model (Dewberry, Inc., 2019). The digital elevation model had a vertical accuracy of ± 10 cm (95 % CI) and a horizontal accuracy of ± 116 cm (95 % CI). Soil samples were collected via coring at low tide in June 2011. From each plot, we collected a composite soil sample consisting of three discrete 5.1 cm diameter soil cores taken at equidistant points to 7.6 cm depth. Cores were taken by tapping a sleeve into the soil until its top was flush with the soil surface, sliding a hand under the core, and lifting it up. Cores were then capped and transferred on ice to our laboratory at the University of South Florida (Tampa, Florida, USA), where they were combined in plastic zipper bags, and homogenized by hand into plot-level composite samples on the day they were collected. A damp soil subsample was immediately taken from each composite sample to initiate 1 y incubations for determination of active C and N (see below). The remainder of each composite sample was then placed in a drying oven (60 °C) for 1 week with frequent mixing of the soil to prevent aggregation and liberate water. Organic wetland soils are sometimes dried at 70 °C, however high drying temperatures can volatilize non-water liquids and oxidize and decompose organic matter, so 50 °C is also a common drying temperature for organic soils (Gardner 1986, "Methods of Soil Analysis: Part 1", Soil Science Society of America); we accordingly chose 60 °C as a compromise between sufficient water removal and avoidance of non-water mass loss. Bulk density was determined as soil dry mass per core volume (adding back the dry mass equivalent of the damp subsample removed prior to drying). Dried subsamples were obtained for determination of soil organic matter (SOM), mineral texture composition, and extractable and total carbon (C) and nitrogen (N) within the following week. Sample analyses. A dried subsample was apportioned from each composite sample to determine SOM as mass loss on ignition at 550 °C for 4 h. After organic matter was removed from soil via ignition, mineral particle size composition was determined using a combination of wet sieving and density separation in 49 mM (3 %) sodium hexametaphosphate ((NaPO_3)_6) following procedures in Kettler et al. (2001, Soil Science Society of America Journal 65, 849-852). The percentage of dry soil mass composed of silt and clay particles (hereafter, fines) was calculated as the mass lost from dispersed mineral soil after sieving (0.053 mm mesh sieve). Fines could have been slightly underestimated if any clay particles were burned off during the preceding ignition of soil. An additional subsample was taken from each composite sample to determine extractable N and organic C concentrations via 0.5 M potassium sulfate (K_2SO_4) extractions. We combined soil and extractant (ratio of 1 g dry soil:5 mL extractant) in plastic bottles, reciprocally shook the slurry for 1 h at 120 rpm, and then gravity filtered it through Fisher G6 (1.6 μm pore size) glass fiber filters, followed by colorimetric detection of nitrite (NO_2^-) + nitrate (NO_3^-) and ammonium (NH_4^+) in the filtrate (Hood Nowotny et al., 2010,Soil Science Society of America Journal 74, 1018-1027) using a microplate spectrophotometer (Biotek Epoch, Winooski, VT, USA). Filtrate was also analyzed for dissolved organic C (referred to hereafter as extractable organic C) and total dissolved N via combustion and oxidation followed by detection of the evolved CO_2 and N oxide gases on a Formacs HT TOC/TN analyzer (Skalar, Breda, The Netherlands). Extractable organic N was then computed as total dissolved N in filtrate minus extractable mineral N (itself the sum of extractable NH_4-N and NO_2-N + NO_3-N). We determined soil total C and N from dried, milled subsamples subjected to elemental analysis (ECS 4010, Costech, Inc., Valencia, CA, USA) at the University of South Florida Stable Isotope Laboratory. Median concentration of inorganic C in unvegetated surface soil at our sites is 0.5 % of soil mass (Anderson, 2019, Univ. of South Florida M.S. thesis via methods in Wang et al., 2011, Environmental Monitoring and Assessment 174, 241-257). Inorganic C concentrations are likely even lower in our samples from under vegetation, where organic matter would dilute the contribution of inorganic C to soil mass. Nevertheless, the presence of a small inorganic C pool in our soils may be counted in the total C values we report. Extractable organic C is necessarily of organic C origin given the method (sparging with HCl) used in detection. Active C and N represent the fractions of organic C and N that are mineralizable by soil microorganisms under aerobic conditions in long-term soil incubations. To quantify active C and N, 60 g of field-moist soil were apportioned from each composite sample, placed in a filtration apparatus, and incubated in the dark at 25 °C and field capacity moisture for 365 d (as in Lewis et al., 2014, Ecosphere 5, art59). Moisture levels were maintained by frequently weighing incubated soil and wetting them up to target mass. Daily CO_2 flux was quantified on 29 occasions at 0.5-3 week intervals during the incubation period (with shorter intervals earlier in the incubation), and these per day flux rates were integrated over the 365 d period to compute an estimate of active C. Observations of per day flux were made by sealing samples overnight in airtight chambers fitted with septa and quantifying headspace CO_2 accumulation by injecting headspace samples (obtained through the septa via needle and syringe) into an infrared gas analyzer (PP Systems EGM 4, Amesbury, MA, USA). To estimate active N, each incubated sample was leached with a C and N free, 35 psu solution containing micronutrients (Nadelhoffer, 1990, Soil Science Society of America Journal 54, 411-415) on 19 occasions at increasing 1-6 week intervals during the 365 d incubation, and then extracted in 0.5 M K_2SO_4 at the end of the incubation in order to remove any residual mineral N. Active N was then quantified as the total mass of mineral N leached and extracted. Mineral N in leached and extracted solutions was detected as NH_4-N and NO_2-N + NO_3-N via colorimetry as above. This incubation technique precludes new C and N inputs and persistently leaches mineral N, forcing microorganisms to meet demand by mineralizing existing pools, and thereby directly assays the potential activity of soil organic C and N pools present at the time of soil sampling. Because this analysis commences with disrupting soil physical structure, it is biased toward higher estimates of active fractions. Calculations. Non-mobile C and N fractions were computed as total C and N concentrations minus the extractable and active fractions of each element. This data package reports surface-soil constituents (moisture, fines, SOM, and C and N pools and fractions) in both gravimetric units (mass constituent / mass soil) and areal units (mass constituent / soil surface area integrated through 7.6 cm soil depth, the depth of sampling). Areal concentrations were computed as X × D × 7.6, where X is the gravimetric concentration of a soil constituent, D is soil bulk density (g dry soil / cm^3), and 7.6 is the sampling depth in cm. 
    more » « less
  4. There are significant disparities between the conferring of science, technology, engineering, and mathematics (STEM) bachelor’s degrees to minoritized groups and the number of STEM faculty that represent minoritized groups at four-year predominantly White institutions (PWIs). Studies show that as of 2019, African American faculty at PWIs have increased by only 2.3% in the last 20 years. This study explores the ways in which this imbalance affects minoritized students in engineering majors. Our research objective is to describe the ways in which African American students navigate their way to success in an engineering program at a PWI where the minoritized faculty representation is less than 10%. In this study, we define success as completion of an undergraduate degree and matriculation into a Ph.D. program. Research shows that African American students struggle with feeling like the “outsider within” in graduate programs and that the engineering culture can permeate from undergraduate to graduate programs. We address our research objective by conducting interviews using navigational capital as our theoretical framework, which can be defined as resilience, academic invulnerability, and skills. These three concepts come together to denote the journey of an individual as they achieve success in an environment not created with them in mind. Navigational capital has been applied in education contexts to study minoritized groups, and specifically in engineering education to study the persistence of students of color. Research on navigational capital often focuses on how participants acquire resources from others. There is a limited focus on the experience of the student as the individual agent exercising their own navigational capital. Drawing from and adapting the framework of navigational capital, this study provides rich descriptions of the lived experiences of African American students in an engineering program at a PWI as they navigated their way to academic success in a system that was not designed with them in mind. This pilot study took place at a research-intensive, land grant PWI in the southeastern United States. We recruited two students who identify as African American and are in the first year of their Ph.D. program in an engineering major. Our interview protocol was adapted from a related study about student motivation, identity, and sense of belonging in engineering. After transcribing interviews with these participants, we began our qualitative analysis with a priori coding, drawing from the framework of navigational capital, to identify the experiences, connections, involvement, and resources the participants tapped into as they maneuvered their way to success in an undergraduate engineering program at a PWI. To identify other aspects of the participants’ experiences that were not reflected in that framework, we also used open coding. The results showed that the participants tapped into their navigational capital when they used experiences, connections, involvement, and resources to be resilient, academically invulnerable, and skillful. They learned from experiences (theirs or others’), capitalized on their connections, positioned themselves through involvement, and used their resources to achieve success in their engineering program. The participants identified their experiences, connections, and involvement. For example, one participant who came from a blended family (African American and White) drew from the experiences she had with her blended family. Her experiences helped her to understand the cultures of Black and White people. She was able to turn that into a skill to connect with others at her PWI. The point at which she took her familial experiences to use as a skill to maneuver her way to success at a PWI was an example of her navigational capital. Another participant capitalized on his connections to develop academic invulnerability. He was able to build his connections by making meaningful relationships with his classmates. He knew the importance of having reliable people to be there for him when he encountered a topic he did not understand. He cultivated an environment through relationships with classmates that set him up to achieve academic invulnerability in his classes. The participants spoke least about how they used their resources. The few mentions of resources were not distinct enough to make any substantial connection to the factors that denote navigational capital. The participants spoke explicitly about the PWI culture in their engineering department. From open coding, we identified the theme that participants did not expect to have role models in their major that looked like them and went into their undergraduate experience with the understanding that they will be the distinct minority in their classes. They did not make notable mention of how a lack of minority faculty affected their success. Upon acceptance, they took on the challenge of being a racial minority in exchange for a well-recognized degree they felt would have more value compared to engineering programs at other universities. They identified ways they maneuvered around their expectation that they would not have representative role models through their use of navigational capital. Integrating knowledge from the framework of navigational capital and its existing applications in engineering and education allows us the opportunity to learn from African American students that have succeeded in engineering programs with low minority faculty representation. The future directions of this work are to outline strategies that could enhance the path of minoritized engineering students towards success and to lay a foundation for understanding the use of navigational capital by minoritized students in engineering at PWIs. Students at PWIs can benefit from understanding their own navigational capital to help them identify ways to successfully navigate educational institutions. Students’ awareness of their capacity to maintain high levels of achievement, their connections to networks that facilitate navigation, and their ability to draw from experiences to enhance resilience provide them with the agency to unleash the invisible factors of their potential to be innovators in their collegiate and work environments. 
    more » « less
  5. Abstract

    As machine learning (ML) has matured, it has opened a new frontier in theoretical and computational chemistry by offering the promise of simultaneous paradigm shifts in accuracy and efficiency. Nowhere is this advance more needed, but also more challenging to achieve, than in the discovery of open‐shell transition metal complexes. Here, localizeddorfelectrons exhibit variable bonding that is challenging to capture even with the most computationally demanding methods. Thus, despite great promise, clear obstacles remain in constructing ML models that can supplement or even replace explicit electronic structure calculations. In this article, I outline the recent advances in building ML models in transition metal chemistry, including the ability to approach sub‐kcal/mol accuracy on a range of properties with tailored representations, to discover and enumerate complexes in large chemical spaces, and to reveal opportunities for design through analysis of feature importance. I discuss unique considerations that have been essential to enabling ML in open‐shell transition metal chemistry, including (a) the relationship of data set size/diversity, model complexity, and representation choice, (b) the importance of quantitative assessments of both theory and model domain of applicability, and (c) the need to enable autonomous generation of reliable, large data sets both for ML model training and in active learning or discovery contexts. Finally, I summarize the next steps toward making ML a mainstream tool in the accelerated discovery of transition metal complexes.

    This article is categorized under:

    Electronic Structure Theory > Density Functional Theory

    Software > Molecular Modeling

    Computer and Information Science > Chemoinformatics

     
    more » « less