skip to main content

Title: Opportunities and Challenges for Machine Learning in Materials Science
Advances in machine learning have impacted myriad areas of materials science, such as the discovery of novel materials and the improvement of molecular simulations, with likely many more important developments to come. Given the rapid changes in this field, it is challenging to understand both the breadth of opportunities and the best practices for their use. In this review, we address aspects of both problems by providing an overview of the areas in which machine learning has recently had significant impact in materials science, and then we provide a more detailed discussion on determining the accuracy and domain of applicability of some common types of machine learning models. Finally, we discuss some opportunities and challenges for the materials community to fully utilize the capabilities of machine learning.  more » « less
Award ID(s):
Author(s) / Creator(s):
Date Published:
Journal Name:
Annual Review of Materials Research
Page Range / eLocation ID:
71 to 103
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Obeid, Iyad Selesnick (Ed.)
    The Temple University Hospital EEG Corpus (TUEG) [1] is the largest publicly available EEG corpus of its type and currently has over 5,000 subscribers (we currently average 35 new subscribers a week). Several valuable subsets of this corpus have been developed including the Temple University Hospital EEG Seizure Corpus (TUSZ) [2] and the Temple University Hospital EEG Artifact Corpus (TUAR) [3]. TUSZ contains manually annotated seizure events and has been widely used to develop seizure detection and prediction technology [4]. TUAR contains manually annotated artifacts and has been used to improve machine learning performance on seizure detection tasks [5]. In this poster, we will discuss recent improvements made to both corpora that are creating opportunities to improve machine learning performance. Two major concerns that were raised when v1.5.2 of TUSZ was released for the Neureka 2020 Epilepsy Challenge were: (1) the subjects contained in the training, development (validation) and blind evaluation sets were not mutually exclusive, and (2) high frequency seizures were not accurately annotated in all files. Regarding (1), there were 50 subjects in dev, 50 subjects in eval, and 592 subjects in train. There was one subject common to dev and eval, five subjects common to dev and train, and 13 subjects common between eval and train. Though this does not substantially influence performance for the current generation of technology, it could be a problem down the line as technology improves. Therefore, we have rebuilt the partitions of the data so that this overlap was removed. This required augmenting the evaluation and development data sets with new subjects that had not been previously annotated so that the size of these subsets remained approximately the same. Since these annotations were done by a new group of annotators, special care was taken to make sure the new annotators followed the same practices as the previous generations of annotators. Part of our quality control process was to have the new annotators review all previous annotations. This rigorous training coupled with a strict quality control process where annotators review a significant amount of each other’s work ensured that there is high interrater agreement between the two groups (kappa statistic greater than 0.8) [6]. In the process of reviewing this data, we also decided to split long files into a series of smaller segments to facilitate processing of the data. Some subscribers found it difficult to process long files using Python code, which tends to be very memory intensive. We also found it inefficient to manipulate these long files in our annotation tool. In this release, the maximum duration of any single file is limited to 60 mins. This increased the number of edf files in the dev set from 1012 to 1832. Regarding (2), as part of discussions of several issues raised by a few subscribers, we discovered some files only had low frequency epileptiform events annotated (defined as events that ranged in frequency from 2.5 Hz to 3 Hz), while others had events annotated that contained significant frequency content above 3 Hz. Though there were not many files that had this type of activity, it was enough of a concern to necessitate reviewing the entire corpus. An example of an epileptiform seizure event with frequency content higher than 3 Hz is shown in Figure 1. Annotating these additional events slightly increased the number of seizure events. In v1.5.2, there were 673 seizures, while in v1.5.3 there are 1239 events. One of the fertile areas for technology improvements is artifact reduction. Artifacts and slowing constitute the two major error modalities in seizure detection [3]. This was a major reason we developed TUAR. It can be used to evaluate artifact detection and suppression technology as well as multimodal background models that explicitly model artifacts. An issue with TUAR was the practicality of the annotation tags used when there are multiple simultaneous events. An example of such an event is shown in Figure 2. In this section of the file, there is an overlap of eye movement, electrode artifact, and muscle artifact events. We previously annotated such events using a convention that included annotating background along with any artifact that is present. The artifacts present would either be annotated with a single tag (e.g., MUSC) or a coupled artifact tag (e.g., MUSC+ELEC). When multiple channels have background, the tags become crowded and difficult to identify. This is one reason we now support a hierarchical annotation format using XML – annotations can be arbitrarily complex and support overlaps in time. Our annotators also reviewed specific eye movement artifacts (e.g., eye flutter, eyeblinks). Eye movements are often mistaken as seizures due to their similar morphology [7][8]. We have improved our understanding of ocular events and it has allowed us to annotate artifacts in the corpus more carefully. In this poster, we will present statistics on the newest releases of these corpora and discuss the impact these improvements have had on machine learning research. We will compare TUSZ v1.5.3 and TUAR v2.0.0 with previous versions of these corpora. We will release v1.5.3 of TUSZ and v2.0.0 of TUAR in Fall 2021 prior to the symposium. ACKNOWLEDGMENTS Research reported in this publication was most recently supported by the National Science Foundation’s Industrial Innovation and Partnerships (IIP) Research Experience for Undergraduates award number 1827565. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the official views of any of these organizations. REFERENCES [1] I. Obeid and J. Picone, “The Temple University Hospital EEG Data Corpus,” in Augmentation of Brain Function: Facts, Fiction and Controversy. Volume I: Brain-Machine Interfaces, 1st ed., vol. 10, M. A. Lebedev, Ed. Lausanne, Switzerland: Frontiers Media S.A., 2016, pp. 394 398. [2] V. Shah et al., “The Temple University Hospital Seizure Detection Corpus,” Frontiers in Neuroinformatics, vol. 12, pp. 1–6, 2018. [3] A. Hamid et, al., “The Temple University Artifact Corpus: An Annotated Corpus of EEG Artifacts.” in Proceedings of the IEEE Signal Processing in Medicine and Biology Symposium (SPMB), 2020, pp. 1-3. [4] Y. Roy, R. Iskander, and J. Picone, “The NeurekaTM 2020 Epilepsy Challenge,” NeuroTechX, 2020. [Online]. Available: [Accessed: 01-Dec-2021]. [5] S. Rahman, A. Hamid, D. Ochal, I. Obeid, and J. Picone, “Improving the Quality of the TUSZ Corpus,” in Proceedings of the IEEE Signal Processing in Medicine and Biology Symposium (SPMB), 2020, pp. 1–5. [6] V. Shah, E. von Weltin, T. Ahsan, I. Obeid, and J. Picone, “On the Use of Non-Experts for Generation of High-Quality Annotations of Seizure Events,” Available: https://www.isip.picone [Accessed: 01-Dec-2021]. [7] D. Ochal, S. Rahman, S. Ferrell, T. Elseify, I. Obeid, and J. Picone, “The Temple University Hospital EEG Corpus: Annotation Guidelines,” Philadelphia, Pennsylvania, USA, 2020. [8] D. Strayhorn, “The Atlas of Adult Electroencephalography,” EEG Atlas Online, 2014. [Online]. Availabl 
    more » « less
  2. This paper describes the development of a facilitator training program that prepares volunteers to offer interactive workshops to build professional skills. This effort to “train the trainers” is part of the CyberAmbassadors workforce development project funded by the National Science Foundation (NSF). The overarching goal of the CyberAmbassadors project is to develop professional skills training that helps participants collaborate more effectively in interdisciplinary settings. The core curriculum for participants includes 20+ hours of materials and activities to build communications, teamwork, and leadership skills. The “train the trainers” project described here is a complementary effort to prepare STEM professionals to facilitate these CyberAmbassadors professional skills trainings for their own workplaces and communities. The facilitator training program was developed and tested with two cohorts, totaling more than 50 participants. Over the course of two days of in-person training, new facilitators had opportunities to experience the core curriculum as participants; to practice facilitation skills and lead group activities; to discuss practical and logistical aspects of offering training in their own communities; and to become familiar with the underlying pedagogy, learning goals, and modular structure of the professional skills curriculum. Surveys were used to collect feedback and evaluate participants’ satisfaction with the CyberAmbassadors professional skills curriculum; their self-assessment of facilitation and professional skills before and after the training; and feedback on the facilitator training experience. Responses from the first cohort of participants were used to refine the facilitator training program and it was offered to a second group of volunteers six months later. In the intervening time, several facilitators from the first cohort implemented CyberAmbassadors trainings at academic institutions, professional conferences, and industry workplaces. Participant surveys were used to provide feedback to the volunteer facilitators and to assist the project coordinators in identifying areas where additional training or support might be helpful. These lessons were used to improve the facilitator training program for the second cohort, and we recruited some of the original volunteers to help lead the second “train the trainers” experience. This approach both provides newer facilitators with additional experience and expands the number of individuals who can “train the trainers” and help to propagate the program for future participants. In addition to describing the experiences and results from this “train the trainers” effort, this paper details the information, planning tools, and supports that are incorporated throughout the CyberAmbassadors professional skills curriculum materials to assist facilitators in offering these trainings. Lessons learned from this project can be adapted to other professional education efforts, both in terms of preparing new instructors and in helping trained facilitators better understand and meet the needs of their audience. 
    more » « less
  3. Abstract

    Data‐driven science and technology have helped achieve meaningful technological advancements in areas such as materials/drug discovery and health care, but efforts to apply high‐end data science algorithms to the areas of glass and ceramics are still limited. Many glass and ceramic researchers are interested in enhancing their work by using more data and data analytics to develop better functional materials more efficiently. Simultaneously, the data science community is looking for a way to access materials data resources to test and validate their advanced computational learning algorithms. To address this issue, The American Ceramic Society (ACerS) convened a Glass and Ceramic Data Science Workshop in February 2018, sponsored by the National Institute for Standards and Technology (NIST) Advanced Manufacturing Technologies (AMTech) program. The workshop brought together a select group of leaders in the data science, informatics, and glass and ceramics communities, ACerS, and Nexight Group to identify the greatest opportunities and mechanisms for facilitating increased collaboration and coordination between these communities. This article summarizes workshop discussions about the current challenges that limit interactions and collaboration between the glass and ceramic and data science communities, opportunities for a coordinated approach that leverages existing knowledge in both communities, and a clear path toward the enhanced use of data science technologies for functional glass and ceramic research and development.

    more » « less
  4. BACKGROUND Optical sensing devices measure the rich physical properties of an incident light beam, such as its power, polarization state, spectrum, and intensity distribution. Most conventional sensors, such as power meters, polarimeters, spectrometers, and cameras, are monofunctional and bulky. For example, classical Fourier-transform infrared spectrometers and polarimeters, which characterize the optical spectrum in the infrared and the polarization state of light, respectively, can occupy a considerable portion of an optical table. Over the past decade, the development of integrated sensing solutions by using miniaturized devices together with advanced machine-learning algorithms has accelerated rapidly, and optical sensing research has evolved into a highly interdisciplinary field that encompasses devices and materials engineering, condensed matter physics, and machine learning. To this end, future optical sensing technologies will benefit from innovations in device architecture, discoveries of new quantum materials, demonstrations of previously uncharacterized optical and optoelectronic phenomena, and rapid advances in the development of tailored machine-learning algorithms. ADVANCES Recently, a number of sensing and imaging demonstrations have emerged that differ substantially from conventional sensing schemes in the way that optical information is detected. A typical example is computational spectroscopy. In this new paradigm, a compact spectrometer first collectively captures the comprehensive spectral information of an incident light beam using multiple elements or a single element under different operational states and generates a high-dimensional photoresponse vector. An advanced algorithm then interprets the vector to achieve reconstruction of the spectrum. This scheme shifts the physical complexity of conventional grating- or interference-based spectrometers to computation. Moreover, many of the recent developments go well beyond optical spectroscopy, and we discuss them within a common framework, dubbed “geometric deep optical sensing.” The term “geometric” is intended to emphasize that in this sensing scheme, the physical properties of an unknown light beam and the corresponding photoresponses can be regarded as points in two respective high-dimensional vector spaces and that the sensing process can be considered to be a mapping from one vector space to the other. The mapping can be linear, nonlinear, or highly entangled; for the latter two cases, deep artificial neural networks represent a natural choice for the encoding and/or decoding processes, from which the term “deep” is derived. In addition to this classical geometric view, the quantum geometry of Bloch electrons in Hilbert space, such as Berry curvature and quantum metrics, is essential for the determination of the polarization-dependent photoresponses in some optical sensors. In this Review, we first present a general perspective of this sensing scheme from the viewpoint of information theory, in which the photoresponse measurement and the extraction of light properties are deemed as information-encoding and -decoding processes, respectively. We then discuss demonstrations in which a reconfigurable sensor (or an array thereof), enabled by device reconfigurability and the implementation of neural networks, can detect the power, polarization state, wavelength, and spatial features of an incident light beam. OUTLOOK As increasingly more computing resources become available, optical sensing is becoming more computational, with device reconfigurability playing a key role. On the one hand, advanced algorithms, including deep neural networks, will enable effective decoding of high-dimensional photoresponse vectors, which reduces the physical complexity of sensors. Therefore, it will be important to integrate memory cells near or within sensors to enable efficient processing and interpretation of a large amount of photoresponse data. On the other hand, analog computation based on neural networks can be performed with an array of reconfigurable devices, which enables direct multiplexing of sensing and computing functions. We anticipate that these two directions will become the engineering frontier of future deep sensing research. On the scientific frontier, exploring quantum geometric and topological properties of new quantum materials in both linear and nonlinear light-matter interactions will enrich the information-encoding pathways for deep optical sensing. In addition, deep sensing schemes will continue to benefit from the latest developments in machine learning. Future highly compact, multifunctional, reconfigurable, and intelligent sensors and imagers will find applications in medical imaging, environmental monitoring, infrared astronomy, and many other areas of our daily lives, especially in the mobile domain and the internet of things. Schematic of deep optical sensing. The n -dimensional unknown information ( w ) is encoded into an m -dimensional photoresponse vector ( x ) by a reconfigurable sensor (or an array thereof), from which w′ is reconstructed by a trained neural network ( n ′ = n and w′   ≈   w ). Alternatively, x may be directly deciphered to capture certain properties of w . Here, w , x , and w′ can be regarded as points in their respective high-dimensional vector spaces ℛ n , ℛ m , and ℛ n ′ . 
    more » « less
  5. Abstract

    Pendant drops of oxide-coated high-surface tension fluids frequently produce perturbed shapes that impede interfacial studies. Eutectic gallium indium or Galinstan are high-surface tension fluids coated with a ∼5 nm gallium oxide (Ga2O3) film and falls under this fluid classification, also known as liquid metals (LMs). The recent emergence of LM-based applications often cannot proceed without analyzing interfacial energetics in different environments. While numerous techniques are available in the literature for interfacial studies- pendant droplet-based analyses are the simplest. However, the perturbed shape of the pendant drops due to the presence of surface oxide has been ignored frequently as a source of error. Also, exploratory investigations of surface oxide leveraging oscillatory pendant droplets have remained untapped. We address both challenges and present two contributing novelties- (a) by utilizing the machine learning (ML) technique, we predict the approximate surface tension value of perturbed pendant droplets, (ii) by leveraging the oscillation-induced bubble tensiometry method, we study the dynamic elastic modulus of the oxide-coated LM droplets. We have created our dataset from LM’s pendant drop shape parameters and trained different models for comparison. We have achieved >99% accuracy with all models and added versatility to work with other fluids. The best-performing model was leveraged further to predict the approximate values of the nonaxisymmetric LM droplets. Then, we analyzed LM’s elastic and viscous moduli in air, harnessing oscillation-induced pendant droplets, which provides complementary opportunities for interfacial studies alternative to expensive rheometers. We believe it will enable more fundamental studies of the oxide layer on LM, leveraging both symmetric and perturbed droplets. Our study broadens the materials science horizon, where researchers from ML and artificial intelligence domains can work synergistically to solve more complex problems related to surface science, interfacial studies, and other studies relevant to LM-based systems.

    more » « less