skip to main content


Title: A Deep Learning Perspective on Dropwise Condensation
Abstract

Condensation is ubiquitous in nature and industry. Heterogeneous condensation on surfaces is typified by the continuous cycle of droplet nucleation, growth, and departure. Central to the mechanistic understanding of the thermofluidic processes governing condensation is the rapid and high‐fidelity extraction of interpretable physical descriptors from the highly transient droplet population. However, extracting quantifiable measures out of dynamic objects with conventional imaging technologies poses a challenge to researchers. Here, an intelligent vision‐based framework is demonstrated that unites classical thermofluidic imaging techniques with deep learning to fundamentally address this challenge. The deep learning framework can autonomously harness physical descriptors and quantify thermal performance at extreme spatio‐temporal resolutions of 300 nm and 200 ms, respectively. The data‐centric analysis conclusively shows that contrary to classical understanding, the overall condensation performance is governed by a key tradeoff between heat transfer rate per individual droplet and droplet population density. The vision‐based approach presents a powerful tool for the study of not only phase‐change processes but also any nucleation‐based process within and beyond the thermal science community through the harnessing of big data.

 
more » « less
Award ID(s):
1752147
NSF-PAR ID:
10366489
Author(s) / Creator(s):
 ;  ;  ;  ;  ;  ;  ;  ;  
Publisher / Repository:
Wiley Blackwell (John Wiley & Sons)
Date Published:
Journal Name:
Advanced Science
Volume:
8
Issue:
22
ISSN:
2198-3844
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    Artificial intelligence (AI) is shifting the paradigm of two-phase heat transfer research. Recent innovations in AI and machine learning uniquely offer the potential for collecting new types of physically meaningful features that have not been addressed in the past, for making their insights available to other domains, and for solving for physical quantities based on first principles for phase-change thermofluidic systems. This review outlines core ideas of current AI technologies connected to thermal energy science to illustrate how they can be used to push the limit of our knowledge boundaries about boiling and condensation phenomena. AI technologies for meta-analysis, data extraction, and data stream analysis are described with their potential challenges, opportunities, and alternative approaches. Finally, we offer outlooks and perspectives regarding physics-centered machine learning, sustainable cyberinfrastructures, and multidisciplinary efforts that will help foster the growing trend of AI for phase-change heat and mass transfer.

     
    more » « less
  2. Abstract

    Streamflow prediction is a long‐standing hydrologic problem. Development of models for streamflow prediction often requires incorporation of catchment physical descriptors to characterize the associated complex hydrological processes. Across different scales of catchments, these physical descriptors also allow models to extrapolate hydrologic information from one catchment to others, a process referred to as “regionalization”. Recently, in gauged basin scenarios, deep learning models have been shown to achieve state of the art regionalization performance by building a global hydrologic model. These models predict streamflow given catchment physical descriptors and weather forcing data. However, these physical descriptors are by their nature uncertain, sometimes incomplete, or even unavailable in certain cases, which limits the applicability of this approach. In this paper, we show that by assigning a vector of random values as a surrogate for catchment physical descriptors, we can achieve robust regionalization performance under a gauged prediction scenario. Our results show that the deep learning model using our proposed random vector approach achieves a predictive performance comparable to that of the model using actual physical descriptors. The random vector approach yields robust performance under different data sparsity scenarios and deep learning model selections. Furthermore, based on the use of random vectors, high‐dimensional characterization improves regionalization performance in gauged basin scenario when physical descriptors are uncertain, or insufficient.

     
    more » « less
  3. Size-controlled polymer nanodomes (PNDs) benefit a broad cross-section of existing and emerging technologies. Condensed droplet polymerization (CDP) is a vacuum-based synthesis technology that produces PNDs from monomer precursors in a single step. However, the effect of synthesis and processing conditions on the PND size distribution remains elusive. Towards size distribution control, we report the effect of substrate temperature, on which monomer droplets condense, on the size distribution of PNDs. We take a reductionist approach and operate the CDP under batch mode to match the conditions commonly used in condensation research. Notably, despite the rich knowledge base in dropwise condensation, the behavior of nonpolar liquids like a common monomer, i.e., 2-hydroxyethyl methacrylate (HEMA), is not well understood. We bridge that gap by demonstrating that dropwise condensation of HEMA follows a two-stage growth process. Early-stage growth is dominated by drop nucleation and growth, giving rise to relatively uniform sizes with a lognormal distribution, whereas late-stage growth is dominated by the combined effect of drop coalescence and renucleation, leading to a bimodal size distribution. This new framework for understanding the PND size distribution enables an unprecedented population of PNDs. Their controlled size distribution has the potential to enable programmable properties for emergent materials.

     
    more » « less
  4. Abstract

    Deep learning-based image reconstruction methods have achieved remarkable success in phase recovery and holographic imaging. However, the generalization of their image reconstruction performance to new types of samples never seen by the network remains a challenge. Here we introduce a deep learning framework, termed Fourier Imager Network (FIN), that can perform end-to-end phase recovery and image reconstruction from raw holograms of new types of samples, exhibiting unprecedented success in external generalization. FIN architecture is based on spatial Fourier transform modules that process the spatial frequencies of its inputs using learnable filters and a global receptive field. Compared with existing convolutional deep neural networks used for hologram reconstruction, FIN exhibits superior generalization to new types of samples, while also being much faster in its image inference speed, completing the hologram reconstruction task in ~0.04 s per 1 mm2of the sample area. We experimentally validated the performance of FIN by training it using human lung tissue samples and blindly testing it on human prostate, salivary gland tissue and Pap smear samples, proving its superior external generalization and image reconstruction speed. Beyond holographic microscopy and quantitative phase imaging, FIN and the underlying neural network architecture might open up various new opportunities to design broadly generalizable deep learning models in computational imaging and machine vision fields.

     
    more » « less
  5. BACKGROUND Optical sensing devices measure the rich physical properties of an incident light beam, such as its power, polarization state, spectrum, and intensity distribution. Most conventional sensors, such as power meters, polarimeters, spectrometers, and cameras, are monofunctional and bulky. For example, classical Fourier-transform infrared spectrometers and polarimeters, which characterize the optical spectrum in the infrared and the polarization state of light, respectively, can occupy a considerable portion of an optical table. Over the past decade, the development of integrated sensing solutions by using miniaturized devices together with advanced machine-learning algorithms has accelerated rapidly, and optical sensing research has evolved into a highly interdisciplinary field that encompasses devices and materials engineering, condensed matter physics, and machine learning. To this end, future optical sensing technologies will benefit from innovations in device architecture, discoveries of new quantum materials, demonstrations of previously uncharacterized optical and optoelectronic phenomena, and rapid advances in the development of tailored machine-learning algorithms. ADVANCES Recently, a number of sensing and imaging demonstrations have emerged that differ substantially from conventional sensing schemes in the way that optical information is detected. A typical example is computational spectroscopy. In this new paradigm, a compact spectrometer first collectively captures the comprehensive spectral information of an incident light beam using multiple elements or a single element under different operational states and generates a high-dimensional photoresponse vector. An advanced algorithm then interprets the vector to achieve reconstruction of the spectrum. This scheme shifts the physical complexity of conventional grating- or interference-based spectrometers to computation. Moreover, many of the recent developments go well beyond optical spectroscopy, and we discuss them within a common framework, dubbed “geometric deep optical sensing.” The term “geometric” is intended to emphasize that in this sensing scheme, the physical properties of an unknown light beam and the corresponding photoresponses can be regarded as points in two respective high-dimensional vector spaces and that the sensing process can be considered to be a mapping from one vector space to the other. The mapping can be linear, nonlinear, or highly entangled; for the latter two cases, deep artificial neural networks represent a natural choice for the encoding and/or decoding processes, from which the term “deep” is derived. In addition to this classical geometric view, the quantum geometry of Bloch electrons in Hilbert space, such as Berry curvature and quantum metrics, is essential for the determination of the polarization-dependent photoresponses in some optical sensors. In this Review, we first present a general perspective of this sensing scheme from the viewpoint of information theory, in which the photoresponse measurement and the extraction of light properties are deemed as information-encoding and -decoding processes, respectively. We then discuss demonstrations in which a reconfigurable sensor (or an array thereof), enabled by device reconfigurability and the implementation of neural networks, can detect the power, polarization state, wavelength, and spatial features of an incident light beam. OUTLOOK As increasingly more computing resources become available, optical sensing is becoming more computational, with device reconfigurability playing a key role. On the one hand, advanced algorithms, including deep neural networks, will enable effective decoding of high-dimensional photoresponse vectors, which reduces the physical complexity of sensors. Therefore, it will be important to integrate memory cells near or within sensors to enable efficient processing and interpretation of a large amount of photoresponse data. On the other hand, analog computation based on neural networks can be performed with an array of reconfigurable devices, which enables direct multiplexing of sensing and computing functions. We anticipate that these two directions will become the engineering frontier of future deep sensing research. On the scientific frontier, exploring quantum geometric and topological properties of new quantum materials in both linear and nonlinear light-matter interactions will enrich the information-encoding pathways for deep optical sensing. In addition, deep sensing schemes will continue to benefit from the latest developments in machine learning. Future highly compact, multifunctional, reconfigurable, and intelligent sensors and imagers will find applications in medical imaging, environmental monitoring, infrared astronomy, and many other areas of our daily lives, especially in the mobile domain and the internet of things. Schematic of deep optical sensing. The n -dimensional unknown information ( w ) is encoded into an m -dimensional photoresponse vector ( x ) by a reconfigurable sensor (or an array thereof), from which w′ is reconstructed by a trained neural network ( n ′ = n and w′   ≈   w ). Alternatively, x may be directly deciphered to capture certain properties of w . Here, w , x , and w′ can be regarded as points in their respective high-dimensional vector spaces ℛ n , ℛ m , and ℛ n ′ . 
    more » « less