skip to main content


Title: VDAM: VAE based domain adaptation for cloud property retrieval from multi-satellite data
Domain adaptation techniques using deep neural networks have been mainly used to solve the distribution shift problem in homogeneous domains where data usually share similar feature spaces and have the same dimensionalities. Nevertheless, real world applications often deal with heterogeneous domains that come from completely different feature spaces with different dimensionalities. In our remote sensing application, two remote sensing datasets collected by an active sensor and a passive one are heterogeneous. In particular, CALIOP actively measures each atmospheric column. In this study, 25 measured variables/features that are sensitive to cloud phase are used and they are fully labeled. VIIRS is an imaging radiometer, which collects radiometric measurements of the surface and atmosphere in the visible and infrared bands. Recent studies have shown that passive sensors may have difficulties in prediction cloud/aerosol types in complicated atmospheres (e.g., overlapping cloud and aerosol layers, cloud over snow/ice surface, etc.). To overcome the challenge of the cloud property retrieval in passive sensor, we develop a novel VAE based approach to learn domain invariant representation that capture the spatial pattern from multiple satellite remote sensing data (VDAM), to build a domain invariant cloud property retrieval method to accurately classify different cloud types (labels) in the passive sensing dataset. We further exploit the weight based alignment method on the label space to learn a powerful domain adaptation technique that is pertinent to the remote sensing application. Experiments demonstrate our method outperforms other state-of-the-art machine learning methods and achieves higher accuracy in cloud property retrieval in the passive satellite dataset.  more » « less
Award ID(s):
1942714 1730250 1948399
NSF-PAR ID:
10400896
Author(s) / Creator(s):
; ; ;
Date Published:
Journal Name:
Proceedings of the 30th International Conference on Advances in Geographic Information Systems (SIGSPATIAL 22)
Page Range / eLocation ID:
1 to 10
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract. During the Chequamegon Heterogeneous Ecosystem Energy-balance Study Enabled by a High-density Extensive Array of Detectors 2019 (CHEESEHEAD19) field campaign, held in the summer of 2019 in northern Wisconsin, USA, active and passive ground-based remote sensing instruments were deployed to understand the response of the planetary boundary layer to heterogeneous land surface forcing. These instruments include radar wind profilers, microwave radiometers, atmospheric emitted radiance interferometers, ceilometers, high spectral resolution lidars, Doppler lidars, and collaborative lower-atmospheric mobile profiling systems that combine several of these instruments. In this study, these ground-based remote sensing instruments are used to estimate the height of the daytime planetary boundary layer, and their performance is compared against independent boundary layer depth estimates obtained from radiosondes launched as part of the field campaign. The impact of clouds (in particular boundary layer clouds) on boundary layer depth estimations is also investigated. We found that while all instruments are overall able to provide reasonable boundary layer depth estimates, each of them shows strengths and weaknesses under certain conditions. For example, radar wind profilers perform well during cloud-free conditions, and microwave radiometers and atmospheric emitted radiance interferometers have a very good agreement during all conditions but are limited by the smoothness of the retrieved thermodynamic profiles. The estimates from ceilometers and high spectral resolution lidars can be hindered by the presence of elevated aerosol layers or clouds, and the multi-instrument retrieval from the collaborative lower atmospheric mobile profiling systems can be constricted to a limited height range in low-aerosol conditions. 
    more » « less
  2. null (Ed.)
    Abstract. We trained two Random Forest (RF) machine learning models for cloud mask andcloud thermodynamic-phase detection using spectral observations from Visible InfraredImaging Radiometer Suite (VIIRS)on board Suomi National Polar-orbiting Partnership (SNPP). Observations from Cloud-Aerosol Lidarwith Orthogonal Polarization (CALIOP) were carefully selected toprovide reference labels. The two RF models were trained for all-day anddaytime-only conditions using a 4-year collocated VIIRS and CALIOP dataset from2013 to 2016. Due to the orbit difference, the collocated CALIOP and SNPPVIIRS training samples cover a broad-viewing zenith angle range, which is agreat benefit to overall model performance. The all-day model uses three VIIRSinfrared (IR) bands (8.6, 11, and 12 µm), and the daytime model uses fiveNear-IR (NIR) and Shortwave-IR (SWIR) bands (0.86, 1.24, 1.38, 1.64, and 2.25 µm) together with the three IR bands to detect clear, liquid water, and icecloud pixels. Up to seven surface types, i.e., ocean water, forest, cropland,grassland, snow and ice, barren desert, and shrubland, were consideredseparately to enhance performance for both models. Detection of cloudypixels and thermodynamic phase with the two RF models was compared againstcollocated CALIOP products from 2017. It is shown that, when using a conservativescreening process that excludes the most challenging cloudy pixels forpassive remote sensing, the two RF models have high accuracy rates incomparison to the CALIOP reference for both cloud detection andthermodynamic phase. Other existing SNPP VIIRS and Aqua MODIS cloud mask andphase products are also evaluated, with results showing that the two RFmodels and the MODIS MYD06 optical property phase product are the top threealgorithms with respect to lidar observations during the daytime. During thenighttime, the RF all-day model works best for both cloud detection andphase, particularly for pixels over snow and ice surfaces. The present RFmodels can be extended to other similar passive instruments if trainingsamples can be collected from CALIOP or other lidars. However, the qualityof reference labels and potential sampling issues that may impact modelperformance would need further attention. 
    more » « less
  3. Human context recognition (HCR) using sensor data is a crucial task in Context-Aware (CA) applications in domains such as healthcare and security. Supervised machine learning HCR models are trained using smartphone HCR datasets that are scripted or gathered in-the-wild. Scripted datasets are most accurate because of their consistent visit patterns. Supervised machine learning HCR models perform well on scripted datasets but poorly on realistic data. In-the-wild datasets are more realistic, but cause HCR models to perform worse due to data imbalance, missing or incorrect labels, and a wide variety of phone placements and device types. Lab-to-field approaches learn a robust data representation from a scripted, high-fidelity dataset, which is then used for enhancing performance on a noisy, in-the-wild dataset with similar labels. This research introduces Triplet-based Domain Adaptation for Context REcognition (Triple-DARE), a lab-to-field neural network method that combines three unique loss functions to enhance intra-class compactness and inter-class separation within the embedding space of multi-labeled datasets: (1) domain alignment loss in order to learn domain-invariant embeddings; (2) classification loss to preserve task-discriminative features; and (3) joint fusion triplet loss. Rigorous evaluations showed that Triple-DARE achieved 6.3% and 4.5% higher F1-score and classification, respectively, than state-of-the-art HCR baselines and outperformed non-adaptive HCR models by 44.6% and 10.7%, respectively. 
    more » « less
  4. Abstract. The role of clouds in the Arctic radiation budget is not well understood. Ground-based and airborne measurements provide valuable data to test and improve our understanding. However, the ground-based measurements are intrinsically sparse, and the airborne observations are snapshots in time and space. Passive remote sensing measurements from satellite sensors offer high spatial coverage and an evolving time series, having lengths potentially of decades. However, detecting clouds by passive satellite remote sensing sensors is challenging over the Arctic because of the brightness of snow and ice in the ultraviolet and visible spectral regions and because of the small brightness temperature contrast to the surface. Consequently, the quality of the resulting cloud data products needs to be assessed quantitatively. In this study, we validate the cloud data products retrieved from the Advanced Very High Resolution Radiometer (AVHRR) post meridiem (PM) data from the polar-orbiting NOAA-19 satellite and compare them with those derived from the ground-based instruments during the sunlit months. The AVHRR cloud data products by the European Space Agency (ESA) Cloud Climate Change Initiative (Cloud_CCI) project uses the observations in the visible and IR bands to determine cloud properties. The ground-based measurements from four high-latitude sites have been selected for this investigation: Hyytiälä (61.84∘ N, 24.29∘ E), North Slope of Alaska (NSA; 71.32∘ N, 156.61∘ W), Ny-Ålesund (Ny-Å; 78.92∘ N, 11.93∘ E), and Summit (72.59∘ N, 38.42∘ W). The liquid water path (LWP) ground-based data are retrieved from microwave radiometers, while the cloud top height (CTH) has been determined from the integrated lidar–radar measurements. The quality of the satellite products, cloud mask and cloud optical depth (COD), has been assessed using data from NSA, whereas LWP and CTH have been investigated over Hyytiälä, NSA, Ny-Å, and Summit. The Cloud_CCI COD results for liquid water clouds are in better agreement with the NSA radiometer data than those for ice clouds. For liquid water clouds, the Cloud_CCI COD is underestimated roughly by 3 optical depth (OD) units. When ice clouds are included, the underestimation increases to about 5 OD units. The Cloud_CCI LWP is overestimated over Hyytiälä by ≈7 g m−2, over NSA by ≈16 g m−2, and over Ny-Å by ≈24 g m−2. Over Summit, CCI LWP is overestimated for values ≤20 g m−2 and underestimated for values >20 g m−2. Overall the results of the CCI LWP retrievals are within the ground-based instrument uncertainties. To understand the effects of multi-layer clouds on the CTH retrievals, the statistics are compared between the single-layer clouds and all types (single-layer + multi-layer). For CTH retrievals, the Cloud_CCI product overestimates the CTH for single-layer clouds. When the multi-layer clouds are included (i.e., all types), the observed CTH overestimation becomes an underestimation of about 360–420 m. The CTH results over Summit station showed the highest biases compared to the other three sites. To understand the scale-dependent differences between the satellite and ground-based data, the Bland–Altman method is applied. This method does not identify any scale-dependent differences for all the selected cloud parameters except for the retrievals over the Summit station. In summary, the Cloud_CCI cloud data products investigated agree reasonably well with those retrieved from ground-based measurements made at the four high-latitude sites.

     
    more » « less
  5. Satellite remote sensing of aerosol optical depth (AOD) is essential for detection, characterization, and forecasting of wildfire smoke. In this work, we evaluate the AOD (550 nm) retrievals during the extreme wildfire events over the western U.S. in September 2020. Three products are analyzed, including the Moderate-resolution Imaging Spectroradiometers (MODIS) Multi-Angle Implementation of Atmospheric Correction (MAIAC) product collections C6.0 and C6.1, and the NOAA-20 Visible Infrared Imaging Radiometer (VIIRS) AOD from the NOAA Enterprise Processing System (EPS) algorithm. Compared with the Aerosol Robotic Network (AERONET) data, all three products show strong linear correlations with MAIAC C6.1 and VIIRS presenting overall low bias (<0.06). The accuracy of MAIAC C6.1 is found to be substantially improved with respect to MAIAC C6.0 that drastically underestimated AOD over thick smoke, which validates the effectiveness of updates made in MAIAC C6.1 in terms of an improved representation of smoke aerosol optical properties. VIIRS AOD exhibits comparable uncertainty with MAIAC C6.1 with a slight tendency of increased positive bias over the AERONET AOD range of 0.5–3.0. Averaging coincident retrievals from MAIAC C6.1 and VIIRS provides a lower root mean square error and higher correlation than for the individual products, motivating the benefit of blending these datasets. MAIAC C6.1 and VIIRS are further compared to provide insights on their retrieval strategy. When gridded at 0.1° resolution, MAIAC C6.1 and VIIRS provide similar monthly AOD distribution patterns and the latter exhibits a slightly higher domain average. On daily scale, over thick plumes near fire sources, MAIAC C6.1 reports more valid retrievals where VIIRS tends to have retrievals designated as low or medium quality, which tends to be due to internal quality checks. Over transported smoke near scattered clouds, VIIRS provides better retrieval coverage than MAIAC C6.1 owing to its higher spatial resolution, pixel-level processing, and less strict cloud masking. These results can be used as a guide for applications of satellite AOD retrievals during wildfire events and provide insights on future improvement of retrieval algorithms under heavy smoke conditions. 
    more » « less