Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Land surface temperature (LST) derived from satellite observations and weather modeling has been widely used for investigating Earth surface-atmosphere energy exchange and radiation budget. However, satellite-derived LST has a trade-off between spatial and temporal resolutions and missing observations caused by clouds, while there are limitations such as potential bias and expensive computation in model calibration and simulation for weather modeling. To mitigate those limitations, we proposed a WRFM framework to estimate LST at a spatial resolution of 1 km and temporal resolution of an hour by integrating the Weather Research and Forecasting (WRF) model and MODIS satellite data using the morphing technique. We tested the framework in eight counties, Iowa, USA, including urban and rural areas, to generate hourly LSTs from June 1st to August 31st, 2019, at a 1 km resolution. Upon evaluation with in-situ LST measurements, our WRFM framework has demonstrated its ability to capture hourly LSTs under both clear and cloudy conditions, with a root mean square error (RMSE) of 2.63 K and 3.75 K, respectively. Additionally, the assessment with satellite LST observations has shown that the WRFM framework can effectively reduce the bias magnitude in LST from the WRF simulation, resulting in a reduction of the average RMSE over the study area from 4.34 K (daytime) and 4.12 K (nighttime) to 2.89 K (daytime) and 2.75 K (nighttime), respectively, while still capturing the hourly patterns of LST. Overall, the WRFM is effective in integrating the complementary advantages of satellite observations and weather modeling and can generate LSTs with high spatiotemporal resolutions in areas with complex landscapes (e.g., urban).more » « lessFree, publicly-accessible full text available November 20, 2025
-
Free, publicly-accessible full text available October 1, 2025
-
Free, publicly-accessible full text available May 26, 2025
-
Free, publicly-accessible full text available May 26, 2025
-
Gibbons, PhillipB; Pekhimenko, Gennady; De_Sa, Christopher (Ed.)The emergence of ML in various cloud system management tasks (e.g., workload autoscaling and job scheduling) has become a core driver of ML-centric cloud platforms. However, there are still numerous algorithmic and systems challenges that prevent ML-centric cloud platforms from being production-ready. In this paper, we focus on the challenges of model performance variability and costly model retraining, introduced by dynamic workload patterns and heterogeneous applications and infrastructures in cloud environments. To address these challenges, we present FLASH, an extensible framework for fast model adaptation in ML-based system management tasks. We show how FLASH leverages existing ML agents and their training data to learn to generalize across applications/environments with meta-learning. FLASH can be easily integrated with an existing ML-based system management agent with a unified API. We demonstrate the use of FLASH by implementing three existing ML agents that manage (1) resource configurations, (2) autoscaling, and (3) server power. Our experiments show that FLASH enables fast adaptation to new, previously unseen applications/environments (e.g., 5.5× faster than transfer learning in the autoscaling task), indicating significant potential for adopting ML-centric cloud platforms in production.more » « lessFree, publicly-accessible full text available September 1, 2025
-
Gibbons, Phillip B; Gennady, P; De_Sa, Christopher (Ed.)The emergence of ML in various cloud system management tasks (e.g., workload autoscaling and job scheduling) has become a core driver of ML-centric cloud platforms. However, there are still numerous algorithmic and systems challenges that prevent ML-centric cloud platforms from being production-ready. In this paper, we focus on the challenges of model performance variability and costly model retraining, introduced by dynamic workload patterns and heterogeneous applications and infrastructures in cloud environments. To address these challenges, we present FLASH, an extensible framework for fast model adaptation in ML-based system management tasks. We show how FLASH leverages existing ML agents and their training data to learn to generalize across applications/environments with meta-learning. FLASH can be easily integrated with an existing ML-based system management agent with a unified API. We demonstrate the use of FLASH by implementing three existing ML agents that manage (1) resource configurations, (2) autoscaling, and (3) server power. Our experiments show that FLASH enables fast adaptation to new, previously unseen applications/environments (e.g., 5.5× faster than transfer learning in the autoscaling task), indicating significant potential for adopting ML-centric cloud platforms in production.more » « lessFree, publicly-accessible full text available September 1, 2025
-
Begnum, Kyrre; Border, Charles (Ed.)With the increasing popularity of large deep learning model serving workloads, there is a pressing need to reduce the energy consumption of a model-serving cluster while maintaining satisfied throughput or model-serving latency requirements. Model multiplexing approaches such as model parallelism, model placement, replication, and batching aim to optimize the model-serving performance. However, they fall short of leveraging the GPU frequency scaling opportunity for power saving. In this paper, we demonstrate (1) the benefits of GPU frequency scaling in power saving for model serving; and (2) the necessity for co-design and optimization of fine grained model multiplexing and GPU frequency scaling. We explore the co-design space and present a novel power-aware model-serving system, μ-Serve. μ-Serve is a model-serving framework that optimizes the power consumption and model serving latency/throughput of serving multiple ML models efficiently in a homogeneous GPU cluster. Evaluation results on production workloads show that μ-Serve achieves 1.2–2.6× power saving by dynamic GPU frequency scaling (up to 61% reduction) without SLO attainment violations.more » « lessFree, publicly-accessible full text available September 1, 2025
-
Begnum, Kyrre; Border, Charles (Ed.)With the increasing popularity of large deep learning model-serving workloads, there is a pressing need to reduce the energy consumption of a model-serving cluster while maintaining satisfied throughput or model-serving latency requirements. Model multiplexing approaches such as model parallelism, model placement, replication, and batching aim to optimize the model-serving performance. However, they fall short of leveraging the GPU frequency scaling opportunity for power saving. In this paper, we demonstrate (1) the benefits of GPU frequency scaling in power saving for model serving; and (2) the necessity for co-design and optimization of fine-grained model multiplexing and GPU frequency scaling. We explore the co-design space and present a novel power-aware model-serving system, μ-Serve. μ-Serve is a model-serving framework that optimizes the power consumption and model-serving latency/throughput of serving multiple ML models efficiently in a homogeneous GPU cluster. Evaluation results on production workloads show that μ-Serve achieves 1.2–2.6× power saving by dynamic GPU frequency scaling (up to 61% reduction) without SLO attainment violations.more » « lessFree, publicly-accessible full text available September 1, 2025
-
Free, publicly-accessible full text available July 24, 2025