skip to main content

Attention:

The NSF Public Access Repository (NSF-PAR) system and access will be unavailable from 11:00 PM ET on Thursday, October 10 until 2:00 AM ET on Friday, October 11 due to maintenance. We apologize for the inconvenience.


Search for: All records

Creators/Authors contains: "Jha, S"

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Begnum, Kyrre ; Border, Charles (Ed.)
    With the increasing popularity of large deep learning model serving workloads, there is a pressing need to reduce the energy consumption of a model-serving cluster while maintaining satisfied throughput or model-serving latency requirements. Model multiplexing approaches such as model parallelism, model placement, replication, and batching aim to optimize the model-serving performance. However, they fall short of leveraging the GPU frequency scaling opportunity for power saving. In this paper, we demonstrate (1) the benefits of GPU frequency scaling in power saving for model serving; and (2) the necessity for co-design and optimization of fine grained model multiplexing and GPU frequency scaling. We explore the co-design space and present a novel power-aware model-serving system, μ-Serve. μ-Serve is a model-serving framework that optimizes the power consumption and model serving latency/throughput of serving multiple ML models efficiently in a homogeneous GPU cluster. Evaluation results on production workloads show that μ-Serve achieves 1.2–2.6× power saving by dynamic GPU frequency scaling (up to 61% reduction) without SLO attainment violations. 
    more » « less
    Free, publicly-accessible full text available September 1, 2025
  2. Free, publicly-accessible full text available March 13, 2025
  3. nd (Ed.)
    This paper addresses the urgent need to transition to global net-zero carbon emissions by 2050 while retaining the ability to meet joint performance and resilience objectives. The focus is on the computing infrastructures, such as hyperscale cloud datacenters, that consume significant power, thus producing increasing amounts of carbon emissions. Our goal is to (1) optimize the usage of green energy sources (e.g., solar energy), which is desirable but expensive and relatively unstable, and (2) continuously reduce the use of fossil fuels, which have a lower cost but a significant negative societal impact. Meanwhile, cloud datacenters strive to meet their customers’ requirements, e.g., service-level objectives (SLOs) in application latency or throughput, which are impacted by infrastructure resilience and availability. We propose a scalable formulation that combines sustainability, cloud resilience, and performance as a joint optimization problem with multiple interdependent objectives to address these issues holistically. Given the complexity and dynamicity of the problem, machine learning (ML) approaches, such as reinforcement learning, are essential for achieving continuous optimization. Our study highlights the challenges of green energy instability which necessitates innovative MLcentric solutions across heterogeneous infrastructures to manage the transition towards green computing. Underlying the MLcentric solutions must be methods to combine classic system resilience techniques with innovations in real-time ML resilience (not addressed heretofore). We believe that this approach will not only set a new direction in the resilient, SLO-driven adoption of green energy but also enable us to manage future sustainable systems in ways that were not possible before. 
    more » « less
    Free, publicly-accessible full text available January 1, 2025
  4. Out-of-distribution (OOD) detection plays a crucial role in ensuring the safe deployment of deep neural network (DNN) classifiers. While a myriad of methods have focused on improving the performance of OOD detectors, a critical gap remains in interpreting their decisions. We help bridge this gap by providing explanations for OOD detectors based on learned high-level concepts. We first propose two new metrics for assessing the effectiveness of a particular set of concepts for explaining OOD detectors: 1) detection completeness, which quantifies the sufficiency of concepts for explaining an OOD-detector’s decisions, and 2) concept separability, which captures the distributional separation between in-distribution and OOD data in the concept space. Based on these metrics, we propose an unsupervised framework for learning a set of concepts that satisfy the desired properties of high detection completeness and concept separability, and demonstrate its effectiveness in providing concept-based explanations for diverse off-the-shelf OOD detectors. We also show how to identify prominent concepts contributing to the detection results, and provide further reasoning about their decisions. 
    more » « less
  5. Abstract

    We present ultraviolet/optical/near-infrared observations and modeling of Type II supernovae (SNe II) whose early time (δt< 2 days) spectra show transient, narrow emission lines from shock ionization of confined (r< 1015cm) circumstellar material (CSM). The observed electron-scattering broadened line profiles (i.e., IIn-like) of Hi, Hei/ii, Civ, and Niii/iv/vfrom the CSM persist on a characteristic timescale (tIIn) that marks a transition to a lower-density CSM and the emergence of Doppler-broadened features from the fast-moving SN ejecta. Our sample, the largest to date, consists of 39 SNe with early time IIn-like features in addition to 35 “comparison” SNe with no evidence of early time IIn-like features, all with ultraviolet observations. The total sample includes 50 unpublished objects with a total of 474 previously unpublished spectra and 50 multiband light curves, collected primarily through the Young Supernova Experiment and Global Supernova Project collaborations. For all sample objects, we find a significant correlation between peak ultraviolet brightness and bothtIInand the rise time, as well as evidence for enhanced peak luminosities in SNe II with IIn-like features. We quantify mass-loss rates and CSM density for the sample through the matching of peak multiband absolute magnitudes, rise times,tIIn, and optical SN spectra with a grid of radiation hydrodynamics and non-local thermodynamic equilibrium radiative-transfer simulations. For our grid of models, all with the same underlying explosion, there is a trend between the duration of the electron-scattering broadened line profiles and inferred mass-loss rate:tIIn3.8[Ṁ/(0.01Myr−1)] days.

     
    more » « less
    Free, publicly-accessible full text available July 31, 2025
  6. ABSTRACT

    Near-infrared (NIR) observations of normal Type Ia supernovae (SNe Ia) obtained between 150 and 500 d past maximum light reveal the existence of an extended plateau. Here, we present observations of the underluminous, 1991bg-like SN 2021qvv. Early, ground-based optical and NIR observations show that SN 2021qvv is similar to SN 2006mr, making it one of the dimmest, fastest evolving 1991bg-like SNe to date. Late-time (170–250 d) Hubble Space Telescope observations of SN 2021qvv reveal no sign of a plateau. An extrapolation of these observations backwards to earlier-phase NIR observations of SN 2006mr suggests the complete absence of an NIR plateau, at least out to 250 d. This absence may be due to a higher ionization state of the ejecta, as predicted by certain sub-Chandrasekhar-mass detonation models, or to the lower temperatures of the ejecta of 1991bg-like SNe, relative to normal SNe Ia, which might preclude their becoming fluorescent and shifting ultraviolet light into the NIR. This suggestion can be tested by acquiring NIR imaging of a sample of 1991bg-like SNe that covers the entire range from slowly evolving to fast-evolving events (0.2 ≲ sBV ≲ 0.6). A detection of the NIR plateau in slower evolving, hotter 1991bg-like SNe would provide further evidence that these SNe exist along a continuum with normal SNe Ia. Theoretical progenitor and explosion scenarios would then have to match the observed properties of both SN Ia subtypes.

     
    more » « less
  7. Abstract

    A bright (mF150W,AB= 24 mag),z= 1.95 supernova (SN) candidate was discovered in JWST/NIRCam imaging acquired on 2023 November 17. The SN is quintuply imaged as a result of strong gravitational lensing by a foreground galaxy cluster, detected in three locations, and remarkably is the second lensed SN found in the same host galaxy. The previous lensed SN was called “Requiem,” and therefore the new SN is named “Encore.” This makes the MACS J0138.0−2155 cluster the first known system to produce more than one multiply imaged SN. Moreover, both SN Requiem and SN Encore are Type Ia SNe (SNe Ia), making this the most distant case of a galaxy hosting two SNe Ia. Using parametric host fitting, we determine the probability of detecting two SNe Ia in this host galaxy over a ∼10 yr window to be ≈3%. These observations have the potential to yield a Hubble constant (H0) measurement with ∼10% precision, only the third lensed SN capable of such a result, using the three visible images of the SN. Both SN Requiem and SN Encore have a fourth image that is expected to appear within a few years of ∼2030, providing an unprecedented baseline for time-delay cosmography.

     
    more » « less
    Free, publicly-accessible full text available May 29, 2025
  8. This paper presents an approach to detect out-of-context (OOC) objects in an image. Given an image with a set of objects, our goal is to determine if an object is inconsistent with the scene context and detect the OOC object with a bounding box. In this work, we consider commonly explored contextual relations such as co-occurrence relations, the relative size of an object with respect to other objects, and the position of the object in the scene. We posit that contextual cues are useful to determine object labels for in-context objects and inconsistent context cues are detrimental to determining object labels for out-of-context objects. To realize this hypothesis, we propose a graph contextual reasoning network (GCRN) to detect OOC objects. GCRN consists of two separate graphs to predict object labels based on the contextual cues in the image: 1) a representation graph to learn object features based on the neighboring objects and 2) a context graph to explicitly capture contextual cues from the neighboring objects. GCRN explicitly captures the contextual cues to improve the detection of in-context objects and identify objects that violate contextual relations. In order to evaluate our approach, we create a large-scale dataset by adding OOC object instances to the COCO images. We also evaluate on recent OCD benchmark. Our results show that GCRN outperforms competitive baselines in detecting OOC objects and correctly detecting in-context objects. 
    more » « less
  9. Normalizing flows map an independent set of latent variables to their samples using a bijective transformation. Despite the exact correspondence between samples and latent variables, their high level relationship is not well understood. In this paper we characterize the geometric structure of flows using principal manifolds and understand the relationship between latent variables and samples using contours. We introduce a novel class of normalizing flows, called principal component flows (PCF), whose contours are its principal manifolds, and a variant for injective flows (iPCF) that is more efficient to train than regular injective flows. PCFs can be constructed using any flow architecture, are trained with a regularized maximum likelihood objective and can perform density estimation on all of their principal manifolds. In our experiments we show that PCFs and iPCFs are able to learn the principal manifolds over a variety of datasets. Additionally, we show that PCFs can perform density estimation on data that lie on a manifold with variable dimensionality, which is not possible with existing normalizing flows. 
    more » « less