Abstract Due to their short timescale, stellar flares are a challenging target for the most modern synoptic sky surveys. The upcoming Vera C. Rubin Legacy Survey of Space and Time (LSST), a project designed to collect more data than any precursor survey, is unlikely to detect flares with more than one data point in its main survey. We developed a methodology to enable LSST studies of stellar flares, with a focus on flare temperature and temperature evolution, which remain poorly constrained compared to flare morphology. By leveraging the sensitivity expected from the Rubin system, differential chromatic refraction (DCR) can be used to constrain flare temperature from a single-epoch detection, which will enable statistical studies of flare temperatures and constrain models of the physical processes behind flare emission using the unprecedentedly high volume of data produced by Rubin over the 10 yr LSST. We model the refraction effect as a function of the atmospheric column density, photometric filter, and temperature of the flare, and show that flare temperatures at or above ∼4000 K can be constrained by a singleg-band observation at air massX≳ 1.2, given the minimum specified requirement on the single-visit relative astrometric accuracy of LSST, and that a surprisingly large number of LSST observations are in fact likely be conducted atX≳ 1.2, in spite of image quality requirements pushing the survey to preferentially lowX. Having failed to measure flare DCR in LSST precursor surveys, we make recommendations on survey design and data products that enable these studies in LSST and other future surveys.
more »
« less
Give Me a Few Hours: Exploring Short Timescales in Rubin Observatory Cadence Simulations
Abstract The limiting temporal resolution of a time-domain survey in detecting transient behavior is set by the time between observations of the same sky area. We analyze the distribution of visit separations for a range of Vera C. Rubin Observatory cadence simulations. Simulations from families v1.5–v1.7.1 are strongly peaked at the 22 minute visit pair separation and provide effectively no constraint on temporal evolution within the night. This choice will necessarily prevent Rubin from discovering a wide range of astrophysical phenomena in time to trigger rapid follow-up. We present a science-agnostic metric to supplement detailed simulations of fast-evolving transients and variables and suggest potential approaches for improving the range of timescales explored.
more »
« less
- Award ID(s):
- 1812779
- PAR ID:
- 10361409
- Publisher / Repository:
- DOI PREFIX: 10.3847
- Date Published:
- Journal Name:
- The Astrophysical Journal Supplement Series
- Volume:
- 258
- Issue:
- 1
- ISSN:
- 0067-0049
- Format(s):
- Medium: X Size: Article No. 13
- Size(s):
- Article No. 13
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Abstract The Vera C. Rubin Legacy Survey of Space and Time (LSST) holds the potential to revolutionize time domain astrophysics, reaching completely unexplored areas of the Universe and mapping variability time scales from minutes to a decade. To prepare to maximize the potential of the Rubin LSST data for the exploration of the transient and variable Universe, one of the four pillars of Rubin LSST science, the Transient and Variable Stars Science Collaboration, one of the eight Rubin LSST Science Collaborations, has identified research areas of interest and requirements, and paths to enable them. While our roadmap is ever-evolving, this document represents a snapshot of our plans and preparatory work in the final years and months leading up to the survey’s first light.more » « less
-
Abstract An occupancy model makes use of data that are structured as sets of repeated visits to each of many sites, in order to estimate the actual probability of occupancy (i.e. proportion of occupied sites) after correcting for imperfect detection using the information contained in the sets of repeated observations. We explore the conditions under which preexisting, volunteer-collected data from the citizen science project eBird can be used for fitting occupancy models. Because the majority of eBird’s data are not collected in the form of repeated observations at individual locations, we explore 2 ways in which the single-visit records could be used in occupancy models. First, we assess the potential for space-for-time substitution: aggregating single-visit records from different locations within a region into pseudo-repeat visits. On average, eBird’s observers did not make their observations at locations that were representative of the habitat in the surrounding area, which would lead to biased estimates of occupancy probabilities when using space-for-time substitution. Thus, the use of space-for-time substitution is not always appropriate. Second, we explored the utility of including data from single-visit records to supplement sets of repeated-visit data. In a simulation study we found that inclusion of single-visit records increased the precision of occupancy estimates, but only when detection probabilities are high. When detection probability was low, the addition of single-visit records exacerbated biases in estimates of occupancy probability. We conclude that subsets of data from eBird, and likely from similar projects, can be used for occupancy modeling either using space-for-time substitution or supplementing repeated-visit data with data from single-visit records. The appropriateness of either alternative will depend on the goals of a study and on the probabilities of detection and occupancy of the species of interest.more » « less
-
Abstract The Vera C. Rubin Legacy Survey of Space and Time will discover thousands of microlensing events across the Milky Way, allowing for the study of populations of exoplanets, stars, and compact objects. We evaluate numerous survey strategies simulated in the Rubin Operation Simulations to assess the discovery and characterization efficiencies of microlensing events. We have implemented three metrics in the Rubin Metric Analysis Framework: a discovery metric and two characterization metrics, where one estimates how well the light curve is covered and the other quantifies how precisely event parameters can be determined. We also assess the characterizability of microlensing parallax, critical for detection of free-floating black hole lenses. We find that, given Rubin’s baseline cadence, the discovery and characterization efficiency will be higher for longer-duration and larger-parallax events. Microlensing discovery efficiency is dominated by the observing footprint, where more time spent looking at regions of high stellar density, including the Galactic bulge, Galactic plane, and Magellanic Clouds, leads to higher discovery and characterization rates. However, if the observations are stretched over too wide an area, including low-priority areas of the Galactic plane with fewer stars and higher extinction, event characterization suffers by >10%. This could impact exoplanet, binary star, and compact object events alike. We find that some rolling strategies (where Rubin focuses on a fraction of the sky in alternating years) in the Galactic bulge can lead to a 15%–20% decrease in microlensing parallax characterization, so rolling strategies should be chosen carefully to minimize losses.more » « less
-
Abstract The empirical constitutive modeling framework of rate‐ and state‐dependent friction (RSF) is commonly used to describe the time‐dependent frictional response of fault gouge to perturbations from steady sliding. In a previous study (Ferdowsi & Rubin, 2020), we found that a granular‐physics‐based model of a fault shear zone, with time‐independent properties at the contact scale, reproduces the phenomenology of laboratory rock and gouge friction experiments in velocity‐step and slide‐hold (SH) protocols. A few slide‐hold‐slide (SHS) simulations further suggested that the granular model might outperform current empirical RSF laws in describing laboratory data. Here, we explore the behavior of the same Discrete Element Method (DEM) model in SH and SHS protocols over a wide range of sliding velocities, hold durations, and system stiffnesses, and provide additional support for this view. We find that, similar to laboratory data, the rate of stress decay during SH simulations is in general agreement with the “Slip law” version of the RSF equations, using parameter values determined independently from velocity step tests. During reslides following long hold times, the model, similar to lab data, produces a nearly constant rate of frictional healing with log hold time, with that rate being in the range of ∼0.5 to 1 times the RSF “state evolution” parameterb. We also find that, as in laboratory experiments, the granular layer undergoes log‐time compaction during holds. This is consistent with the traditional understanding of state evolution under the Aging law, even though the associated stress decay is similar to that predicted by the Slip and not the Aging law.more » « less
An official website of the United States government
