skip to main content


Title: COVID-19 Brings Data Equity Challenges to the Fore
The COVID-19 pandemic is compelling us to make crucial data-driven decisions quickly, bringing together diverse and unreliable sources of information without the usual quality control mechanisms we may employ. These decisions are consequential at multiple levels: They can inform local, state, and national government policy, be used to schedule access to physical resources such as elevators and workspaces within an organization, and inform contact tracing and quarantine actions for individuals. In all these cases, significant inequities are likely to arise and to be propagated and reinforced by data-driven decision systems. In this article, we propose a framework, called FIDES, for surfacing and reasoning about data equity in these systems.  more » « less
Award ID(s):
1934464 1926250 1934565 1741022
NSF-PAR ID:
10287320
Author(s) / Creator(s):
; ;
Date Published:
Journal Name:
Digital Government: Research and Practice
Volume:
2
Issue:
2
ISSN:
2691-199X
Page Range / eLocation ID:
1 to 7
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    In this paper, we investigate the dichotomy between system design delegation driven by requirement allocation and delegation driven by objective allocation. Specifically, we investigate this dichotomy through the lens of agency theory, which addresses cases where an agent makes decisions on behalf of another, that is, a principal. In current practice, design delegation largely involves requirement allocation as a means to inform agents of the desirable system characteristics. The value‐driven design paradigm proposes replacing requirements with objective, or trade‐off, functions to better guide agents toward optimal systems. We apply and adapt the principal–agent mathematical model to the design delegation problem to determine if a principal, that is, the delegator, should communicate using requirements or objectives with her agent. In this model, we assume the case of a single principal and single agent where the agent has certain domain knowledge the principal does not have and the agent accrues costs while solving a delegated design problem. Under the assumptions of the mathematical model, we show that the requirement allocation paradigm can yield greater value to the principal over objective despite limitations requirement allocation places on the principal to learn information from the agent. However, relaxing model assumptions can impact the value proposition of requirement allocation in favor of objective allocation. Therefore, a resolution to the requirement–objective dichotomy may be context dependent. The results and the analytical framework used to derive them provide a new, foundational perspective with which to investigate allocation strategies.

     
    more » « less
  2. Abstract

    Merging multiple data streams together can improve the overall length of record and achieve the number of observations required for robust statistical analysis. We merge complementary information from different data streams with a regression-based approach to estimate the 1 April snow water equivalent (SWE) volume over Sierra Nevada, USA. We more than double the length of available data-driven SWE volume records by leveragingin-situsnow depth observations from longer-length snow course records and SWE volumes from a shorter-length snow reanalysis. With the resulting data-driven merged time series (1940–2018), we conduct frequency analysis to estimate return periods and associated uncertainty, which can inform decisions about the water supply, drought response, and flood control. We show that the shorter (~30-year) reanalysis results in an underestimation of the 100-year return period by ~25 years (relative to the ~80-year merged dataset). Drought and flood risk and water resources planning can be substantially affected if return periods of SWE, which are closely related to potential flooding in spring and water availability in summer, are misrepresented.

     
    more » « less
  3. Livestock industry is daily producing large amounts of multi-scale data (pathogen-, animal-, site-, system-, regional- level) from different sources such as diagnostic laboratories, trade and production records, management and environmental monitoring systems; however, all these data are still presented and used separately and are largely infra-utilized to timely (i.e., near real-time) inform livestock health decisions. Recent advances in the automation of data capture, standardization, multi-scale integration and sharing/communication (i.e. The Internet Of Things) as well as in the development of novel data mining analytical and visualization capabilities specifically adapted to the livestock industry are dramatically changing this paradigm. As a result, we expect vertical advances in the way we prevent and manage livestock diseases both locally and globally. Our team at the Center for Animal Disease Modeling and Surveillance (CADMS), in collaboration with researchers at Iowa State University and industry leaders at Boehringer Ingelheim and GlobalVetLINK have been working in an exceptional research-industry partnership to develop key data connections and novel Big Data capabilities within the Disease BioPortal (http://bioportal.ucdavis.edu/). This web-based platform includes automation of diagnostic interpretations and facilitates the combined analysis of health, production and trade data using novel space-time-genomic visualization and data mining tools. Access to confidential databases is individually granted with different levels of secure access, visualization and editing capabilities for participating producers, labs, veterinarians and other stakeholders. Each user can create and share customized dashboards and reports to inform risk-based, more cost-effective, decisions at site, system or regional level. Here we will provide practical examples of applications in the swine, poultry and aquaculture industries. We hope to contribute to the more coordinated and effective prevention and control of infectious diseases locally and globally. 
    more » « less
  4. The push to make computer science (CS) education available to all students has been closely followed by increased efforts to collect and report better data on where CS is offered, who is teaching CS, and which students have access to, enroll in, and ultimately benefit from learning CS. These efforts can be highly influential on the evolution of CS education policy, as education leaders and policymakers often rely heavily on data to make decisions. Because of this, it is critical that CS education researchers understand how to collect, analyze, and report data in ways that reflect reality without masking disparities between subpopulations. Similarly, it is important that CS education leaders and policymakers understand how to judiciously interpret the data and translate information into action to scale CS education in ways designed to eliminate inequities. To that end, this article expands on recent research regarding the use of data to assess and inform progress in scaling and broadening participation in CS education. We describe the CAPE framework for assessing equity with respect to the capacity for, access to, participation in, and experience of CS education and explicate how it can be applied to analyze and interpret data to inform policy decisions at multiple levels of educational systems. We provide examples using large, statewide datasets containing educational and demographic information for K-12 students and schools, thereby giving leaders and policymakers a roadmap to assess and address issues of equity in their own schools, districts, or states. We compare and contrast different approaches to measuring and reporting inequities and discuss how data can influence the future of CS education through its impact on policy. 
    more » « less
  5. A new housing development in a familiar neighborhood, a wrong turn that ends up lengthening a Sunday stroll: our internal representation of the world requires constant updating, and we need to be able to associate events separated by long intervals of time to finetune future outcome. This often requires neural connections to be altered. A brain region known as the hippocampus is involved in building and maintaining a map of our environment. However, signals from other brain areas can activate silent neurons in the hippocampus when the body is in a specific location by triggering cellular events called dendritic calcium spikes. Milstein et al. explored whether dendritic calcium spikes in the hippocampus could also help the brain to update its map of the world by enabling neurons to stop being active at one location and to start responding at a new position. Experiments in mice showed that calcium spikes could change which features of the environment individual neurons respond to by strengthening or weaking connections between specific cells. Crucially, this mechanism allowed neurons to associate event sequences that unfold over a longer timescale that was more relevant to the ones encountered in day-to-day life. A computational model was then put together, and it demonstrated that dendritic calcium spikes in the hippocampus could enable the brain to make better spatial decisions in future. Indeed, these spikes are driven by inputs from brain regions involved in complex cognitive processes, potentially enabling the delayed outcomes of navigational choices to guide changes in the activity and wiring of neurons. Overall, the work by Milstein et al. advances the understanding of learning and memory in the brain and may inform the design of better systems for artificial learning. 
    more » « less