skip to main content


Title: Scalable Workflow-Driven Hydrologic Analysis in HydroFrame
The HydroFrame project is a community platform designed to facilitate integrated hydrologic modeling across the US. As a part of HydroFrame, we seek to design innovative workflow solutions that create pathways to enable hydrologic analysis for three target user groups: the modeler, the analyzer, and the domain science educator. We present the initial progress on the HydroFrame community platform using an automated Kepler workflow. This workflow performs end-to-end hydrology simulations involving data ingestion, preprocessing, analysis, modeling, and visualization. We demonstrate how different modules of the workflow can be reused and repurposed for the three target user groups. The Kepler workflow ensures complete reproducibility through a built-in provenance framework that collects workflow specific parameters, software versions, and hardware system configuration. In addition, we aim to optimize the utilization of large-scale computational resources to adjust to the needs of all three user groups. Towards this goal, we present a design that leverages provenance data and machine learning techniques to predict performance and forecast failures using an automatic performance collection component of the pipeline.  more » « less
Award ID(s):
1835855
NSF-PAR ID:
10200442
Author(s) / Creator(s):
; ; ; ;
Editor(s):
Krzhizhanovskaya, Valeria V.; Závodszky, Gábor; Lees, Michael H.; Dongarra, Jack J.; Sloot, Peter M.; Brissos, Sérgio; Teixeira, João
Date Published:
Journal Name:
Computational Science ICCS 2020 (Springer International Publishing)
Volume:
2020
Page Range / eLocation ID:
276-289
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. The creation of high-quality curricular materials requires knowledge of curriculum design and a considerable time commitment. Instructors often have limited time to dedicate to the creation of curricular materials. Additionally, the knowledge and skills needed to develop high-quality materials are often not taught to instructors. Furthermore, similar learning material is often prepared by multiple instructors working at separate institutions, leading to unnecessary duplication of effort and inefficiency that can impact quality. To address these problems, we established the HydroLearn platform and associated professional learning experiences for hydrology and water resources instructors. HydroLearn is an online platform for developing and sharing high-quality curricular materials, or learning modules, focused on hydrology and water resources. The HydroLearn team has worked with three cohorts of instructors from around the world who were dedicated to creating high-quality curricular materials to support both their students and the broader community. In order to overcome some of the aforementioned barriers, we tested and revised several different models of professional learning with these cohorts. These models ranged from (a) instructors working individually with periodic guidance from the HydroLearn team, to (b) small groups of instructors collaborating on topics of shared interests guided through an intensive HydroLearn training workshop. We found the following factors to contribute to the success of instructors in creating modules: (1) instructor pairs co-creating modules enhanced the usability and transferability of modules between universities and courses, (2) dedicating an intensive block of time (∼63 h over 9 days) to both learning about and implementing curriculum design principles, (3) implementing structures for continuous feedback throughout that time, (4) designing modules for use in one’s own course, and (5) instituting a peer-review process to refine modules. A comprehensive set of learning modules were produced covering a wide range of topics that target undergraduate and early graduate students, such as: floodplain analysis, hydrologic droughts, remote sensing applications in hydrology, urbanization and stormwater runoff, evapotranspiration, snow and climate, groundwater flow, saltwater intrusion in coastal regions, and stream solute tracers. We share specifics regarding how we structured the professional learning models, as well as lessons learned and challenges faced. 
    more » « less
  2. We are witnessing a race to meet the ever-growing computation requirements of emerging AI applications to provide perception and control in autonomous vehicles — e.g., self-driving cars and UAVs. To remain competitive, vendors are packing more processing units (CPUs, programmable logic, GPUs, and hardware accelerators) into next-generation multiprocessor systems-on-a-chip (MPSoC). As a result, modern embedded platforms are achieving new heights in peak computational capacity. Unfortunately, however, the collateral and inevitable increase in complexity represents a major obstacle for the development of correct-by-design safety-critical real-time applications. Due to the ever-growing gap between fast-paced hardware evolution and comparatively slower evolution of real-time operating systems (RTOS), there is a need for real-time oriented full-platform management frameworks to complement traditional RTOS designs. In this work, we propose one such framework, namely the X-Stream framework, for the definition, synthesis, and analysis of real-time workloads targeting state-of-the-art accelerator-augmented embedded platforms. Our X-Stream framework is designed around two cardinal principles. First, computation and data movements are orchestrated to achieve predictability by design. For this purpose, iterative computation over large data chunks is divided into subsequent segments. These segments are then streamed leveraging the three-phase execution model (load, execute and unload). Second, the framework is workflow-centric: system designers can specify their workflow and the necessary code for workflow orchestration is automatically generated. In addition to automating the deployment of user-defined hardware-accelerated workloads, X-Stream supports the deployment of some computation segments on traditional CPUs. Finally, X-Stream allows the definition of real-time partitions. Each partition groups applications belonging to the same criticality level and that share the same set of hardware resources, with support for preemptive priority-driven scheduling. Conversely, freedom from interference for applications deployed in different partitions is guaranteed by design. We provide a full-system implementation that includes RTOS integration and showcase the proposed X-Stream framework on a Xilinx Ultrascale+ platform by focusing on a matrix-multiplication and addition kernel use-case. 
    more » « less
  3. Abstract

    Recovery‐based design links building‐level engineering and broader community resilience objectives. However, the relationship between above‐code engineering improvements and recovery performance is highly nonlinear and varies on a building‐ and site‐specific basis, presenting a challenge to both individual owners and code developers. In addition, downtime simulations are computationally expensive and hinder exploration of the full design space. In this paper, we present an optimization framework to identify optimal above‐code design improvements to achieve building‐specific recovery objectives. We supplement the optimization with a workflow to develop surrogate models that (i) rapidly estimate recovery performance under a range of user‐defined improvements, and (ii) enable complex and informative optimization techniques that can be repeated for different stakeholder priorities. We explore the implementation of the framework using a case study office building, with a 50th percentile baseline functional recovery time of 155 days at the 475‐year ground‐motion return period. To optimally achieve a target recovery time of 21 days, we find that nonstructural component enhancements are required, and that increasing structural strength (through increase of the importance factor) can be detrimental. However, for less ambitious target recovery times, we find that the use of larger importance factors eliminates the need for nonstructural component improvements. Such results demonstrate that the relative efficacy of a given recovery‐based design strategy will depend strongly on the design criteria set by the user.

     
    more » « less
  4. null (Ed.)
    With the demand and abundance of information increasing over the last two decades, generations of computer scientists are trying to improve the whole process of information searching, retrieval, and storage. With the diversification of the information sources, users' demand for various requirements of the data has also changed drastically both in terms of usability and performance. Due to the growth of the source material and requirements, correctly sorting, filtering, and storing has given rise to many new challenges in the field. With the help of all four other teams on this project, we are developing an information retrieval, analysis, and storage system to retrieve data from Virginia Tech's Electronic Thesis and Dissertation (ETD), Twitter, and Web Page archives. We seek to provide an appropriate data research and management tool to the users to access specific data. The system will also give certain users the authority to manage and add more data to the system. This project's deliverable will be combined with four others to produce a system usable by Virginia Tech's library system to manage, maintain, and analyze these archives. This report attempts to introduce the system components and design decisions regarding how it has been planned and implemented. Our team has developed a front end web interface that is able to search, retrieve, and manage three important content collection types: ETDs, tweets, and web pages. The interface incorporates a simple hierarchical user permission system, providing different levels of access to its users. In order to facilitate the workflow with other teams, we have containerized this system and made it available on the Virginia Tech cloud server. The system also makes use of a dynamic workflow system using a KnowledgeGraph and Apache Airflow, providing high levels of functional extensibility to the system. This allows curators and researchers to use containerised services for crawling, pre-processing, parsing, and indexing their custom corpora and collections that are available to them in the system. 
    more » « less
  5. null (Ed.)
    The first major goal of this project is to build a state-of-the-art information storage, retrieval, and analysis system that utilizes the latest technology and industry methods. This system is leveraged to accomplish another major goal, supporting modern search and browse capabilities for a large collection of tweets from the Twitter social media platform, web pages, and electronic theses and dissertations (ETDs). The backbone of the information system is a Docker container cluster running with Rancher and Kubernetes. Information retrieval and visualization is accomplished with containers in a pipelined fashion, whether in the cluster or on virtual machines, for Elasticsearch and Kibana, respectively. In addition to traditional searching and browsing, the system supports full-text and metadata searching. Search results include facets as a modern means of browsing among related documents. The system supports text analysis and machine learning to reveal new properties of collection data. These new properties assist in the generation of available facets. Recommendations are also presented with search results based on associations among documents and with logged user activity. The information system is co-designed by five teams of Virginia Tech graduate students, all members of the same computer science class, CS 5604. Although the project is an academic exercise, it is the practice of the teams to work and interact as though they are groups within a company developing a product. The teams on this project include three collection management groups -- Electronic Theses and Dissertations (ETD), Tweets (TWT), and Web-Pages (WP) -- as well as the Front-end (FE) group and the Integration (INT) group to help provide the overarching structure for the application. This submission focuses on the work of the Integration (INT) team, which creates and administers Docker containers for each team in addition to administering the cluster infrastructure. Each container is a customized application environment that is specific to the needs of the corresponding team. Each team will have several of these containers set up in a pipeline formation to allow scaling and extension of the current system. The INT team also contributes to a cross-team effort for exploring the use of Elasticsearch and its internally associated database. The INT team administers the integration of the Ceph data storage system into the CS Department Cloud and provides support for interactions between containers and the Ceph filesystem. During formative stages of development, the INT team also has a role in guiding team evaluations of prospective container components and workflows. The INT team is responsible for the overall project architecture and facilitating the tools and tutorials that assist the other teams in deploying containers in a development environment according to mutual specifications agreed upon with each team. The INT team maintains the status of the Kubernetes cluster, deploying new containers and pods as needed by the collection management teams as they expand their workflows. This team is responsible for utilizing a continuous integration process to update existing containers. During the development stage the INT team collaborates specifically with the collection management teams to create the pipeline for the ingestion and processing of new collection documents, crossing services between those teams as needed. The INT team develops a reasoner engine to construct workflows with information goal as input, which are then programmatically authored, scheduled, and monitored using Apache Airflow. The INT team is responsible for the flow, management, and logging of system performance data and making any adjustments necessary based on the analysis of testing results. The INT team has established a Gitlab repository for archival code related to the entire project and has provided the other groups with the documentation to deposit their code in the repository. This repository will be expanded using Gitlab CI in order to provide continuous integration and testing once it is available. Finally, the INT team will provide a production distribution that includes all embedded Docker containers and sub-embedded Git source code repositories. The INT team will archive this distribution on the Virginia Tech Docker Container Registry and deploy it on the Virginia Tech CS Cloud. The INT-2020 team owes a sincere debt of gratitude to the work of the INT-2019 team. This is a very large undertaking and the wrangling of all of the products and processes would not have been possible without their guidance in both direct and written form. We have relied heavily on the foundation they and their predecessors have provided for us. We continue their work with systematic improvements, but also want to acknowledge their efforts Ibid. Without them, our progress to date would not have been possible. 
    more » « less