skip to main content


Title: Using machine learning techniques to aid environmental policy analysis: a teaching case in big data and electric vehicle infrastructure
For a growing class of prediction problems, big data and machine learning analyses can greatly enhance our understanding of the effectiveness of public investments and public policy. However, the outputs of many machine learning models are often abstract and inaccessible to policy communities or the general public. In this article, we describe a hands-on teaching case that is suitable for use in a graduate or advanced undergraduate public policy, public affairs or environmental studies classroom. Students will engage on the use of increasingly popular machine learning classification algorithms and cloud-based data visualization tools to support policy and planning on the theme of electric vehicle mobility and connected infrastructure. By using these tools, students will critically evaluate and convert large and complex datasets into human understandable visualization for communication and decision-making. The tools also enable user flexibility to engage with streaming data sources in a new creative design with little technical background.  more » « less
Award ID(s):
1931980 1945332
NSF-PAR ID:
10165859
Author(s) / Creator(s):
; ;
Date Published:
Journal Name:
Case studies in the environment
ISSN:
2473-9510
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Composable infrastructure holds the promise of accelerating the pace of academic research and discovery by enabling researchers to tailor the resources of a machine (e.g., GPUs, storage, NICs), on-demand, to address application needs. We were first introduced to composable infrastructure in 2018, and at the same time, there was growing demand among our College of Engineering faculty for GPU systems for data science, artificial intelligence / machine learning / deep learning, and visualization. Many purchased their own individual desktop or deskside systems, a few pursued more costly cloud and HPC solutions, and others looked to the College or campus computer center for GPU resources which, at the time, were scarce. After surveying the diverse needs of our faculty and studying product offerings by a few nascent startups in the composable infrastructure sector, we applied for and received a grant from the National Science Foundation in November 2019 to purchase a mid-scale system, configured to our specifications, for use by faculty and students for research and research training. This paper describes our composable infrastructure solution and implementation for our academic community. Given how modern workflows are progressively moving to containers and cloud frameworks (using Kubernetes) and to programming notebooks (primarily Jupyter), both for ease of use and for ensuring reproducible experiments, we initially adapted these tools for our system. We have since made it simpler to use our system, and now provide our users with a public facing JupyterHub server. We also added an expansion chassis to our system to enable composable co-location, which is a shared central architecture in which our researchers can insert and integrate specialized resources (GPUs, accelerators, networking cards, etc.) needed for their research. In February 2020, installation of our system was finalized and made operational and we began providing access to faculty in the College of Engineering. Now, two years later, it is used by over 40 faculty and students plus some external collaborators for research and research training. Their use cases and experiences are briefly described in this paper. Composable infrastructure has proven to be a useful computational system for workload variability, uneven applications, and modern workflows in academic environments. 
    more » « less
  2. Abstract Practitioner notes

    What is already known about this topic

    Scholarly attention has turned to examining Artificial Intelligence (AI) literacy in K‐12 to help students understand the working mechanism of AI technologies and critically evaluate automated decisions made by computer models.

    While efforts have been made to engage students in understanding AI through building machine learning models with data, few of them go in‐depth into teaching and learning of feature engineering, a critical concept in modelling data.

    There is a need for research to examine students' data modelling processes, particularly in the little‐researched realm of unstructured data.

    What this paper adds

    Results show that students developed nuanced understandings of models learning patterns in data for automated decision making.

    Results demonstrate that students drew on prior experience and knowledge in creating features from unstructured data in the learning task of building text classification models.

    Students needed support in performing feature engineering practices, reasoning about noisy features and exploring features in rich social contexts that the data set is situated in.

    Implications for practice and/or policy

    It is important for schools to provide hands‐on model building experiences for students to understand and evaluate automated decisions from AI technologies.

    Students should be empowered to draw on their cultural and social backgrounds as they create models and evaluate data sources.

    To extend this work, educators should consider opportunities to integrate AI learning in other disciplinary subjects (ie, outside of computer science classes).

     
    more » « less
  3. Abstract  
    more » « less
  4. Abstract

    The use of two‐dimensional images to teach students about three‐dimensional molecules continues to be a prevalent issue in many classrooms. As affordable visualization technologies continue to advance, there has been an increasing interest to utilize novel technology, such as augmented reality (AR), in the development of molecular visualization tools. Existing evaluations of these visual–spatial learning tools focus primarily on student performance and attitude, with little attention toward potential inequity in student participation. Our study adds to the current literature on introducing molecular visualization technology in biochemistry classrooms by examining the potential inequity in a group activity mediated by AR technology. Adapting the participatory equity framework to our specific context, we view equity and inequity in terms of access to the technological conversational floor, a social space created when people enter technology‐mediated joint endeavors. We explore three questions: What are the different ways students interact with an AR model of the potassium channel? What are salient patterns of participation that may signify inequity in classroom technology use? What is the interplay between group social dynamics and the introduction of AR technology in the context of a technology‐mediated group activity? Pairing qualitative analysis with quantitative metrics, our mixed‐methods approach produced a complex story of student participation in an AR‐mediated group activity. The patterns of student participation showed that equity and inequity in an AR‐mediated biochemistry group learning activity are fluid and multifaceted. It was observed that students who gave more explanations during group discussion also had more interactions with the AR model (i.e., they had greater access to the technological conversational floor), and their opinion of the AR model may have greater influence on how their group engage with the AR model. This study provides more nuanced ways of conceptualizing equity and inequity in biochemistry learning settings.

     
    more » « less
  5. Abstract Background

    Machine learning (ML) has emerged as a vital asset for researchers to analyze and extract valuable information from complex datasets. However, developing an effective and robust ML pipeline can present a real challenge, demanding considerable time and effort, thereby impeding research progress. Existing tools in this landscape require a profound understanding of ML principles and programming skills. Furthermore, users are required to engage in the comprehensive configuration of their ML pipeline to obtain optimal performance.

    Results

    To address these challenges, we have developed a novel tool called Machine Learning Made Easy (MLme) that streamlines the use of ML in research, specifically focusing on classification problems at present. By integrating 4 essential functionalities—namely, Data Exploration, AutoML, CustomML, and Visualization—MLme fulfills the diverse requirements of researchers while eliminating the need for extensive coding efforts. To demonstrate the applicability of MLme, we conducted rigorous testing on 6 distinct datasets, each presenting unique characteristics and challenges. Our results consistently showed promising performance across different datasets, reaffirming the versatility and effectiveness of the tool. Additionally, by utilizing MLme’s feature selection functionality, we successfully identified significant markers for CD8+ naive (BACH2), CD16+ (CD16), and CD14+ (VCAN) cell populations.

    Conclusion

    MLme serves as a valuable resource for leveraging ML to facilitate insightful data analysis and enhance research outcomes, while alleviating concerns related to complex coding scripts. The source code and a detailed tutorial for MLme are available at https://github.com/FunctionalUrology/MLme.

     
    more » « less