skip to main content

Title: Unequal Impacts of Augmented Reality on Learning and Collaboration During Robot Programming with Peers
Augmented reality (AR) applications are growing in popularity in educational settings. While the effects of AR experiences on learning have been widely studied, there is relatively less research on understanding the impact of AR on the dynamics of co-located collaborative learning, specifically in the context of novices programming robots. Educational robotics are a powerful learning context because they engage students with problem solving, critical thinking, STEM (Science, Technology, Engineering, Mathematics) concepts, and collaboration skills. However, such collaborations can suffer due to students having unequal access to resources or dominant peers. In this research we investigate how augmented reality impacts learning and collaboration while peers engage in robot programming activities. We use a mixed methods approach to measure how participants are learning, manipulating resources, and engaging in problem solving activities with peers. We investigate how these behaviors are impacted by the presence of augmented reality visualizations, and by participants? proximity to resources. We find that augmented reality improved overall group learning and collaboration. Detailed analysis shows that AR strongly helps one participant more than the other, by improving their ability to learn and contribute while remaining engaged with the robot. Furthermore, augmented reality helps both participants maintain a common ground and more » balance contributions during problem solving activities. We discuss the implications of these results for designing AR and non-AR collaborative interfaces. « less
Authors:
; ;
Award ID(s):
1917716
Publication Date:
NSF-PAR ID:
10276293
Journal Name:
Proceedings of the ACM on Human-Computer Interaction
Volume:
4
Issue:
CSCW3
Page Range or eLocation-ID:
1 to 23
ISSN:
2573-0142
Sponsoring Org:
National Science Foundation
More Like this
  1. Over the past two decades, educators have used computer-supported collaborative learning (CSCL) to integrate technology with pedagogy to improve student engagement and learning outcomes. Researchers have also explored the diverse affordances of CSCL, its contributions to engineering instruction, and its effectiveness in K-12 STEM education. However, the question of how students use CSCL resources in undergraduate engineering classrooms remains largely unexplored. This study examines the affordances of a CSCL environment utilized in a sophomore dynamics course with particular attention given to the undergraduate engineering students’ use of various CSCL resources. The resources include a course lecturebook, instructor office hours, amore »teaching assistant help room, online discussion board, peer collaboration, and demonstration videos. This qualitative study uses semi-structured interview data collected from nine mechanical engineering students (four women and five men) who were enrolled in a dynamics course at a large public research university in Eastern Canada. The interviews focused on the individual student’s perceptions of the school, faculty, students, engineering courses, and implemented CSCL learning environment. The thematic analysis was conducted to analyze the transcribed interviews using a qualitative data analysis software (Nvivo). The analysis followed a six step process: (1) reading interview transcripts multiple times and preliminary in vivo codes; (2) conducting open coding by coding interesting or salient features of the data; (3) collecting codes and searching for themes; (4) reviewing themes and creating a thematic map; (5) finalizing themes and their definitions; and (6) compiling findings. This study found that the students’ use of CSCL resources varied depending on the students’ personal preferences, as well as their perceptions of the given resource’s value and its potential to enhance their learning. For example, the dynamics lecturebook, which had been redesigned to encourage problem solving and note-taking, fostered student collaborative problem solving with their peers. In contrast, the professor’s example video solutions had much more of an influence on students’ independent problem-solving processes. The least frequently used resource was the course’s online discussion forum, which could be used as a means of communication. The findings reveal how computer-supported collaborative learning (CSCL) environments enable engineering students to engage in multiple learning opportunities with diverse and flexible resources to both address and to clarify their personal learning needs. This study strongly recommends engineering instructors adapt a CSCL environment for implementation in their own unique classroom context.« less
  2. The DeepLearningEpilepsyDetectionChallenge: design, implementation, andtestofanewcrowd-sourced AIchallengeecosystem Isabell Kiral*, Subhrajit Roy*, Todd Mummert*, Alan Braz*, Jason Tsay, Jianbin Tang, Umar Asif, Thomas Schaffter, Eren Mehmet, The IBM Epilepsy Consortium◊ , Joseph Picone, Iyad Obeid, Bruno De Assis Marques, Stefan Maetschke, Rania Khalaf†, Michal Rosen-Zvi† , Gustavo Stolovitzky† , Mahtab Mirmomeni† , Stefan Harrer† * These authors contributed equally to this work † Corresponding authors: rkhalaf@us.ibm.com, rosen@il.ibm.com, gustavo@us.ibm.com, mahtabm@au1.ibm.com, sharrer@au.ibm.com ◊ Members of the IBM Epilepsy Consortium are listed in the Acknowledgements section J. Picone and I. Obeid are with Temple University, USA. T. Schaffter is with Sage Bionetworks, USA. E. Mehmetmore »is with the University of Illinois at Urbana-Champaign, USA. All other authors are with IBM Research in USA, Israel and Australia. Introduction This decade has seen an ever-growing number of scientific fields benefitting from the advances in machine learning technology and tooling. More recently, this trend reached the medical domain, with applications reaching from cancer diagnosis [1] to the development of brain-machine-interfaces [2]. While Kaggle has pioneered the crowd-sourcing of machine learning challenges to incentivise data scientists from around the world to advance algorithm and model design, the increasing complexity of problem statements demands of participants to be expert data scientists, deeply knowledgeable in at least one other scientific domain, and competent software engineers with access to large compute resources. People who match this description are few and far between, unfortunately leading to a shrinking pool of possible participants and a loss of experts dedicating their time to solving important problems. Participation is even further restricted in the context of any challenge run on confidential use cases or with sensitive data. Recently, we designed and ran a deep learning challenge to crowd-source the development of an automated labelling system for brain recordings, aiming to advance epilepsy research. A focus of this challenge, run internally in IBM, was the development of a platform that lowers the barrier of entry and therefore mitigates the risk of excluding interested parties from participating. The challenge: enabling wide participation With the goal to run a challenge that mobilises the largest possible pool of participants from IBM (global), we designed a use case around previous work in epileptic seizure prediction [3]. In this “Deep Learning Epilepsy Detection Challenge”, participants were asked to develop an automatic labelling system to reduce the time a clinician would need to diagnose patients with epilepsy. Labelled training and blind validation data for the challenge were generously provided by Temple University Hospital (TUH) [4]. TUH also devised a novel scoring metric for the detection of seizures that was used as basis for algorithm evaluation [5]. In order to provide an experience with a low barrier of entry, we designed a generalisable challenge platform under the following principles: 1. No participant should need to have in-depth knowledge of the specific domain. (i.e. no participant should need to be a neuroscientist or epileptologist.) 2. No participant should need to be an expert data scientist. 3. No participant should need more than basic programming knowledge. (i.e. no participant should need to learn how to process fringe data formats and stream data efficiently.) 4. No participant should need to provide their own computing resources. In addition to the above, our platform should further • guide participants through the entire process from sign-up to model submission, • facilitate collaboration, and • provide instant feedback to the participants through data visualisation and intermediate online leaderboards. The platform The architecture of the platform that was designed and developed is shown in Figure 1. The entire system consists of a number of interacting components. (1) A web portal serves as the entry point to challenge participation, providing challenge information, such as timelines and challenge rules, and scientific background. The portal also facilitated the formation of teams and provided participants with an intermediate leaderboard of submitted results and a final leaderboard at the end of the challenge. (2) IBM Watson Studio [6] is the umbrella term for a number of services offered by IBM. Upon creation of a user account through the web portal, an IBM Watson Studio account was automatically created for each participant that allowed users access to IBM's Data Science Experience (DSX), the analytics engine Watson Machine Learning (WML), and IBM's Cloud Object Storage (COS) [7], all of which will be described in more detail in further sections. (3) The user interface and starter kit were hosted on IBM's Data Science Experience platform (DSX) and formed the main component for designing and testing models during the challenge. DSX allows for real-time collaboration on shared notebooks between team members. A starter kit in the form of a Python notebook, supporting the popular deep learning libraries TensorFLow [8] and PyTorch [9], was provided to all teams to guide them through the challenge process. Upon instantiation, the starter kit loaded necessary python libraries and custom functions for the invisible integration with COS and WML. In dedicated spots in the notebook, participants could write custom pre-processing code, machine learning models, and post-processing algorithms. The starter kit provided instant feedback about participants' custom routines through data visualisations. Using the notebook only, teams were able to run the code on WML, making use of a compute cluster of IBM's resources. The starter kit also enabled submission of the final code to a data storage to which only the challenge team had access. (4) Watson Machine Learning provided access to shared compute resources (GPUs). Code was bundled up automatically in the starter kit and deployed to and run on WML. WML in turn had access to shared storage from which it requested recorded data and to which it stored the participant's code and trained models. (5) IBM's Cloud Object Storage held the data for this challenge. Using the starter kit, participants could investigate their results as well as data samples in order to better design custom algorithms. (6) Utility Functions were loaded into the starter kit at instantiation. This set of functions included code to pre-process data into a more common format, to optimise streaming through the use of the NutsFlow and NutsML libraries [10], and to provide seamless access to the all IBM services used. Not captured in the diagram is the final code evaluation, which was conducted in an automated way as soon as code was submitted though the starter kit, minimising the burden on the challenge organising team. Figure 1: High-level architecture of the challenge platform Measuring success The competitive phase of the "Deep Learning Epilepsy Detection Challenge" ran for 6 months. Twenty-five teams, with a total number of 87 scientists and software engineers from 14 global locations participated. All participants made use of the starter kit we provided and ran algorithms on IBM's infrastructure WML. Seven teams persisted until the end of the challenge and submitted final solutions. The best performing solutions reached seizure detection performances which allow to reduce hundred-fold the time eliptologists need to annotate continuous EEG recordings. Thus, we expect the developed algorithms to aid in the diagnosis of epilepsy by significantly shortening manual labelling time. Detailed results are currently in preparation for publication. Equally important to solving the scientific challenge, however, was to understand whether we managed to encourage participation from non-expert data scientists. Figure 2: Primary occupation as reported by challenge participants Out of the 40 participants for whom we have occupational information, 23 reported Data Science or AI as their main job description, 11 reported being a Software Engineer, and 2 people had expertise in Neuroscience. Figure 2 shows that participants had a variety of specialisations, including some that are in no way related to data science, software engineering, or neuroscience. No participant had deep knowledge and experience in data science, software engineering and neuroscience. Conclusion Given the growing complexity of data science problems and increasing dataset sizes, in order to solve these problems, it is imperative to enable collaboration between people with differences in expertise with a focus on inclusiveness and having a low barrier of entry. We designed, implemented, and tested a challenge platform to address exactly this. Using our platform, we ran a deep-learning challenge for epileptic seizure detection. 87 IBM employees from several business units including but not limited to IBM Research with a variety of skills, including sales and design, participated in this highly technical challenge.« less
  3. In this paper we present an evaluation and lessons learned from a joint Research Experience for Undergraduates (REU) and Research Experience for Teachers (RET) program focused on energy and sustainability topics within a Materials Science and Engineering program at a public university. This program brought eleven undergraduate science and engineering students with diverse educational and institutional backgrounds and four local middle and high school teachers on campus for an 8-week research experience working in established lab groups at the university. Using the Qualtrics online survey software, we conducted pre-experience and post-experience surveys of the participants to assess the effects ofmore »participating in this summer research program. At the beginning of the summer, all participants provided their definition of technical research and described what they hoped to get out of their research experience, and the undergraduate students described their future career and educational plans. At the conclusion of the summer, a post-experience survey presented participants’ with their answers from the beginning of the summer and asked them to reflect on how their understanding of research and future plans involving research changed over the course of the summer experience. Many participants evolved a new understanding of research as a result of participating in the summer experience. In particular, they better recognized the collaborative nature of research and the challenges that can arise as part of the process of doing research. Participants acquired both technical and professional skills that they found useful, such as learning new programming languages, becoming proficient at using new pieces of equipment, reviewing technical literature, and improving presentation and communication skills. Undergraduates benefited from developing new relationships with their peers, while the teacher participants benefited from developing relationships with faculty and staff at the university. While most of the participants felt that they were better prepared for future studies or employment, they did not feel like the summer research experience had a significant impact on their future career or degree plans. Finally, while almost all of the participants described their summer research experience as positive, areas for improvement included better planning and access to mentors, as well as more structured activities for the teachers to adapt their research activities for the classroom.« less
  4. In successful collaborative paradigms such as pair programming, students engage in productive dialogue and work to resolve con- flicts as they arise. However, little is known about how elementary students engage in collaborative dialogue for computer science learning. Early findings indicate that these younger students may struggle to manage conflicts that arise during pair programming. To investigate collaborative dialogue that elementary learners use and the conflicts that they encounter, we analyzed videos of twelve pairs of fifth grade students completing pair programming activities. We developed a novel annotation scheme with a focus on collab- orative dialogue and conflicts. We foundmore »that student pairs used best-practice dialogue moves such as self-explanation, question generation, uptake, and praise in less than 23% of their dialogue. High-conflict pairs antagonized their partner, whereas this behav- ior was not observed with low-conflict pairs. We also observed more praise (e.g., “We did it!”) and uptake (e.g., “Yeah and. . . ”) in low-conflict pairs than high-conflict pairs. All pairs exhibited some conflicts about the task, but high-conflict pairs also engaged in conflicts about control of the computer and their partner’s con- tributions. The results presented here provide insights into the collaborative process of young learners in CS problem solving, and also hold implications for educators as we move toward building learning environments that support students in this context.« less
  5. Purpose In response to the evolving COVID-19 pandemic, many universities have transitioned to online instruction. With learning promising to be online, at least in part, for the near future, instructors may be thinking of providing online collaborative learning opportunities to their students who are increasingly isolated from their peers because of social distancing guidelines. This paper aims to provide design recommendations for online collaborative project-based learning exercises based on this research in a software engineering course at the university level. Design/methodology/approach Through joint work between learning scientists, course instructors and software engineering practitioners, instructional design best practices of alignment betweenmore »the context of the learners, the learning objectives, the task and the assessment are actualized in the design of collaborative programming projects for supporting learning. The design, first segments a short real-time collaborative exercise into tasks, each with a problem-solving phase where students participate in collaborative programming, and a reflection phase for reflecting on what they learned in the task. Within these phases, a role-assignment paradigm scaffolds collaboration by assigning groups of four students to four complementary roles that rotate after each task. Findings By aligning each task with granular learning objectives, significant pre- to post-test learning from the exercise as well as each task is observed. Originality/value The roles used in the paradigm discourage divide-and-conquer tendencies often associated with collaborative projects. By requiring students to discuss conflicting ideas to arrive at a consensus implementation, their ideas are made explicit, thus providing opportunities for clarifying misconceptions through discussion and learning from the collaboration.« less