skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Formative assessment through think alouds
We describe how teachers used a formative assessment approach called whole class think alouds, which may happen at any time during everyday instruction and foster evidence-based instructional practices. They allow students to collaborate and orally communicate their problem solving with a goal of assessing to promote learning.  more » « less
Award ID(s):
1720646
PAR ID:
10252691
Author(s) / Creator(s):
;
Editor(s):
Barlow, A.
Date Published:
Journal Name:
Mathematics teaching and learning
Volume:
114
Issue:
8
ISSN:
2589-6016
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. With the rapid proliferation of small unmanned aircraft systems (UAS), the risk of mid-air collisions is growing, as is the risk associated with the malicious use of these systems. Airborne Detect-and-Avoid (ABDAA) and counter-UAS technologies have similar sensing requirements to detect and track airborne threats, albeit for different purposes: to avoid a collision or to neutralize a threat, respectively. These systems typically include a variety of sensors, such as electro-optical or infrared (EO/IR) cameras, RADAR, or LiDAR, and they fuse the data from these sensors to detect and track a given threat and to predict its trajectory. Camera imagery can be an effective method for detection as well as for pose estimation and threat classification, though a single camera cannot resolve range to a threat without additional information, such as knowledge of the threat geometry. To support ABDAA and counter-UAS applications, we consider a merger of two image-based sensing methods that mimic human vision: (1) a "peripheral vision" camera (i.e., with a fisheye lens) to provide a large field-of-view and (2) a "central vision" camera (i.e., with a perspective lens) to provide high resolution imagery of a specific target. Beyond the complementary ability of the two cameras to support detection and classification, the pair form a heterogeneous stereo vision system that can support range resolution. This paper describes the initial development and testing of a peripheral-central vision system to detect, localize, and classify an airborne threat and finally to predict its path using knowledge of the threat class. 
    more » « less
  2. null (Ed.)
    Like many natural sciences, a critical component of archaeology is field work. Despite its importance, field opportunities are available to few students for financial and logistical reasons. With little exposure to archaeological research, fewer students are entering archaeology, particularly minority students (Smith 2004; Wilson 2015). To counter these trends, we have leveraged the ongoing revolution in consumer electronics for the current, digitally-empowered generation by creating a game-based, virtual archaeology curriculum to 1) teach foundational principles of a discipline that is challenging to present in a traditional classroom by using sensory and cognitive immersion; and, 2) allow wider access to a field science that has previously been limited to only select students. Virtual reality (VR) is computer technology that creates a simulated three-dimensional world for a user to experience in a bodily way, thereby transforming data analysis into a sensory and cognitive experience. Using a widely-available, room-scale, VR platform, we have created a virtual archaeological excavation experience that conveys two overarching classroom objectives: 1) teach the physical methods of archaeological excavation by providing the setting and tools for a student to actively engage in field work; and, 2) teach archaeological concepts using a scientific approach to problem solving by couching them within a role-playing game. The current prototype was developed with the HTC Vive VR platform, which includes a headset, hand controllers, and two base stations to track the position and orientation of the user’s head and hands within a 4x4 meter area. Environments were developed using Unreal Engine 4, an open source gaming engine, to maximize usability for different audiences, learning objectives, and skill levels. Given the inherent fun of games and widespread interest in archaeology and cultural heritage, the results of this research are adaptable and applicable to learners of all ages in formal and informal educational settings. 
    more » « less
  3. Assistive robot manipulators must be able to autonomously pick and place a wide range of novel objects to be truly useful. However, current assistive robots lack this capability. Additionally, assistive systems need to have an interface that is easy to learn, to use, and to understand. This paper takes a step forward in this direction. We present a robot system comprised of a robotic arm and a mobility scooter that provides both pick-and-drop and pick-and-place functionality for open world environments without modeling the objects or environment. The system uses a laser pointer to directly select an object in the world, with feedback to the user via projecting an interface into the world. Our evaluation over several experimental scenarios shows a significant improvement in both runtime and grasp success rate relative to a baseline from the literature, and furthermore demonstrates accurate pick and place capabilities for tabletop scenarios. 
    more » « less
  4. The ability of quadrupedal robots to follow commanded velocities is important for navigating in constrained environments such as homes and warehouses. This paper presents a simple, scalable approach to realize high fidelity speed regulation and demonstrates its efficacy on a quadrupedal robot. Using analytical inverse kinematics and gravity compensation, a task-level controller calculates joint torques based on the prescribed motion of the torso. Due to filtering and feedback gains in this controller, there is an error in tracking the velocity. To ensure scalability, these errors are corrected at the time scale of a step using a Poincar´e map (a mapping of states and control between consecutive steps). A data-driven approach is used to identify a decoupled Poincar´e map, and to correct for the tracking error in simulation. However, due to model imperfections, the simulation-derived Poincar´e map-based controller leads to tracking errors on hardware. Three modeling approaches – a polynomial, a Gaussian process, and a neural network – are used to identify a correction to the simulation-based Poincar´e map and to reduce the tracking error on hardware. The advantages of our approach are the computational simplicity of the task-level controller (uses analytical computations and avoids numerical searches) and scalability of the sim-to-real transfer (use of low-dimensional Poincar´e map for sim-to-real transfer). A video is in this shortened link: http://tiny.cc/humanoids23 
    more » « less
  5. Interfacing between robots and humans has come a long way in the past few years, and new methods for smart, robust interaction are needed. Typically, a technician has to program a routine for a robot in order for the robot to be useful. This puts up a significant barrier to entry into the field of automating tasks using robots—not only is a technician and a computer required, but the robot is not adaptive to the immediate needs of the user. The robot is only capable of executing a pre-determined task and for any change to be made the entire system needs to be paused. This project seeks to bridge the gap between user and robot interface, creating an easy-to-use system that allows for adaptive robot control. Using a combination of computer vision and a monocular camera system and integrated LiDAR sensor on an iPhone, gesture recognition and pose estimation was conducted within an independent system to control the Baxter humanoid robot. The gathered data was sent wirelessly to the robot to be interpreted and then replay actions performed by the user. 
    more » « less