Abstract This study examined putative benefits of testing and production for learning new languages. Undergraduates (N= 156) were exposed to Turkish spoken dialogues under varying learning conditions (retrieval practice, comprehension, verbal repetition) in a computer‐assisted language learning session. Participants completed pre‐ and posttests of number‐ and case‐marking comprehension, a vocabulary test, and an explicit awareness questionnaire. Controlling for nonverbal ability and pretest scores, the retrieval‐practice group performed highest overall. For number/case marking, the comprehension and retrieval‐practice groups outperformed the verbal‐repetition group, suggesting benefits of either recognition‐ or recall‐based testing. For vocabulary, the verbal‐repetition and retrieval‐practice groups outperformed the comprehension group, indicating benefits of overt production. Case marking was easier to learn than number marking, suggesting advantages for learning word‐final inflections. Explicit awareness correlated with comprehension accuracy, yet some participants demonstrated above‐chance comprehension without showing awareness. Findings indicate the value of incorporating both practice tests and overt production in language pedagogy.
more »
« less
Executable Examples for Programming Problem Comprehension
Flawed problem comprehension leads students to produce flawed implementations. However, testing alone is inadequate for checking comprehension: if a student develops both their tests and implementation with the same misunderstanding, running their tests against their implementation will not reveal the issue. As a solution, some pedagogies encourage the creation of input-output examples independent of testing-but seldom provide students with any mechanism to check that their examples are correct and thorough. We propose a mechanism that provides students with instant feedback on their examples, independent of their implementation progress. We assess the impact of such an interface on an introductory programming course and find several positive impacts, some more neutral outcomes, and no identified negative effects.
more »
« less
- PAR ID:
- 10111823
- Date Published:
- Journal Name:
- ICER '19 Proceedings of the 2019 ACM Conference on International Computing Education Research
- Page Range / eLocation ID:
- 131 to 139
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
In this full research paper, we examine various grading policies for second-chance testing. Second-chance testing refers to giving students the opportunity to take a second version of a test for some form of grade replacement. Second-chance testing as a pedagogical strategy bears some similarities to mastery learning, but second-chance testing is less expensive to implement. Previous work has shown that second-chance testing is associated with improved performance, but there is still a lack of clarity regarding the optimal grading policies for this testing strategy. We interviewed seven instructors who use second-chance testing in their courses to collect data on why they chose specific policies. We then conducted structured interviews with some students (N = 11) to capture more nuance about students’ decision making processes under the different grading policies. Afterwards, we conducted a quasi-experimental study to compare two second-chance testing grading policies and determine how they influenced students across multiple dimensions. We varied the grading policies used in two similar sophomore-level engineering courses. We collected assessment data and administered a survey that queried students (N = 513) about their behavior and reactions to both grading policies. Surprisingly, we found that the students’ preference between these two policies were almost perfectly split. We conclude that there are likely many policies that perform well by being simple and encouraging serious attempts on both tests.more » « less
-
null (Ed.)We present in this paper a summary analysis of log files collected during an experiment designed to test the hypothesis that prompting for free self-explanations leads to better comprehension of computer code examples. Indeed, the results indicate that students who were prompted to self-explain while trying to understand code examples performed significantly better at predicting the correct output of the examples than students who were just prompted to read the code examples and predict their output.more » « less
-
The Success and Retention of Students using Multiple-Attempt Testing in Fundamental Engineering Courses: Dynamics and Thermodynamics First name Last name1, First name Last name 1, First name Last name 2, First name Last name 2, First name Last name 3, and First name Last name 1 1Department One 2Department two 3Department Three University Abstract The notion behind Multiple-Attempt Testing continues to be investigated for its benefits in terms of students’ overall success and their retention in fundamental engineering courses. Two engineering courses were delivered in mixed-mode in Spring 2023 (post-COVID): Dynamics and Thermodynamics, whose results were compared to the same courses given in the same semester, four years earlier, delivered in mixed-mode in Spring 2019 (pre-COVID). All four courses were large classes ranging from 167 students for Spring 2023 in Dynamics to 267 students in Thermodynamics for the same Spring 2023 semester. For both courses, there were three tests during the semester. In Spring 2019, students were given a five-day window to conduct their tests in the testing center (TC). Facilitated by the Learning Management System (LMS), the grades were instantly uploaded into CANVAS. Once the test closed, students were allowed to see their work with a teaching-assistant to learn from mistakes and claim some partial credit where possible. However, in Spring 2023, for both courses, students were given three tests during the semester with three attempts each, as well as a make-up final cumulative examination, also with three attempts, for those who wanted to improve their grades. No partial credit was given in any attempt of any test or the final examination. Each attempt was open for two days and the students were allowed to see their tests after each attempt, learn from mistakes, and prepare better for the next attempt. The effectiveness of this testing-interwoven-learning method lies in the fact that students are comfortable and less anxious to do their tests knowing they have other chances, can learn from their mistakes and focus their attention on their weaknesses, enhance their knowledge, and do better in the next attempt. With this self-paced method students learn a lot on their own given the amount of videos provided them. The study shows a substantial decrease in the failure rate, 65% and the overall DWF decreased by more than 40% in both courses. This suggests students aspired to do well in every attempt, or even if they failed all three tests, they would still have a final examination that could save them, which reduced the overall DWF. A survey was also conducted, revealing more than 70% of students preferred this method of testing and learning in future courses.more » « less
-
Most people dislike taking multiple-choice tests, so why are they the default way we evaluate NLP systems? This position paper argues that, despite its simplicity and popularity, multiple-choice evaluation is flawed, both in its format and the datasets it relies on. Drawing from educational testing theory, we propose practical fixes for these issues, helping us build evaluations that better test knowledge and reflect how humans use NLP systems.more » « less
An official website of the United States government

