For help with their spelling errors, children often turn to spellcheckers integrated in software applications like word processors and search engines. However, existing spellcheckers are usually tuned to the needs of traditional users (i.e., adults) and generally prove unsatisfactory for children. Motivated by this issue, we introduce KidSpell, an English spellchecker oriented to the spelling needs of children. KidSpell applies (i) an encoding strategy for mapping both misspelled words and spelling suggestions to their phonetic keys and (ii) a selection process that prioritizes candidate spelling suggestions that closely align with the misspelled word based on their respective keys. To assess the effectiveness of, we compare the model’s performance against several popular, mainstream spellcheckers in a number of offline experiments using existing and novel datasets. The results of these experiments show that KidSpell outperforms existing spellcheckers, as it accurately prioritizes relevant spelling corrections when handling misspellings generated by children in both essay writing and online search tasks. As a byproduct of our study, we create two new datasets comprised of spelling errors generated by children from hand-written essays and web search inquiries, which we make available to the research community.
more »
« less
Guiding the selection of child spellchecker suggestions using audio and visual cues
Spellchecking functionality embedded in existing search tools can assist children by offering a list of spelling alternatives when a spelling error is detected. Unfortunately, children tend to generally select the first alternative when presented with a list of options, as opposed to the one that matches their intent. In this paper, we describe a study we conducted with 191 children ages 6-12 in order to offer empirical evidence of: (1) their selection habits when identifying spelling suggestions that match the word they meant to type, and (2) the degree of influence multimodal cues, i.e., synthesized speech and images, have in prompting children to select the correct spelling suggestion. The results from our study reveal that multimodal cues, primarily synthesized speech, have a positive impact on the children's ability to identify their intended word from a list of spelling suggestions.
more »
« less
- Award ID(s):
- 1763649
- PAR ID:
- 10337071
- Date Published:
- Journal Name:
- Proceedings of the Interaction Design and Children Conference
- Page Range / eLocation ID:
- 398-408
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Abstract Psycholinguistic research on children's early language environments has revealed many potential challenges for language acquisition. One is that in many cases, referents of linguistic expressions are hard to identify without prior knowledge of the language. Likewise, the speech signal itself varies substantially in clarity, with some productions being very clear, and others being phonetically reduced, even to the point of uninterpretability. In this study, we sought to better characterize the language‐learning environment of American English‐learning toddlers by testing how well phonetic clarity and referential clarity align in infant‐directed speech. Using an existing Human Simulation Paradigm (HSP) corpus with referential transparency measurements and adding new measures of phonetic clarity, we found that the phonetic clarity of words’ first mentions significantly predicted referential clarity (how easy it was to guess the intended referent from visual information alone) at that moment. Thus, when parents’ speech was especially clear, the referential semantics were also clearer. This suggests that young children could use the phonetics of speech to identify globally valuable instances that support better referential hypotheses, by homing in on clearer instances and filtering out less‐clear ones. Such multimodal “gems” offer special opportunities for early word learning. Research HighlightsIn parent‐infant interaction, parents’ referential intentions are sometimes clear and sometimes unclear; likewise, parents’ pronunciation is sometimes clear and sometimes quite difficult to understand.We find that clearer referential instances go along with clearer phonetic instances, more so than expected by chance.Thus, there are globally valuable instances (“gems”) from which children could learn about words’ pronunciations and words’ meanings at the same time.Homing in on clear phonetic instances and filtering out less‐clear ones would help children identify these multimodal “gems” during word learning.more » « less
-
Abstract What is vision's role in driving early word production? To answer this, we assessed parent‐report vocabulary questionnaires administered to congenitally blind children (N = 40, Mean age = 24 months [R: 7–57 months]) and compared the size and contents of their productive vocabulary to those of a large normative sample of sighted children (N = 6574). We found that on average, blind children showed a roughly half‐year vocabulary delay relative to sighted children, amid considerable variability. However, the content of blind and sighted children's vocabulary was statistically indistinguishable in word length, part of speech, semantic category, concreteness, interactiveness, and perceptual modality. At a finer‐grained level, we also found that words’ perceptual properties intersect with children's perceptual abilities. Our findings suggest that while an absence of visual input may initially make vocabulary development more difficult, the content of the early productive vocabulary is largely resilient to differences in perceptual access. Research HighlightsInfants and toddlers born blind (with no other diagnoses) show a 7.5 month productive vocabulary delay on average, with wide variability.Across the studied age range (7–57 months), vocabulary delays widened with age.Blind and sighted children's early vocabularies contain similar distributions of word lengths, parts of speech, semantic categories, and perceptual modalities.Blind children (but not sighted children) were more likely to say visual words which could also be experienced through other senses.more » « less
-
Misspellings in queries used to initiate online searches is an everyday occurrence. When this happens, users either rely on the search engine's ability to understand their query or they turn to spellcheckers. Spellcheckers are usually based on popular dictionaries or past query logs, leading to spelling suggestions that often better resonate with adult users because that data is more readily available. Based on an educational perspective, previous research reports, and initial analyses of sample search logs, we hypothesize that existing spellcheckers are not suitable for young users who frequently encounter spelling challenges when searching for information online. We present early results of our ongoing research focused on identifying the needs and expectations children have regarding spellcheckers.more » « less
-
Abstract Previous research has shown that when domain‐general transitional probability (TP) cues to word segmentation are in conflict with language‐specific stress cues, English‐learning 5‐ and 7‐month‐olds rely on TP, whereas 9‐month‐olds rely on stress. In two artificial languages, we evaluated English‐learning infants’ sensitivity to TP cues to word segmentation vis‐a‐vis language‐specific vowel phonotactic (VP) cues—English words do not end in lax vowels. These cues were either consistent or conflicting. When these cues were in conflict, 10‐month‐olds relied on the VP cues, whereas 5‐month‐olds relied on TP. These findings align with statistical bootstrapping accounts, where infants initially use domain‐general distributional information for word segmentation, and subsequently discover language‐specific patterns based on segmented words. Research HighlightsResearch indicates that when transitional probability (TP) conflicts with stress cues for word segmentation, English‐learning 9‐month‐olds rely on stress, whereas younger infants rely on TP.In two artificial languages, we evaluated English‐learning infants’ sensitivity to TP versus vowel phonotactic (VP) cues for word segmentation.When these cues conflicted, 10‐month‐olds relied on VPs, whereas 5‐month‐olds relied on TP.These findings align with statistical bootstrapping accounts, where infants first utilize domain‐general distributional information for word segmentation, and then identify language‐specific patterns from segmented words.more » « less
An official website of the United States government

