skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: THE LINGUISTIC AND THE SOCIAL INTERTWINED: LINGUISTIC CONVERGENCE TOWARD SOUTHERN SPEECH
Award ID(s):
1917900
PAR ID:
10161663
Author(s) / Creator(s):
Date Published:
Journal Name:
Dissertation
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. This article examines linguistic variation in relation to the critical social institution and social domain of education, with an emphasis on linguistic inclusion, focusing on the United States. Education is imbued with power dynamics, and language often serves as a gatekeeping mechanism for students from minoritized backgrounds, which helps create, sustain, and perpetuate educational inequalities. Grounded in this context, the article reviews intersecting factors related to linguistic variation that affect student academic performance. Empirical and applied models of effective partnerships among researchers, educators, and students are presented, which provide road maps to advance linguistic inclusion in schools within the broader social movement for equity in education. 
    more » « less
  2. The wider genealogical affiliations of the Timor-Alor-Pantar languages have been the subject of much speculation. These languages are surrounded by unrelated Austronesian languages, and attempts to locate related languages have focused on Papuan languages 800 km or more distant. This chapter draws on typological, pronominal, and especially lexical evidence to examine three hypotheses regarding the higher-level affiliations of the Timor-Alor-Pantar languages: (1) the languages are related to the North Halmaheran (West Papuan) languages; (2) the languages are part of the Trans New Guinea family; and (3) the languages are related to the West Bomberai family, with no link to Trans New-Guinea more broadly. We rely in particular on recent reconstructions of proto-Timor-Alor-Pantar vocabulary (chapter 3). Of the hypotheses evaluated here, we find the most striking similarities between TAP and the West Bomberai family. However, we conclude that the evidence currently available is insufficient to confirm a genealogical relationship with West Bomberai or any other family, and hence, TAP must be considered a family-level isolate. 
    more » « less
  3. Linguistic analysis is a core task in the process of documenting, analyzing, and describing endangered and less-studied languages. In addition to providing insight into the properties of the language being studied, having tools to automatically label words in a language for grammatical category and morphological features can support a range of applications useful for language pedagogy and revitalization. At the same time, most modern NLP methods for these tasks require both large amounts of data in the language and compute costs well beyond the capacity of most research groups and language communities. In this paper, we present a gloss-to-gloss (g2g) model for linguistic analysis (specifically, morphological analysis and part-of-speech tagging) that is lightweight in terms of both data requirements and computational expense. The model is designed for the interlinear glossed text (IGT) format, in which we expect the source text of a sentence in a low-resource language, a translation of that sentence into a language of wider communication, and a detailed glossing of the morphological properties of each word in the sentence. We first produce silver standard parallel glossed data by automatically labeling the high-resource translation. The model then learns to transform source language morphological labels into output labels for the target language, mediated by a structured linguistic representation layer. We test the model on both low-resource and high-resource languages, and find that our simple CNN-based model achieves comparable performance to a state-of-the-art transformer-based model, at a fraction of the computational cost. 
    more » « less