Abstract A prior study found that mailing prepaid incentives with $5 cash visible from outside the envelope increased the response rate to a mail survey by 4 percentage points compared to cash that was not externally visible. This “visible cash effect” suggests opportunities to improve survey response at little or no cost, but many unknowns remain. Among them: Does the visible cash effect generalize to different survey modes, respondent burdens, and cash amounts? Does it differ between fresh samples and reinterview samples? Does it affect data quality or survey costs? This article examines these questions using two linked studies where incentive visibility was randomized in a large probability sample for the American National Election Studies. The first study used $10 incentives with invitations to a long web questionnaire (median 71 minutes, n = 17,849). Visible cash increased response rates in a fresh sample for both screener and extended interview response (by 6.7 and 4.8 percentage points, respectively). Visible cash did not increase the response rate in a reinterview sample where the baseline reinterview response rate was very high (72 percent). The second study used $5 incentives with invitations to a mail-back paper questionnaire (n = 8,000). Visible cash increased the response rate in a sample of prior nonrespondents by 4.0 percentage points (from 31.5 to 35.5), but it did not increase the response rate in a reinterview sample where the baseline reinterview rate was very high (84 percent). In the two studies, several aspects of data quality were investigated, including speeding, non-differentiation, item nonresponse, nonserious responses, noncredible responses, sample composition, and predictive validity; no adverse effects of visible cash were detected, and sample composition improved marginally. Effects on survey costs were either negligible or resulted in net savings. Accumulated evidence now shows that visible cash can increase incentives’ effectiveness in several circumstances.
more »
« less
Data Exclusion in Policy Survey and Questionnaire Data: Aberrant Responses and Missingness
Data preprocessing is an integral step prior to analyzing data in psychological science, with implications for its potentially guiding policy. This article reports how psychological researchers address data preprocessing or quality concerns, with a focus on aberrant responses and missing data in self-report measures. 240 articles were sampled from four journals: Psychological Science, Journal of Personality and Social Psychology, Developmental Psychology, and Abnormal Psychology from 2012 to 2018. Nearly half of the studies did not report any missing data treatment (111/240; 46.25%), and if they did, the most common approach was listwise deletion (71/240; 29.6%). Studies that remove data due to missingness removed, on average, 12% of the sample. Likewise, most studies do not report any aberrant responses (194/240; 80%), but if they did, they classified 4% of the sample as suspect. Most studies are either not transparent enough about their data preprocessing steps or may be leveraging suboptimal procedures. Recommendations can improve transparency and data quality.
more »
« less
- Award ID(s):
- 1853166
- PAR ID:
- 10401828
- Publisher / Repository:
- SAGE Publications
- Date Published:
- Journal Name:
- Policy Insights from the Behavioral and Brain Sciences
- Volume:
- 10
- Issue:
- 1
- ISSN:
- 2372-7322
- Format(s):
- Medium: X Size: p. 11-17
- Size(s):
- p. 11-17
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Psychological science can benefit from and contribute to emerging approaches from the computing and information sciences driven by the availability of real-world data and advances in sensing and computing. We focus on one such approach, machine-learned computational models (MLCMs)—computer programs learned from data, typically with human supervision. We introduce MLCMs and discuss how they contrast with traditional computational models and assessment in the psychological sciences. Examples of MLCMs from cognitive and affective science, neuroscience, education, organizational psychology, and personality and social psychology are provided. We consider the accuracy and generalizability of MLCM-based measures, cautioning researchers to consider the underlying context and intended use when interpreting their performance. We conclude that in addition to known data privacy and security concerns, the use of MLCMs entails a reconceptualization of fairness, bias, interpretability, and responsible use.more » « less
-
Raw datasets collected for fake news detection usually contain some noise such as missing values. In order to improve the performance of machine learning based fake news detection, a novel data preprocessing method is proposed in this paper to process the missing values. Specifically, we have successfully handled the missing values problem by using data imputation for both categorical and numerical features. For categorical features, we imputed missing values with the most frequent value in the columns. For numerical features, the mean value of the column is used to impute numerical missing values. In addition, TF-IDF vectorization is applied in feature extraction to filter out irrelevant features. Experimental results show that Multi-Layer Perceptron (MLP) classifier with the proposed data preprocessing method outperforms baselines and improves the prediction accuracy by more than 15%.more » « less
-
Research suggests that psychological factors are related to persistence in science, especially for underrepresented students; however, most psychological instruments have been validated through studies conducted at predominately White institutions. In the current study, we report reliability estimates for measures of science identity, science motivation, and science self-efficacy with a sample of undergraduate college students from a Hispanic Serving Institution (N = 309). Internal consistency and test-retest reliability were estimated with Cronbach’s alpha and intra-class correlation coefficients, respectively. We report Cronbach's alpha values separately for male (N = 152), female (N = 152), Hispanic (N = 111), and White (N = 115) students. We also examined whether there were statistically significant differences in the Cronbach’s alpha values between these groups. The results demonstrated good to excellent reliability estimates for internal consistency (α ranged from .89 to .96) and test-retest reliability (ICC ranged from .76 to .80) for all groups. There were no significant differences in Cronbach’s alpha values between students identifying as male versus female or between Hispanic and White identifying students. We conclude by urging science education researchers to examine, report, and interpret reliability estimates for their measures for each dataset.more » « less
-
null (Ed.)Children and adolescents interact in peer groups, which are known to influence a range of psychological and behavioral outcomes. In developmental psychology and related disciplines, social cognitive mapping (SCM), as implemented with the SCM 4.0 software, is the most commonly used method for identifying peer groups from peer report data. However, in a series of four studies, we demonstrate that SCM has an unacceptably high risk of false positives. Specifically, we show that SCM will identify peer groups even when applied to random data. We introduce backbone extraction and community detection as one promising alternative to SCM, and offer several recommendations for researchers seeking to identify peer groups from peer report data.more » « less
An official website of the United States government
