skip to main content


Title: Curating for Accessibility
Accessibility of research data to disabled users has received scant attention in literature and practice. In this paper we briefly survey the current state of accessibility for research data and suggest some first steps that repositories should take to make their holdings more accessible. We then describe in depth how those steps were implemented at the Qualitative Data Repository (QDR), a domain repository for qualitative social-science data. The paper discusses accessibility testing and improvements on the repository and its underlying software, changes to the curation process to improve accessibility, as well as efforts to retroactively improve the accessibility of existing collections. We conclude by describing key lessons learned during this process as well as next steps.  more » « less
Award ID(s):
1823950 2116935
NSF-PAR ID:
10403706
Author(s) / Creator(s):
; ; ;
Date Published:
Journal Name:
International Journal of Digital Curation
Volume:
17
Issue:
1
ISSN:
1746-8256
Page Range / eLocation ID:
10
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. null (Ed.)
    This paper describes the development of services and tools for scaling data curation services at the Qualitative Data Repository (QDR). Through a set of open-source tools, semi-automated workflows, and extensions to the Dataverse platform, our team has built services for curators to efficiently and effectively publish collections of qualitatively derived data. The contributions we seek to make in this paper are as follows: 1. We describe ‘human-in-the-loop’ curation and the tools that facilitate this model at QDR; 2. We provide an in-depth discussion of the design and implementation of these tools, including applications specific to the Dataverse software repository, as well as standalone archiving tools written in R; and 3. We highlight the role of providing a service layer for data discovery and accessibility of qualitative data. 
    more » « less
  2. How can authors using many individual pieces of qualitative data throughout a publication make their research transparent? In this paper we introduce Annotation for Transparent Inquiry (ATI), an approach to enhance transparency in qualitative research. ATI allows authors to connect specific passages in their publication with an annotation. These annotations provide additional information relevant to the passage and, when possible, include a link to one or more data sources underlying a claim; data sources are housed in a repository. After describing ATI’s conceptual and technological implementation, we report on its evaluation through a series of workshops conducted by the Qualitative Data Repository (QDR) and present initial results of the evaluation. The article ends with an outlook on next steps for the project. 
    more » « less
  3. Abstract. As cloud-based web services get more and more capable, available, and powerful (CAP), data science and engineering is pulled toward the frontline because DATA means almost anything-as-a-service (XaaS) via Digital Archiving and Transformed Analytics. In general, a web service (via a website) serves customers with web documents in HTML, JSON, XML, and multimedia via interactive (request) and responsive (reply) ways for specific domain problem solving over the Internet. In particular, a web service is deeply involved with UI & UX (user interface and user experience) plus considerate regulations on QoS (Quality of Service) as well, which refers to both information synthesis and security, namely availability and reliability for providential web services. This paper, based on the novel wiseCIO as a Platform-as-a-Service (PaaS), presents digital archiving 3 and transformed analytics (DATA) via machine learning, one of the most practical aspects of artificial intelligence. Machine learning is the science of data analysis that automates analytical model building and online analytical processing (OLAP) that enables computers to act without being explicitly programmed through CTMP. Computational thinking combined with manageable processing is 4 thoroughly discussed and utilized for FAST solutions in a feasible, analytical, scalable and testable approach. DATA is central to information synthesis and analytics (ISA), and digitized archives plays a key role in transformed analytics on intelligence for business, education and entertainment (iBEE). Case studies as applicable examples are discussed over broad fields where archival digitization is required for analytical transformation via machine learning, such as scalable ARM (archival repository for manageable accessibility), visual BUS (biological understanding from STEM), schooling DIGIA (digital intelligence governing instruction and administering), viewable HARP (historical archives & religious preachings), vivid MATH (mathematical apps in teaching and hands-on exercise), and SHARE (studies via hands-on assignment, revision and evaluation). As a result, wiseCIO promotes DATA service by providing ubiquitous web services of analytical processing via universal interface and user-centric experience in favor of logical organization of web content and relational information groupings that are vital steps in the ability of an archivist or librarian to recommend and retrieve information for a researcher. More important, wiseCIO also plays a key role as a content management system and delivery platform with capacity of hosting 10,000+ traditional web pages with great ease. 
    more » « less
  4. Background: While a primary goal of education research is discovering and disseminating scholarly knowledge, traditional dissemination alone is insufficient to foster sustainable educational change. Journals and conferences target a particular audience invested in a specific topic; it is often not practical for stakeholders to engage with research outside of their specific areas of specialization. Thus, the research-to-practice gap continually widens as education research findings fail to influence wider audiences. Purpose: In this paper, we highlight audio as a promising alternative format for dissemination. Audio dissemination has the potential to multiply the impacts of qualitative research by disseminating findings with more immediacy and accessibility than traditional research publications. Approach: We summarize one specific audio narrative dissemination approach conducted as part of the pilot phase of the Audio for Inclusion Project, a recent National Science Foundation-funded project to foster inclusive understandings for engineering faculty. We organize the discussion around orienting goals and challenges encountered, as well as lessons learned and suggestions for future improvements. Findings: Lessons learned for audio narrative dissemination include paying close attention to creating a coherent and cohesive narrative by removing distracting details, and aligning student actors with participants so that the tone, affect, and emphasis remain true to the participant. Implications: This paper presents new possibilities for qualitative researchers on broadening participation, to repurpose their interview content to form practical resources and training that can improve faculty’s knowledge, empathy, and understanding of students’ diverse and marginalized backgrounds. Additionally, these findings will be useful for all researchers seeking new methods of translating research findings into actionable impact. 
    more » « less
  5. Maintaining data quality is a fundamental requirement for any successful and long-term data management. Providing high-quality, reliable, and statistically sound data is a primary goal for clinical research informatics. In addition, effective data governance and management are essential to ensuring accurate data counts, reports, and validation. As a crucial step of the clinical research process, it is important to establish and maintain organization-wide standards for data quality management to ensure consistency across all systems designed primarily for cohort identification, allowing users to perform an enterprise-wide search on a clinical research data repository to determine the existence of a set of patients meeting certain inclusion or exclusion criteria. Some of the clinical research tools are referred to as de-identified data tools. Assessing and improving the quality of data used by clinical research informatics tools are both important and difficult tasks. For an increasing number of users who rely on information as one of their most important assets, enforcing high data quality levels represents a strategic investment to preserve the value of the data. In clinical research informatics, better data quality translates into better research results and better patient care. However, achieving high-quality data standards is a major task because of the variety of ways that errors might be introduced in a system and the difficulty of correcting them systematically. Problems with data quality tend to fall into two categories. The first category is related to inconsistency among data resources such as format, syntax, and semantic inconsistencies. The second category is related to poor ETL and data mapping processes. In this paper, we describe a real-life case study on assessing and improving the data quality at one of healthcare organizations. This paper compares between the results obtained from two de-identified data systems i2b2, and Epic Slicedicer, and discuss the data quality dimensions' specific to the clinical research informatics context, and the possible data quality issues between the de-identified systems. This work in paper aims to propose steps/rules for maintaining the data quality among different systems to help data managers, information systems teams, and informaticists at any health care organization to monitor and sustain data quality as part of their business intelligence, data governance, and data democratization processes. 
    more » « less