[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

Published on in Vol 26 (2024)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/57615, first published .
Data Quality–Driven Improvement in Health Care: Systematic Literature Review

Data Quality–Driven Improvement in Health Care: Systematic Literature Review

Data Quality–Driven Improvement in Health Care: Systematic Literature Review

Review

1Clinical Outcomes and Data Unit, The Christie NHS Foundation Trust, Manchester, United Kingdom

2Radiotherapy Related Research Group, University of Manchester, Manchester, United Kingdom

Corresponding Author:

Anthony Lighterness, BBiomedSc, MPhil, MSc

Clinical Outcomes and Data Unit

The Christie NHS Foundation Trust

550 Wilmslow Road

Manchester, M20 4BX

United Kingdom

Phone: 44 7305054646

Email: anthony.lighterness@nhs.net


Background: The promise of real-world evidence and the learning health care system primarily depends on access to high-quality data. Despite widespread awareness of the prevalence and potential impacts of poor data quality (DQ), best practices for its assessment and improvement are unknown.

Objective: This review aims to investigate how existing research studies define, assess, and improve the quality of structured real-world health care data.

Methods: A systematic literature search of studies in the English language was implemented in the Embase and PubMed databases to select studies that specifically aimed to measure and improve the quality of structured real-world data within any clinical setting. The time frame for the analysis was from January 1945 to June 2023. We standardized DQ concepts according to the Data Management Association (DAMA) DQ framework to enable comparison between studies. After screening and filtering by 2 independent authors, we identified 39 relevant articles reporting DQ improvement initiatives.

Results: The studies were characterized by considerable heterogeneity in settings and approaches to DQ assessment and improvement. Affiliated institutions were from 18 different countries and 18 different health domains. DQ assessment methods were largely manual and targeted completeness and 1 other DQ dimension. Use of DQ frameworks was limited to the Weiskopf and Weng (3/6, 50%) or Kahn harmonized model (3/6, 50%). Use of standardized methodologies to design and implement quality improvement was lacking, but mainly included plan-do-study-act (PDSA) or define-measure-analyze-improve-control (DMAIC) cycles. Most studies reported DQ improvements using multiple interventions, which included either DQ reporting and personalized feedback (24/39, 61%), IT-related solutions (21/39, 54%), training (17/39, 44%), improvements in workflows (5/39, 13%), or data cleaning (3/39, 8%). Most studies reported improvements in DQ through a combination of these interventions. Statistical methods were used to determine significance of treatment effect (22/39, 56% times), but only 1 study implemented a randomized controlled study design. Variability in study designs, approaches to delivering interventions, and reporting DQ changes hindered a robust meta-analysis of treatment effects.

Conclusions: There is an urgent need for standardized guidelines in DQ improvement research to enable comparison and effective synthesis of lessons learned. Frameworks such as PDSA learning cycles and the DAMA DQ framework can facilitate this unmet need. In addition, DQ improvement studies can also benefit from prioritizing root cause analysis of DQ issues to ensure the most appropriate intervention is implemented, thereby ensuring long-term, sustainable improvement. Despite the rise in DQ improvement studies in the last decade, significant heterogeneity in methodologies and reporting remains a challenge. Adopting standardized frameworks for DQ assessment, analysis, and improvement can enhance the effectiveness, comparability, and generalizability of DQ improvement initiatives.

J Med Internet Res 2024;26:e57615

doi:10.2196/57615

Keywords



Background

The landscape of health care, improvement science, and digital technologies increasingly hinges on real-world data (RWD) to improve patient care and outcomes [1,2]. RWD encompasses a vast and dynamic collection of health-related information generated by means of routine clinical care from a diverse range of sources, such as electronic health records (EHRs), electronic medical records (EMRs), hospital information systems (HISs), picture archiving and communication systems (PACSs), national registries, claims data, and wearable devices [2-4]. Despite their long history, the adoption and use of EHRs have become widespread only during the last decade [4,5]. EHRs and EMRs are often used interchangeably in the literature discussing health-related RWD, where some suggest that EMRs are a subset of EHRs [6], but the prominence of EHRs have positioned them as a primary source of RWD due to the comprehensive spectrum of patient information, including genetic testing, treatment modalities, and clinical outcomes [7,8]. To reflect this primary focus, the term EHR-RWD will be used throughout this review to denote RWD derived from EHRs.

Real-world evidence (RWE) generated from EHR-RWD holds unprecedented potential to bridge the unmet gaps that exist between controlled clinical trial studies and the complexities of health care delivery in the real world [1,2,7,9]. While randomized control trial studies remain the “gold standard” for assessing the efficacy of new interventions, the essence of RWE lies in its potential to reflect the diversity, heterogeneity, and nuances of patient populations and care settings, thus enabling a more holistic understanding of health outcomes and interventions [7,10]. Data and RWE studies can support the life cycle of drug development, clinical and regulatory decision-making, and health technology assessment [3,8,11]. Moreover, RWE underpins the vision of the learning health care system (LHS), which is a paradigm built upon the cycle of continuous learning to achieve personalized medicine [12]. The transformative potential of RWE, however, hinges on a pivotal caveat—high-quality data.

Despite its potential, EHR-RWD, and by extension RWE, grapple with formidable barriers, and chief among these is data quality (DQ) [1,2,9,10,13,14]. The need for high-quality data was exemplified during the COVID-19 pandemic, when EHR-RWD was critical for research and planning [15]. Regulatory authorities, such as the Food and Drug Administration (FDA), European Medicines Agency (EMA), National Institute for Health and Care Excellence (NICE), and Medicines and Healthcare Products Regulatory Agency (MHRA), recommend the reporting of DQ metrics and dimensions to provide additional context to real-world study outcomes, thus serving as the foundation for trustworthy RWE [3,14,16-19]. These guidelines mostly promote ad hoc DQ assessment and reporting, with the exception that EMA briefly notes the importance of assessing DQ as close as possible to the moment of data capture to help with collection errors [3,16-18]. While ad hoc DQ measurement and reporting can increase transparency and awareness of the limitations of real-world study outcomes, understanding the causes of poor data capture is needed for long-term, sustainable improvement in DQ and, subsequently, the impact of RWE to support decision-making in health care.

In this study, we aimed to evaluate the robustness of studies seeking to improve the quality of structured EHR-derived data. Incorporating quality improvement best practices, we aimed to assess how studies measure DQ, identify which interventions are implemented, and summarize the outcomes to understand which interventions are useful to improve DQ.

DQ Theory

The quality of data describes if the data meet the expectations of a data consumer and, therefore, if they are fit for purpose [20]. This expected behavior can be documented and understood using metadata, that is, additional data that provide meaning and context by describing how the data they are associated with should have been captured, defined, structured, and represented [20]. Metadata, in turn, can inform the design of quantifiable metrics that measure the compliance of data against a set of relevant business rules and constraints during DQ assessment [21]. DQ metrics, which can act as the method by which to measure the respective DQ dimensions, can serve a crucial purpose in understanding areas needing improvement.

Scholars have proposed various multidimensional frameworks for comprehensive DQ assessment. While there may be disagreements on semantic choices and definitions across these frameworks, certain DQ concepts are consistently studied and well represented in the literature, popular frameworks, and DQ profiling software [22]. Despite the challenge, some common concepts can be mapped across frameworks, as exemplified by the 6 core dimensions defined by the Data Management Association (DAMA): completeness, validity, consistency, uniqueness, timeliness, and accuracy (Textbox 1) [20].

Textbox 1 outlines the 6 fundamental DQ dimensions defined by DAMA to enable standardization and comparability of DQ concepts defined, assessed, and improved in the included DQ improvement literature. These dimensions are essential for assessing and improving the quality of RWD.

In an exploration of the DQ literature, we found 10 other reviews that summarized the most frequently represented DQ dimensions across DQ studies, software, and theoretical frameworks [19,21,23-30]. All 10 reviews demonstrated the concept of completeness to be well represented. Of these, 4 reviews agreed that data accuracy and consistency were popular DQ concepts [23,24,26,28]. Timeliness and validity were said to be popular by 3 different reviews, and uniqueness was only raised in the study by Gordon et al [29].

Textbox 1. Core dimensions of data quality as defined by the Data Management Association.

Data quality dimension and description

  • Completeness: the presence of the expected data
  • Uniqueness: uniqueness of records where duplication is not expected
  • Timeliness: a measure of data freshness
  • Consistency: a check of consistency between multiple sources of the same data elements
  • Validity: the validity of data against data standards or plausible values, ranges, or patterns
  • Accuracy: a check of consistency of source data against a referencegold standard

Completeness, also known as missingness, is often reported to be popular among DQ frameworks, tooling, and studies [22-27,29,30]. In general, it refers to the degree to which all required or expected data values or records are present [20]. The most common method to measure completeness consists of counting records with blank, unknown, empty, “NULL,” or “NaN” values, though variations may include a measurement of data availability [22,25,26]. Reviews by Weiskopf and Weng [22] and Syed et al [28] also found variations of data completeness assessment involving triangulation of multiple sources to create a gold standard. However, this approach risks assuming the accuracy of available data. Previous work demonstrate that missing data can lower statistical power of research outcomes and lead to biased assumptions with improper use of imputation methods [10,31,32].

Data accuracy is also well studied in DQ literature [22,25,28]. The accuracy dimension measures the extent to which data reflect the truth of events and conform to their actual value [20,22,25,28]. Other terms used to describe accuracy include error, correctness, integrity, trustworthiness, reliability, and validity [22,28]. The most common method to assess accuracy in health care involves the comparison of EHR data to a reference gold standard, which may include paper records, manual data reviews, triangulation of data from multiple sources, or interviews with patients [22]. Measurement of data accuracy can identify issues such as lack of specificity or precision [33]. Previous work found that code precision can be related to staff training or use of multiple EHR systems [33-36]. As Cook et al [37] noted in a review of DQ issues affecting social determinants data, imprecise codified data may affect minority groups disproportionately, which in turn may affect secondary research outcomes.

Various terms and definitions for validity exist in the literature [22,28]. It generally describes the conformance of data to expected value ranges, patterns, formats, general medical knowledge, or data standards as set by local or external authorities [22,26,28,38]. Validity is also termed plausibility, conformance, and integrity and can also be separated into internal and external validity or be incorporated with other data elements such as temporal validity [22,38]. As EHR-derived data contain large volumes of categorical data, such as patient demographic, diagnostic, and treatment-related information, validity constraints are needed to identify areas needing standardization [39,40]. Data standardization has been shown to correlate positively to data sharing capabilities and emergency care [41]. However, improper design of standardized data entry user interfaces, such as the use of excessively long drop-down lists for diagnostic codes, can also increase cognitive demand, lower workflow efficiency, and correlate to clinician burnout [42].

Timeliness refers to several time-related characteristics of data and is, therefore, also termed currency, recency, or freshness [22,38]. For example, time-related data items can measure how closely the recorded information corresponds to the actual event. Factors affecting timely capture of EHR data include workflow inefficiencies, documentation burden, limited access to hardware, and interruptions [43-48]. Batch processing of data long after the event may indicate a lack of timeliness and affects other DQ dimensions such as completeness, accuracy, and validity [28,44,46,48,49].

Consistency, otherwise known as concordance, describes the agreement of similar data elements between multiple sources [20,22]. The existence of multiple data capture systems and RWD sources can give rise to inconsistent data for a given patient, and in the absence of a defined gold standard, the consistency dimension can identify potentially erroneous data [10,50]. Botsis et al [36] identified multiple inconsistencies during a DQ analysis of a cohort of patients with pancreatic cancer stored in the Columbia University Medical Center’s EHR data warehouse. These included pancreatitis recorded as chronic in pathology reports but acute in clinical notes and patients with diabetes receiving both codes for type 1 and type 2 in the same EHR source [36]. von Lucadou et al [51] found similar discrepancies when comparing data items between different systems, adding that inconsistencies may be caused by individual documentation habits. Measurement of data consistency highlights potential duplication and redundancy between different EHR sources and within the same EHR system and can thus help improve data capture or data engineering workflows.

The uniqueness dimension identifies where duplication of objects, events, or values are not expected [10,26]. Duplication of patient EHRs frequently occur when disparate data flows that contain overlapping objects are combined [10,38]. Similar to consistency, the uniqueness dimension can identify and resolve redundant and inefficient workflows and processes [52]. This is particularly relevant given that 60% to 90% of clinicians routinely copy and paste data between systems [53]. The “copy and paste” phenomenon is pervasive in health care and is known to promote inconsistencies, propagate errors, and contribute to documentation burden and clinician fatigue [42,48,53]. As such, the uniqueness dimension is related to consistency and accuracy.

DQ Tooling

DQ measurement involves the process by which defective values are identified and labeled through the application of business rules or automated tooling [21]. The subsequent analysis of DQ results can then be aggregated, analyzed, and summarized, providing key insights for improvement. Tools to support these activities are widely reported and well studied [21,27,29,30,54,55].

While the availability of DQ tools is abundant, the literature reveals a considerable gap in the effective support for DQ improvement efforts, particularly in the realm of health care. Evaluations of DQ profiling software in the studies by Ehrlinger and Wöß [30], Gordon et al [29], and Ozonze et al [21] highlight limitations in the range of DQ metrics offered for assessment, interoperability issues, and complex configuration requirements.

Root cause analysis, a pivotal aspect of DQ management, is also notably lacking in demonstrations within this landscape. Eden et al [48] demonstrated the utility of the Odigos framework in qualitative root cause analysis, which classifies DQ issues that emanate from the material world, such as digital infrastructure or access to hardware; personal world, that is, staff behaviors; and societal world, that is, job roles and social norms. The legal and technical implications associated with data cleaning, as opposed to addressing the root causes of poor data, underscore potential risks to patient safety [56-58]. Consequently, the overarching trend in DQ tool development leans toward prioritizing technical features, leaving a noticeable gap in the demonstration of their utility in the prevention and improvement of poor-quality data capture in real-world health settings.

Quality Improvement

Quality improvement describes the use of systematic continuous approaches to create positive changes in an area of need [59]. Various structured, iterative learning frameworks, such as plan-do-study-act (PDSA), total data quality management (TDQM), define-measure-analyze-improve-control (DMAIC), and the LHS, exist [12,60-63]. Lacking a universally agreed-upon model, each methodology focuses on enhancing different areas, ranging from service evaluation to treatment standards [59].

Developed from the earlier plan-do-check-act cycle by Deming [64], the PDSA cycle enhances the traditional model by prioritizing the “study” stage—a deeper analysis rather than a simple check [65]. This adaptation roots the PDSA cycle firmly in the scientific method, encouraging a disciplined approach to testing and monitoring changes over time [65]. Its flexible and qualitative nature makes it particularly suitable for health care settings where adaptability to complex and variable processes is crucial.

Proposed by Wang [61] in the late 1990s, TDQM adapts traditional total quality management principles specifically to data management, highlighting the importance of data as a key asset or product. In health care, where decision-making increasingly relies on accurate and timely data, TDQM offers a robust framework to ensure the integrity and usability of data. This focus on DQ management is critical as health care systems integrate more digital processes and data-driven decision-making frameworks. TDQM adapts the PDSA planning stage to specifically target the improvement of DQ [61].

A product of Motorola engineers in the 1980s, DMAIC provides a structured, data-driven quality improvement methodology [61]. Unlike the more qualitative PDSA, DMAIC emphasizes quantifiable metrics and statistical analysis to identify and mitigate variations in processes. This makes DMAIC highly suitable for health care areas requiring high levels of measurement precision and control, such as clinical laboratories or any clinical process where outcomes need to meet high standards of care.

While all 3 methodologies share a structured, iterative approach and a reliance on empirical data to drive improvements, they cater to different needs within the health care sector. PDSA’s qualitative and flexible nature is best suited for areas requiring rapid change and adaptability. In contrast, DMAIC’s rigorous, statistical approach fits environments where precision and control are paramount. TDQM’s specific focus on DQ fills a critical niche in ensuring the reliability of health care data systems. What remains unknown is their implementations in the real world. In a review of PDSA cycles aimed at improving treatment standards, Taylor et al [60] demonstrated that <20% of implementations comply with the core features including running multiple iterative learning cycles, the notion of small-scale change, and the use of quantitative data at monthly or more frequent intervals to inform progression of cycles. The stages of each methodology are listed and described in Table 1.

Table 1 presents a comparison of the stages involved in 3 iterative learning frameworks present in health quality improvement and other related literature. The comparison aims to understand common themes in quality improvement methodology and how these can be applied to DQ improvement.

Table 1. Detailed comparison of 3 QIa frameworks for iterative learning.
Framework and stageDescription
PDSAb [60]

PlanIdentify a change hypothesis and plan a small test.

DoConduct a study plan with the collection of data.

StudyAnalyze and interpret the results.

ActAdapt the change based on feedback and plan the next iteration.
TDQMc [61]

DefineDefine target data requirements and DQd dimensions.

MeasureCreate metrics to evaluate these dimensions.

AnalyzeInvestigate root causes for DQ issues.

ImproveIdentify key areas for improvement based on DQ root cause analysis.
DMAICe [63]

DefineDefine project scope and objectives.

MeasureIdentify and measure baseline service indicators.

ActAnalyze baseline metrics and identify causes of errors.

ImproveImplement changes to reduce or remove root causes of defects.

ControlPut mechanisms in place to ensure sustained improvement.

aQI: quality improvement.

bPDSA: plan-do-study-act.

cTDQM: total data quality management.

dDQ: data quality.

eDMAIC: define-measure-analyze-improve-control.

Objectives

The rapidly growing body of DQ publications and software tools indicate that this field has gained significant traction, and recent publications illustrate that there is no shortage of DQ concepts, frameworks, and tools [21,24,28-30,54,55,66,67]. While these surveys already provide comprehensive theoretical and functional evaluations on existing DQ concepts and tools for definition and measurement, this represents only the early stages of a bigger picture in DQ improvement and management. Our aim was to evaluate the robustness of studies seeking to use DQ measurement as part of DQ improvement initiatives, focusing on improving the quality of structured EHR-derived data.

Brouwer et al [67], Wiebe et al [68], and Lemma et al [69] have previously published studies on DQ-driven improvement in health care. These studies are compared in Table 2. Wiebe et al [68] included 24 studies aiming to improve EHR documents such as operative reports or discharge summaries. The authors reported that heterogeneity in tools or metrics used to measure the quality of unstructured clinical notes made it difficult to evaluate outcomes. However, 8 included studies used an ad hoc questionnaire and 1 used the validated Physician Documentation Quality Instrument (PDQI-9) tool. Although unstructured notes in health care is a pervasive and ubiquitous source of important patient information, this scope limits the exploration of semiautomated or automated DQ assessment tools or methods [68].

Table 2 aims to compare various literature reviews focused on DQ improvement, identifying current knowledge gaps and evaluating the existing body of research. The goal is to understand current progress and unmet needs.

Brouwer et al [67] evaluated studies published up to 2005 and limited to a general practice setting. With digital health care technology and culture evolving rapidly, a more recent and broader evaluation is needed. Lemma et al [69] focused on low- and middle-income countries, where initiatives generally targeted broader and less-specific DQ improvement compared to high-income countries or technologically advanced institutions. The review expands on these works by evaluating contemporary DQ improvement studies targeting structured her-derived RWD agnostic of health care settings. Our evaluation is guided by quality improvement best practices to understand how studies measure and seek to improve DQ dimensions as defined by the well-recognized DAMA framework [20]. Specifically, we addressed the following three questions:

  1. How do quality improvement studies define and measure the quality of data?
  2. What interventions are being implemented to improve the quality of RWD?
  3. What are the outcomes reported?
Table 2. Comparison of literature reviews evaluating DQa improvement studies.
ComparisonReview paper
Brouwer et al [67] Wiebe et al [68] Lemma et al [69]
Period covered Up to 20052004-20162008-2020
Number of studies122420
Structured RWDbc
Completedness
Accuracy
Timeliness
Consistency
Validity
Uniqueness
QId framework

aDQ: data quality.

bRWD: real-world data.

cDid evaluate.

dQI: quality improvement.


Search Strategy and Information Sources

In this review, studies seeking to improve the quality of structured EHR data were examined using the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines (refer to Multimedia Appendix 1) [70]. The Population - Intervention - Comparison - Outcome - Context framework was used to identify relevant keywords and Medical Subject Headings (MeSH) terms [71,72]. These were combined using Boolean operators to create strategic search queries, which were then used to search the Ovid MEDLINE and PubMed databases for articles published from 1945 to July 2023 (refer to Multimedia Appendix 2 for more information). Additional relevant papers were identified from other publications and manual searches through Google Scholar.

Literature Selection Process

The Ovid MEDLINE and PubMed search results were downloaded as a research information systems file and PubMed text file, respectively. These were then imported into the Mendeley reference manager (Mendeley Ltd) and Rayyan (Rayyan Systems Inc) software for iterative analysis [73]. The Rayyan web app was used to streamline the selection process. Articles were selected based on the following criteria: (1) they describe a DQ assessment or measurement process, (2) they focus on data from an EHR or EMR system, and (3) they involve an intervention aimed at improving DQ over time. The search strategies and article selection process were performed independently by 2 reviewers: AL and MA. Excluded articles were nonempirical studies, improvement studies focusing on quality of care or treatment standards instead of quality of data, studies targeting semistructured or unstructured data or data not captured by an RWD source such as an EHR or EMR system, and studies without an intervention seeking to improve DQ. Table 3 summarizes the key inclusion and exclusion criteria for paper selection. In total, 39 studies were included in the review, as presented by the PRISMA flow diagram for RWD (Figure 1).

Table 3 details the specific criteria used to include or exclude studies in the evaluation of current methods for assessing and improving the quality of structured health RWD. These criteria help to systematically assess the landscape of DQ assessment and improvement strategies.

Table 3. Inclusion and exclusion criteria to evaluate the current landscape of DQa assessment and improvement approaches.
CategoryInclusion criteriaExclusion criteria
Article typeEmpirical, original, or review articles where tools, frameworks, or interventions seek to measure and improve DQNonempirical studies, thesis papers, and non–peer-reviewed publications
LanguagePublished in EnglishArticles not published in English
AccessPeer-reviewed and open access articlesPapers that are not free to access
Primary target for quality improvementStudies that primarily aim to improve the quality of dataStudies that primarily target improvement of treatment standards, standard of care, and clinical workflows without a DQ focus
Study populationStudies targeting structured, tabular dataStudies targeting semistructured or unstructured data
Data sourceData from RWDb sources such as EHRc, EMRd, PACSe, or HISf-like systemsData generated by clinical trial studies
DQ assessment and reportingStudies that describe a DQ assessment, quantification, or measurement process and implemented an intervention seeking to improve DQ over timeStudies that focus on DQ tool development without demonstration of measurement or improvement of DQ over time
Location or health contextNo criteria appliedNo criteria applied
Time frameStudies published since 1945Studies published before 1945

aDQ: data quality.

bRWD: real-world data.

cEHR: electronic health record.

dEMR: electronic medical record.

ePACS: picture archiving and communication system.

fHIS: hospital information system.

Figure 1. PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flowchart for study selection to evaluate the current literature targeting assessment, analysis, and improvement of the quality of structured real-world data in health care. This PRISMA flowchart illustrates the selection process for reports included in the review, detailing the steps of inclusion and exclusion needed to accurately achieve the intended scope of the study.

Data Extraction and Synthesis

Following paper selection, we defined a set of data elements essential for addressing the defined research questions. Subsequently, 2 authors (AL and MA) independently extracted and documented this information from each study. The collected data were then cross-checked for notable discrepancies, and any disparities were resolved through consensus. Five key sections of information were extracted: (1) study characteristics (eg, year of publication and health domain), (2) study plans outlined (eg, descriptions of target data and relevant metadata to plan DQ assessment), (3) DQ assessment (eg, methods and dimensions), (4) interventions (eg, which interventions were implemented seeking to improve DQ?), and (5) outcomes (eg, how are results reported?).


Overview of the Study Characteristics

We identified 39 studies describing DQ improvement initiatives in health care that targeted structured RWD sources [74-112]. These are listed in Tables 4 and 5. Most were published between 2009 and 2022 (37/39, 95%), with 2 published in 2002. Studies were affiliated with institutions in the United States (15/39, 38%), followed by Kenya (4/39, 10%), Australia (2/39, 5%), or 3% (1/39) each with 15 other countries. We classified the affiliated institutions into different levels of health care, including primary (11/39, 28%), secondary (10/39, 26%), tertiary (15/39, 38%), and community (3/39, 8%). The primary domains of health care were general practice (4/39, 10%), HIV care (4/39, 10%), intensive care (4/39, 10%), tropical medicine (4/39, 10%), oncology (3/39, 8%), surgery (3/39, 8%), or 3% (1/39) each from 12 other domains.

Table 4 lists the 6 DQ dimensions defined by DAMA, as described, assessed, and/or improved in all included studied. It highlights the framework’s role in standardizing and comparing DQ concepts within the reviewed literature.

DQ improvement studies targeted RWD sources that were generated by varying numbers of institutions. Target data were generated by either a single organization (9/39, 23%) or multiple different sites, which ranged from 2 to 10 (11/39, 28%), 11 to 50 (9/39, 23%), or >51 (9/39, 23%) different organizations. A variety of terminology were used to describe the source systems, including EMR (12/39, 31%), national registries or databases (8/39, 21%), EHR (7/39, 18%), HIS (7/39, 18%), clinical information networks (6/39, 15%), PACS (1/39, 3%), or claims data (1/39, 3%).

Table 4. Standardized data quality (DQ) dimensions defined by DAMAa across the included studies.
IDStudy, yearCompleteness (n=31)Validity (n=12)Uniqueness (n=4)Consistency (n=11)Timeliness (n=6)Accuracy (n=12)
1de Lusignan et al [109], 2002b

2Wallace et al [108], 2002



3Nassaralla et al [105], 2009



4Amoroso et al [112], 2010





5Griever et al [107], 2011



6Ahlbrandt et al [96], 2012




7Mphatswe et al [110], 2012



8Rahbar et al [106], 2013

9Knight et al [101], 2014



10Siegel et al [100], 2014




11Benard et al [111], 2015




12Genet et al [104], 2015





13Haskew et al [99], 2015




14Smith et al [95], 2015



15Soto et al [102], 2015




16Taggart et al [98], 2015

17Ewing et al [103], 2016




18Ma et al [77], 2016

19Tuti et al [97], 2016




20Qin et al [93], 2017




21Edgerton et al [92], 2018




22Miyoshi et al [89], 2018


23Muthee et al [94], 2018



24Qualls et al [90], 2018



25Daniel et al [88], 2019

26Bhattacharya et al [82], 2020

27Dean et al [75], 2020



28Koo et al [81], 2020





29Moomba et al [85], 2020



30Ng et al [78], 2020



31Njugunaet al [91], 2020




32Sinaiko et al [87], 2020




33Larrow et al [74], 2020



34Manesen et al [83], 2021



35Olagundoye et al [84], 2021




36Tizifa et al [86], 2021



37Kiogou et al [80], 2022

38Pfaff et al [79], 2022


39Tuti et al [76], 2022



aDAMA: Data Management Association.

bDid assess and target the DQ dimension for improvement.

Table 5. Themes of interventions for DQa improvement in included studies.
IDStudy, yearIT or technical (n=21)Training (n=17)DQ report and feedback (n=24)Workflow (n=5)Cleaning (n=3)
1de Lusignan et al [109], 2002b
2Wallace et al [108], 2002


3Nassaralla et al [105], 2009

4Amoroso et al [112], 2010

5Griever et al [107], 2011

6Ahlbrandt et al [96], 2012


7Mphatswe et al [110], 2012

8Rahbar et al [106], 2013
9Knight et al [101], 2014

10Siegel et al [100], 2014



11Benard et al [111], 2015



12Genet et al [104], 2015



13Haskew et al [99], 2015


14Smith et al [95], 2015

15Soto et al [102], 2015

16Taggart et al [98], 2015
17Ewing et al [103], 2016


18Ma et al [77], 2016
19Tuti et al [97], 2016

20Qin et al [93], 2017

21Edgerton et al [92], 2018


22Miyoshi et al [89], 2018
23Muthee et al [94], 2018

24Qualls et al [90], 2018

25Daniel et al [88], 2019
26Bhattacharya et al [82], 2020
27Dean et al [75], 2020



28Koo et al [81], 2020

29Moomba et al [85], 2020

30Ng et al [78], 2020
31Njugunaet al [91], 2020


32Sinaiko et al [87], 2020


33Larrow et al [74], 2020

34Manesen et al [83], 2021

35Olagundoye et al [84], 2021

36Tizifa et al [86], 2021

37Kiogou et al [80], 2022
38Pfaff et al [79], 2022
39Tuti et al [76], 2022

aDQ: data quality.

bDid implement.

DQ Assessment Methods

We found various approaches to DQ assessment. The duration of studies ranged from 1 month to 9 years, as did the frequency of DQ assessment. Most studies measured and reported DQ before and after the intervention (38/39, 97%) at varying intervals, including a single before and after comparison (19/39, 49%), yearly (4/39, 10%), quarterly (2/39, 5%), monthly (11/39, 28%), fortnightly (1/39, 3%), weekly (1/39, 3%), or specified data cycles (1/39, 3%). DQ assessment was achieved using manual (15/39, 38%), automated (3/39, 8%), or semiautomatic (13/39, 33%) methods, whereas some methods lacked sufficient description (6/39, 15%) to be classified. Semiautomated methods for DQ assessment mostly involved the scheduling of ad hoc, manually curated programmatic scripts using either R, SAS, or structured query language (SQL) programming languages [76,85,88,90,91,96-99,101,106-108]. Furthermore, 3 studies applied automated methods that used the World Health Organization (WHO) DQ assessment tool and the Open Data Kit [82,91,94] but did not explicitly describe how this was implemented.

The approaches to defining and assessing DQ dimensions varied. We found that 6 studies explicitly referenced 1 of 2 DQ frameworks, including Weiskopf and Weng [22] (3/39, 8%) and Kahn et al [38] (3/39, 8%) [79,80,82,88,90,94]. To enable comparison between studies, the DQ metrics and dimensions reported were extracted and classified according to the DAMA DQ framework. Some DQ concepts lacked sufficient detail to allow classification (eg, studies reporting “error rate,” “wrong data,” and “percentages of correctly coded” [93,96,112]). These were classified as “unclear” (8/39, 21%). DQ improvement studies assessed the dimensions of completeness (31/39, 79%), accuracy (12/39, 31%), validity (12/39, 31%), consistency (11/39, 28%), timeliness (6/39, 15%), and uniqueness (4/39, 10%). The number of DQ dimensions targeted per study were 1 (13/39, 33%); 2 (14/39, 36%); 3 (2/39, 5%); or 4 (7/39, 18%).

Studies reported inconsistent terminology and definitions for DQ dimensions. For example, although the completeness dimension was generally assessed as the presence or absence of expected data, variations included the proportion of linkage of records between systems [109], use of a gold standard to identify missing patients [92], or overlapping completeness with other dimensions such as validity [76] or accuracy [108]. Validity, also termed conformance or plausibility, was targeted by 12 studies. Of these 39 studies, 6 (15%) used data standards such as WHO International Classification for Disease (ICD) version 9 or 10 [77,85,89,90], SNOMED [77,79,89], Health Language 7 (HL7) [77,88,89], or RxNorm [90]. Others assessed validity by defining business rules that incorporated expected values, formats, or ranges based on local or general medical knowledge [85,89,90,98]. Studies occasionally equated validity with accuracy or correctness [76,89,98].

Of the 12 studies that targeted accuracy, 8 (67%) reported the development and/or use of a gold standard for reference. However, varying definitions for what studies deemed to be a “gold standard” were provided. This included paper charts [93,108]; national data [91,110]; a manually curated data set [84,106]; or manual validation by a trained, expert clinical coder [103]. For example, the “gold standard” in the study by Rahbar et al [106] included 30 patient records that were manually abstracted by a team of experts that included a vascular neurologist clinician before comparing to national stroke registry records. In another study by Ahlbrandt et al [96], the gold standard was described as “the documented and encoded (using OPS [Operationen und Prozedurenschlüssel] Classification, the German modification of ICPM [International Classification of Procedures in Medicine]) surgical procedure,” which could be interpreted either as a data standard or patient data in electronic or paper form.

Sometimes DQ dimensions were subsumed by another. For example, data could only be deemed accurate when they were both complete and correct across multiple data elements [105]. In the absence of what is deemed a “gold standard,” data consistency was similarly assessed by comparison with paper records [77,83,94], multiple registers [86], or national data [82].

We found that studies’ assessment of uniqueness and timeliness were generally consistent with the DAMA definitions. For uniqueness, studies assessed if records were unexpectedly duplicated, for example, in primary keys [79] or patient names [95]. Similarly, timeliness was consistently assessed as the difference in time between point of data capture versus actual timing of events [74,75,77,81,82,104]. In contrast, some of the DQ concepts reported could not be classified according to the DAMA DQ framework, including simplicity, acceptability, flexibility, stability, usefulness [77], and conformance to a specified data model [90]. These were collectively classified as “other” (3/39, 8%).

Interventions for Improvement

Studies varied in their approaches to plan and deliver DQ improvement interventions. In total, 20 studies reported using quantitative or qualitative data analysis before planning an intervention. Qualitative analysis involved assessment of clinical workflow inefficiencies through process mapping techniques or staff surveys [74,108]. In contrast, quantitative analysis involved an assessment of DQ with interpretations of possible root causes [95,96].

To understand the types of interventions studied, we identified 5 common themes, including DQ reporting and feedback (24/39, 62%), IT-related or technical solutions (21/39, 54%), training (17/39, 44%), workflow (5/39, 13%), or data cleaning (3/39, 8%). All studies implemented at least 1 intervention with most implementing multiple interventions (23/39, 59%). DQ reporting and feedback involved assessing DQ and sharing curated results with a specific stakeholder with the aim of encouraging improved data capture behavior. These stakeholders included individual clinical staff or managers [87,98,105] or health care institutions as a whole [82,88,90,101,109,110].

Taggart et al [98] implemented structured DQ reports combined with feedback sessions to improve the quality of EHR data in general practice settings. This approach leveraged regular assessments and direct feedback to practice managers to foster ongoing improvements in data recording practices, illustrating a practical application of DQ feedback mechanisms in a real-world health care setting [98]. In contrast, Sinaiko et al [87], studied peer comparison feedback emails in a randomized controlled study to assess its effectiveness on improving cancer stage data completeness, underscoring the importance of control groups in validating the impact of DQ interventions.

We found a range of subthemes under the IT-related or technical-based interventions. These improvements involved either introducing a new electronic data capture system, upgrading an existing one [77,81,93,99,102,103,108,111], enhancing front-end user interfaces [74,78,96], or refining back-end data flow processes. Ahlbrandt et al [96] introduced an intervention focusing on improving the graphical user interface of anesthesia information management systems to enhance the validity of the data captured. By shifting from drop-down lists to radio buttons, rearranging the graphical user interface layout, and limiting user options to a set list, they aimed to reduce invalid data entry by making the interface more intuitive and compliant with data standards. This study exemplifies how interface design can directly influence data validity and highlights the impact of front-end modifications [96].

Technology-based interventions often overlapped with training and workflow changes. Ewing et al [103] implemented a browser-assisted clinical coding software along with training, which, in turn, improved efficiencies in clinical workflows. Other studies mainly targeted workflow inefficiencies [74,85,104,107], with Greiver et al [107] introducing a data entry clerk, whereas Moomba et al [85] shifting data entry responsibility from data entry clerks to frontline clinical staff.

To plan, implement, and assess the impact of these DQ improvement interventions, we found that of the 39 studies, only 5 (13%) used a standardized quality improvement framework or iterative learning cycle, such as PDSA (4/39, 10%) [74,81,84,101] or DMAIC (1/39, 3%) [88]. All 4 PDSA studies completed multiple cycles, ranging from 3 to 8. One study reportedly conducted 421 PDSA cycles across 54 different sites [101]. Larrow et al [74] applied the PDSA method to enhance the timeliness of discharge summaries at a pediatric hospital. The study team initiated their quality improvement project by identifying key barriers through qualitative analysis of staff surveys, leading to the strategic redesign of the EHR structured discharge summary template. Notable enhancements included embedded writing tips and standardized drop-down menus for common diagnoses.

Daniel et al [88] applied the DMAIC methodology to define and assess DQ issues. These authors correlated specific DQ dimensions to possible technical issues, for example, data lacking standardization or valid entries may be caused by “errors from data originators, ETL issues or limitations of the EHR data entry tool (inadequate value set constrains, lack of DQ checks)” [88]. By measuring and analyzing these problems in a structured methodology, the team identified key areas that required targeted interventions, such as use of data standards to enforce data validation rules to the data entry system [88].

Table 5 summarizes the various intervention themes implemented to improve the quality of structured RWD in various health care contexts. It provides insights into common strategies in DQ improvement.

Reported Outcomes

To understand and compare the outcomes of DQ improvement initiatives, we identified whether studies reported DQ changes that were better, worse, or showed no change over time. Most studies reported improvements in DQ over time (36/39, 92%). This excludes 3 studies due to results being reported as preliminary [88], potential improvements as opposed to actual [92], or without sufficient detail [79]. Of the 36 studies showing improvement, 9 (23%) also report decreases in DQ [96,100,106,109,111], of which another 4 (10%) also report no changes [82,85,98]. These changes were reported at varying levels of granularity. While most studies reported DQ metrics for specific data items, such as validity of surgical procedure codes [96], others aggregated multiple metrics or dimensions into higher level entities, such as “92% reduction in error rate” [112] and mean monthly accuracy for pediatric early warning scores [75].

We also assessed whether statistical tests were used to demonstrate significance of effect and whether studies compared intervention groups with a control group. When determining significance of treatment effect, 22 studies used at least 1 statistical test or method [74-76,78,81-83,87,93,94, 96,98-100,103-107,109,110]. These ranged from chi-square (7/39, 18%) and statistical process control charts (4/39, 10%) to multivariable linear (1/39, 3%) and logistic regression (1/39, 3%). Of these 22 studies, only 1 (5%) study compared the intervention group to a concurrent control group, which reported improvement in completeness of cancer stage data [87].


Principal Findings

In this paper, we conducted a systematic literature review to understand the current practices in DQ improvement of structured RWD in a health care context. We found substantial heterogeneity in the approaches to definition, assessment, and interventions across the reviewed literature. The range of definitions for DQ concepts, quality improvement methodologies, and reported outcomes have made synthesis and comparison of the results challenging. In the following sections, we explore these 3 points in greater depth.

DQ Assessment

A key issue in the exploration of DQ is the lack of consensus on theoretical definitions for DQ assessment. Despite the existence of several DQ frameworks, there are no agreed recommendations or guidelines on which frameworks should be used or on how dimensions should be defined, measured, or used to understand real-world issues in data capture, processing, and utility for high-quality RWE generation. This has been demonstrated in a wealth of previous reviews on DQ theory [22,23,25,26,38,48], but to a lesser extent in a quality improvement context. In our review, we found that while some studies did in fact reference theoretical frameworks by Weiskopf and Weng [22] and Kahn et al [38], these account for <15% of all included studies. This indicates a severe lack of uptake of standardized DQ theory in the wider literature and explains the substantial variation and lack of consensus. In turn, the lack of agreement and consistency makes it difficult to harness the true purpose of DQ assessment, which pertains to its ability to identify issues in real-world processes, behaviors, and resources. While some studies demonstrate qualitative correlations between DQ issues and underlying real-world problems [48], we found that only a small minority of studies implemented a quantitative approach to make a similar connection.

As DQ is a complex, multidimensional construct, each dimension serves to identify context-specific issues in the real-world needing remediation. We found that some authors made this correlation either directly or indirectly; for example, data validity is affected by a lack of standardization of front-end user interfaces on electronic data capture forms [96], timeliness of data indicates possible workflow inefficiencies that delays the point of data capture [74,104], duplication highlights redundant data sources [95], inaccurate data underpin lack of training on medical coding standards [103], and inconsistencies between data sources indicate possible capture of inaccurate data [83,86]. This raises 2 important points: the need to assess DQ beyond completeness or missingness and the importance of standardized frameworks. Without these, crucial error-prone processes in complex clinical pathways may go undiagnosed and continue to generate poor-quality data. This is particularly important given the growing demand for and expectations of real-world health care data, the hype in artificial intelligence, and the growing awareness that maintaining patient records is the leading cause of clinician burnout [42,113].

There was limited reporting of the tools or software used for DQ assessment included in this review. Only 2 DQ assessment tools were reported: the WHO DQ assessment toolkit [82,91,94] and the “Open Data Kit” [86]. Neither of these are explained in sufficient depth to discern how they work or their applicability to other environments. Further investigation into the referenced material also lacked sufficient information. Other tools reported were scheduled programmatic scripts using R, SQL, or SAS software for DQ assessment. Some of these methods are considered “automated” solutions for DQ assessment. This indicates a significant gap between the vast range of DQ software available and the practical implementation of these tools for DQ assessment, causal analysis, and improvement. DQ software must be capable of profiling large volumes of structured data, provide both automated and user specified DQ assessment methods, and facilitate meaningful analysis of possible root causes of poor DQ [21,30]. The limited adoption of existing DQ software might suggest a deficiency in technical proficiency, inadequate documentation clarifying its utility or use cases, or a lack of awareness regarding its availability or relevance.

Quality Improvement Cycles

Another characteristic of the studies included in this review was the limited use of quality improvement frameworks. Only 5 studies [74,81,84,88,101] referenced a quality improvement methodology to plan and implement DQ improvement interventions. This is surprising given the potential benefits these frameworks offer, particularly in fostering systematic, structured, and dynamic approaches to improvement in complex environments.

Quality improvement frameworks, such as PDSA, DMAIC, and TDQM, if implemented robustly, can significantly improve comparability and knowledge sharing between studies, institutions, and organizational teams. This is important given that Siegel et al [100] observed varying improvements across different organizations, stating that systematic and organized quality improvement efforts are needed.

However, the strengths of these frameworks extend beyond iterative learning; they also encourage a deep dive into DQ analysis, helping to unravel the complex relationships between various real-world factors and the root causes of poor DQ. In this way, interventions can be designed in collaboration with the affected stakeholders, that is, frontline clinical staff, to maximize the opportunity for DQ improvement. Knight et al [101] particularly emphasized this point, stating that the quality improvement model, that is, PDSA, “used in this project facilitated the identification and correction of difficulties with the technology of the innovation.” Quality improvement frameworks, such as PDSA, TDQM, or DMAIC, can also be adapted to improve the quality of real-world health care data incorporating DQ-driven quantitative analysis alongside real-world issues that can be identified using the Odigos framework [48].

Despite their strengths, the application of these frameworks is not without challenges. One significant constraint is the need for substantial upfront planning and stakeholder engagement, which can be resource intensive. Furthermore, these frameworks require a culture of continuous improvement and openness to change and adoption of data governance practices, which may not be present in all health care settings. This can limit their applicability and effectiveness. In addition, the lack of consistent application and reporting on the use of these frameworks can make it difficult to evaluate their true effectiveness.

Outcomes

We sought to investigate the current approaches to DQ assessment and improvement to synthesize and summarize the lessons learnt from these endeavors. In general, studies reported positive changes in DQ through the implementation of multiple interventions. Lemma et al [69] associated the benefits to DQ when interventions such as training, technical innovation, and DQ feedback were combined. The same authors reported that studies that focused only on single interventions did not generate equally positive DQ changes. In contrast, we found that 17 studies focused on a single intervention showing mostly positive results. For example, Sinaiko et al [87] demonstrated the positive impact of peer comparison emails to completion of cancer stage data when compared to a control group. Similarly, studies that demonstrated a combination of improvements, reductions, and no changes in DQ often implemented multiple interventions.

Studies highlighted the importance of close collaboration with clinical users’ needs when implementing digital technologies [101,108,112] and the importance of detailed, personalized feedback on data capture performance provided to data capture end users [87,107,109]. In contrast, Taggart et al [98] found that peer comparison and feedback sessions did not result in better DQ and suggested the need for randomized controlled studies. This raises an important question regarding the need for control groups in DQ improvement studies. In our review, we found only 1 study that compared an intervention group with a concurrent control group, yet most report successful improvement in DQ.

While some studies highlighted the need for controlled environments, we observed that interventions were chosen for reasons other than baseline quantitative or qualitative analysis or analysis of underlying causes. In other words, only a few studies planned an intervention based on a data-driven approach. When comparing outcomes, most studies compared average baselines before and after the intervention, where only 1 study compared the intervention group to a concurrent nonintervention control group. A simultaneous control group for comparison can facilitate analysis of the cause of effect of treatment interventions along with the implementation of multiple interventions, which occurred in most studies. This, combined with the lack of data-driven approaches when planning or designing interventions, indicates a significant gap in robust, standardized DQ improvement methodologies. Therefore, the reported outcomes should be considered with caution.

The methodological rigor in DQ improvement studies often suffers from a lack of randomized controls and consistent statistical methodologies. As discussed before, only 1 reviewed study included a randomized control group to demonstrate significant improvement [87], and just 22 (56%) out of 39 studies applied diverse statistical tests, such as chi-square and logistic regression. This inconsistency in applying robust analytical techniques can introduce biases, misattributing improvements to interventions rather than actual effects. Moreover, the absence of uniform experimental designs across various health care settings undermines the robustness and generalizability of findings.

Incorporating structured methodologies such as PDSA or DMAIC could significantly enhance the methodological rigor of these studies. These frameworks support systematic implementations and evaluations, facilitating the use of control groups and statistical analysis to reliably isolate intervention effects. By adopting such standardized approaches, future research could more effectively ensure the credibility and applicability of the findings, fostering the development of evidence-based interventions suitable for diverse health care environments.

Future Recommendations

This review highlights the need for standardized and systematic approaches to DQ assessment, analysis, and improvement. This can be addressed in future studies by following quality improvement methodologies, such as the PDSA [114], TDQM [61], or DMAIC [63] iterative learning cycles, and DQ frameworks, such as DAMA [20], Weiskopf and Weng [22], or Kahn et al [38]. Furthermore, understanding the root causes of poor DQ is essential for planning the most appropriate intervention. This intervention should aim to address issues as close to the point of data capture as possible.

The need for standardized DQ assessment is evident. Future research and development should focus on the development and demonstration of DQ tools that are not only grounded in theoretical frameworks, such as those offered by DAMA [20], Weiskopf and Weng [22], or Kahn et al [38], but are also highly accessible and user-friendly. DQ tools should come with comprehensive documentation and practical examples that enable users in making informed decisions about their applicability and relevance in specific health care settings.

As discussed in the Introduction section, DQ tooling currently lack in usability and usefulness [21,29,30]. Ease of use can be overcome by introducing “plug-and-play” functionality that is combined with useful customizable features. This duality can allow users to quickly test and assess the tool’s immediate value and adjust and extend its functionality to fit more complex, specific needs over time. By extension, DQ tools should produce results that are useful for meaningful, in-depth analysis and monitoring of DQ errors.

Current best practices in root cause analysis of poor real-world health care data are unknown but could be facilitated using a framework such as the Odigos framework [48]. Furthermore, understanding causes of poor data can facilitate the design and selection of more relevant interventions needed—an aspect of DQ management that was demonstrated by few articles in this review. Future studies may also wish to compare intervention groups to concurrent control groups and explore 1 intervention at a time instead of multiple. This may help to control for external factors and increase understanding of barriers to high-quality data capture.

Limitations

This review aimed to summarize the lessons learnt from DQ improvement studies. As an abundance of literature already highlights the substantial variation in terminology for DQ concepts, we used the DAMA DQ framework to standardize the heterogeneity in DQ terms and definitions. In doing so, some DQ concepts could not be classified, potentially affecting the frequency counts of DQ dimensions assessed. Another limitation is that we were unable to perform a comprehensive meta-analysis of the methodological constrains and the effect measures of the reported outcomes. We believe the significant scope of this work warrants future research. This was due to considerable variation in the methods for assessment, analysis, and reporting of DQ metrics and changes over time. This posed significant challenges when attempting to objectively elucidate the effect of treatment interventions.

Moreover, while this review captures a rise in DQ improvement studies, with 31 (79%) out of 39 studies published in the last decade, it also includes 8 studies that were published in or before 2013, potentially missing recent advancements in digital health technologies. In addition, the discovery of 6 additional studies from manual searches indicates the likely exclusion of other relevant work. This is particularly the case due to the lack of consistency in DQ terminology and definitions, which made it difficult to capture all possible variations of DQ terms in the search strategy. Despite these challenges, our review included twice as many studies compared to other related reviews, indicating a thorough coverage within the constraints identified. We believe the significant scope of this work warrants future updates to include emerging trends and methodologies in DQ improvement.

Conclusions

The reviewed studies demonstrate that approaches to DQ improvement vary in their methodologies, definitions, and reporting of DQ dimensions. In general, studies implemented multiple interventions and reported positive changes in the quality of structured real-world health care data. In addition to “going paperless” initiatives, studies demonstrated the benefits of engagement with frontline clinical end users, provision of personalized DQ feedback, streamlining clinical workflows, and raising awareness of DQ and data standards aimed at improving DQ in health care settings. Despite this, heterogeneity is a major limitation among DQ literature in general, and we recommend that studies refer to standardized frameworks, such as PDSA cycles for quality improvement and the DAMA DQ framework for assessing DQ dimensions. This would lead to greater consistency and comparison in the reported outcomes.

Acknowledgments

This work was supported by Cancer Research UK RadNet Manchester (C1994/A28701). GP was supported by the NIHR (National Institute for Health and Care Research) Manchester Biomedical Research Centre (NIHR203308). The authors would like to thank Dr Catherine O’Hara, Head of Analytics from the Clinical Outcomes and Data Unit, for her support and guidance in this project, as well as Daniel Livesey, Librarian at The Christie NHS Foundation Trust, for his invaluable training and assistance on defining and implementing the literature search strategy.

Data Availability

All data generated or analyzed during this study are included in this manuscript and its multimedia appendices.

Conflicts of Interest

None declared.

Multimedia Appendix 1

PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) checklist.

DOCX File , 22 KB

Multimedia Appendix 2

Search strategy and results.

DOCX File , 16 KB

  1. Knevel R, Liao KP. From real-world electronic health record data to real-world results using artificial intelligence. Ann Rheum Dis. Mar 23, 2023;82(3):306-311. [FREE Full text] [CrossRef] [Medline]
  2. Liu F, Panagiotakos D. Real-world data: a brief review of the methods, applications, challenges and opportunities. BMC Med Res Methodol. Nov 05, 2022;22(1):287. [FREE Full text] [CrossRef] [Medline]
  3. Real-world evidence. U.S. Food & Drug Administration. May 2, 2023. URL: https://www.fda.gov/science-research/science-and-research-special-topics/real-world-evidence [accessed 2023-06-21]
  4. Atasoy H, Greenwood BN, McCullough JS. The digitization of patient care: a review of the effects of electronic health records on health care quality and utilization. Annu Rev Public Health. Apr 01, 2019;40:487-500. [CrossRef] [Medline]
  5. Benson T. Why general practitioners use computers and hospital doctors do not--part 1: incentives. BMJ. Nov 09, 2002;325(7372):1086-1089. [FREE Full text] [CrossRef] [Medline]
  6. Anshari M. Redefining electronic health records (EHR) and electronic medical records (EMR) to promote patient empowerment. Int J Inform Dev. Jun 22, 2019;8(1):35-39. [CrossRef]
  7. Blonde L, Khunti K, Harris SB, Meizinger C, Skolnik NS. Interpretation and impact of real-world clinical data for the practicing clinician. Adv Ther. Nov 2018;35(11):1763-1774. [FREE Full text] [CrossRef] [Medline]
  8. Guinn D, Wilhelm EE, Lieberman G, Khozin S. Assessing function of electronic health records for real-world data generation. BMJ Evid Based Med. Jun 2019;24(3):95-98. [CrossRef] [Medline]
  9. Kim HS, Lee S, Kim JH. Real-world evidence versus randomized controlled trial: clinical research based on electronic medical records. J Korean Med Sci. Jun 26, 2018;33(34):e213. [FREE Full text] [CrossRef] [Medline]
  10. Dang A. Real-world evidence: a primer. Pharmaceut Med. Jan 2023;37(1):25-36. [FREE Full text] [CrossRef] [Medline]
  11. NICE real-world evidence framework corporate document. National Institute for Health and Care Excellence. Jun 23, 2022. URL: https://www.nice.org.uk/corporate/ecd9/chapter/overview [accessed 2024-08-06]
  12. Price G, Mackay R, Aznar M, McWilliam A, Johnson-Hart C, van Herk M, et al. Learning healthcare systems and rapid learning in radiation oncology: where are we and where are we going? Radiother Oncol. Nov 2021;164:183-195. [CrossRef] [Medline]
  13. Zhang J, Symons J, Agapow P, Teo JT, Paxton CA, Abdi J, et al. Best practices in the real-world data life cycle. PLOS Digit Health. Jan 18, 2022;1(1):e0000003. [FREE Full text] [CrossRef] [Medline]
  14. Blacketer C, Defalco FJ, Ryan PB, Rijnbeek PR. Increasing trust in real-world evidence through evaluation of observational data quality. J Am Med Inform Assoc. Sep 18, 2021;28(10):2251-2257. [FREE Full text] [CrossRef] [Medline]
  15. Thygesen JH, Tomlinson C, Hollings S, Mizani MA, Handy A, Akbari A, et al. COVID-19 trajectories among 57 million adults in England: a cohort study using electronic health records. Lancet Digit Health. Jul 2022;4(7):e542-e557. [FREE Full text] [CrossRef] [Medline]
  16. NICE real-world evidence framework. National Institute for Health and Care Excellence. Jun 23, 2022. URL: https://www.nice.org.uk/corporate/ecd9/chapter/overview [accessed 2023-06-21]
  17. Data quality framework for medicines regulation. European Medicines Agency. URL: https:/​/www.​ema.europa.eu/​en/​about-us/​how-we-work/​big-data/​data-quality-framework-medicines-regulation [accessed 2024-08-06]
  18. MHRA guidance on the use of real-world data in clinical studies to support regulatory decisions. United Kingdom Government. Dec 16, 2021. URL: https://tinyurl.com/2rkjjnfs [accessed 2023-10-09]
  19. Castellanos EH, Wittmershaus BK, Chandwani S. Raising the bar for real-world data in oncology: approaches to quality across multiple dimensions. JCO Clin Cancer Inform. Jan 2024;8:e2300046. [FREE Full text] [CrossRef] [Medline]
  20. Earley S, Henderson D, Sebastian-Coleman L. The DAMA Guide to the Data Management Body of Knowledge (DAMA-DM BOK). Bradley Beach, NJ. Technics Publications, LLC; 2017.
  21. Ozonze O, Scott PJ, Hopgood AA. Automating electronic health record data quality assessment. J Med Syst. Feb 13, 2023;47(1):23. [FREE Full text] [CrossRef] [Medline]
  22. Weiskopf NG, Weng C. Methods and dimensions of electronic health record data quality assessment: enabling reuse for clinical research. J Am Med Inform Assoc. Jan 01, 2013;20(1):144-151. [FREE Full text] [CrossRef] [Medline]
  23. Liaw ST, Rahimi A, Ray P, Taggart J, Dennis S, de Lusignan S, et al. Towards an ontology for data quality in integrated chronic disease management: a realist review of the literature. Int J Med Inform. Jan 2013;82(1):10-24. [CrossRef] [Medline]
  24. Cichy C, Rass S. An overview of data quality frameworks. IEEE Access. Feb 15, 2019;7:24634-24648. [CrossRef]
  25. Chen H, Hailey D, Wang N, Yu P. A review of data quality assessment methods for public health information systems. Int J Environ Res Public Health. May 14, 2014;11(5):5170-5207. [FREE Full text] [CrossRef] [Medline]
  26. Bian J, Lyu T, Loiacono A, Viramontes TM, Lipori G, Guo Y, et al. Assessing the practice of data quality evaluation in a national clinical data research network through a systematic scoping review in the era of real-world data. J Am Med Inform Assoc. Dec 09, 2020;27(12):1999-2010. [FREE Full text] [CrossRef] [Medline]
  27. Mariño J, Kasbohm E, Struckmann S, Kapsner LA, Schmidt CO. R packages for data quality assessments and data monitoring: a software scoping review with recommendations for future developments. Appl Sci. Apr 22, 2022;12(9):4238. [CrossRef]
  28. Syed R, Eden R, Makasi T, Chukwudi I, Mamudu A, Kamalpour M, et al. Digital health data quality issues: systematic review. J Med Internet Res. Mar 31, 2023;25:e42615. [FREE Full text] [CrossRef] [Medline]
  29. Gordon B, Fennessy C, Varma S, Barrett J, McCondochie E, Heritage T, et al. Evaluation of freely available data profiling tools for health data research application: a functional evaluation review. BMJ Open. May 09, 2022;12(5):e054186. [FREE Full text] [CrossRef] [Medline]
  30. Ehrlinger L, Wöß W. A survey of data quality measurement and monitoring tools. Front Big Data. Mar 31, 2022;5:850611. [FREE Full text] [CrossRef] [Medline]
  31. Parry MG, Sujenthiran A, Cowling TE, Charman S, Nossiter J, Aggarwal A, et al. Imputation of missing prostate cancer stage in English cancer registry data based on clinical assumptions. Cancer Epidemiol. Feb 2019;58:44-51. [CrossRef] [Medline]
  32. Westerberg M, Beckmann K, Gedeborg R, Irenaeus S, Holmberg L, Garmo H, et al. Choice of imputation method for missing metastatic status affected estimates of metastatic prostate cancer incidence. J Clin Epidemiol. Mar 2023;155:22-30. [FREE Full text] [CrossRef] [Medline]
  33. Daniels B, Havard A, Myton R, Lee C, Chidwick K. Evaluating the accuracy of data extracted from electronic health records into MedicineInsight, a national Australian general practice database. Int J Popul Data Sci. Jun 29, 2022;7(1):1713. [FREE Full text] [CrossRef] [Medline]
  34. Palestine AG, Merrill PT, Saleem SM, Jabs DA, Thorne JE. Assessing the precision of ICD-10 codes for uveitis in 2 electronic health record systems. JAMA Ophthalmol. Oct 01, 2018;136(10):1186-1190. [FREE Full text] [CrossRef] [Medline]
  35. Horsky J, Drucker EA, Ramelson HZ. Accuracy and completeness of clinical coding using ICD-10 for ambulatory visits. AMIA Annu Symp Proc. Apr 16, 2018;2017:912-920. [FREE Full text] [Medline]
  36. Botsis T, Hartvigsen G, Chen F, Weng C. Secondary use of EHR: data quality issues and informatics opportunities. Summit Transl Bioinform. Mar 01, 2010;2010:1-5. [FREE Full text] [Medline]
  37. Cook LA, Sachs J, Weiskopf NG. The quality of social determinants data in the electronic health record: a systematic review. J Am Med Inform Assoc. Dec 28, 2021;29(1):187-196. [FREE Full text] [CrossRef] [Medline]
  38. Kahn MG, Callahan TJ, Barnard J, Bauck AE, Brown J, Davidson BN, et al. A harmonized data quality assessment terminology and framework for the secondary use of electronic health record data. EGEMS (Wash DC). Sep 11, 2016;4(1):1244. [FREE Full text] [CrossRef] [Medline]
  39. Bright CJ, Lawton S, Benson S, Bomb M, Dodwell D, Henson KE, et al. Data resource profile: the systemic anti-cancer therapy (SACT) dataset. Int J Epidemiol. Feb 01, 2020;49(1):15-1l. [FREE Full text] [CrossRef] [Medline]
  40. Sarwar T, Seifollahi S, Chan J, Zhang X, Aksakalli V, Hudson I, et al. The secondary use of electronic health records for data mining: data characteristics and challenges. ACM Comput Surv. Jan 18, 2022;55(2):1-40. [CrossRef]
  41. Zhang J, Ashrafian H, Delaney B, Darzi A. Impact of primary to secondary care data sharing on care quality in NHS England hospitals. NPJ Digit Med. Aug 14, 2023;6(1):144. [FREE Full text] [CrossRef] [Medline]
  42. Diaz-Garelli F, Strowd R, Ahmed T, Lycan TWJ, Daley S, Wells BJ, et al. What oncologists want: identifying challenges and preferences on diagnosis data entry to reduce EHR-induced burden and improve clinical data quality. JCO Clin Cancer Inform. May 2021;5:527-540. [FREE Full text] [CrossRef] [Medline]
  43. Ullah E, Albrett J, Khan O, Matthews C, Perry I, GholamHosseini H, et al. Workload involved in vital signs-based monitoring & responding to deteriorating patients: a single site experience from a regional New Zealand hospital. Heliyon. Oct 06, 2022;8(10):e10955. [FREE Full text] [CrossRef] [Medline]
  44. Pérez-Martí M, Casadó-Marín L, Guillén-Villar A. Electronic records with tablets at the point of care in an internal medicine unit: before-after time motion study. JMIR Hum Factors. Feb 10, 2022;9(1):e30512. [FREE Full text] [CrossRef] [Medline]
  45. Byrd JB, Vigen R, Plomondon ME, Rumsfeld JS, Box TL, Fihn SD, et al. Data quality of an electronic health record tool to support VA cardiac catheterization laboratory quality improvement: the VA Clinical Assessment, Reporting, and Tracking System for Cath Labs (CART) program. Am Heart J. Mar 2013;165(3):434-440. [CrossRef] [Medline]
  46. Skyttberg N, Vicente J, Chen R, Blomqvist H, Koch S. How to improve vital sign data quality for use in clinical decision support systems? A qualitative study in nine Swedish emergency departments. BMC Med Inform Decis Mak. Jun 04, 2016;16:61. [FREE Full text] [CrossRef] [Medline]
  47. Dall'Ora C, Griffiths P, Hope J, Briggs J, Jeremy J, Gerry S, et al. How long do nursing staff take to measure and record patients' vital signs observations in hospital? A time-and-motion study. Int J Nurs Stud. Jun 2021;118:103921. [FREE Full text] [CrossRef] [Medline]
  48. Eden R, Syed R, Makasi T, Andrews R, Ter Hofstede A, Wynn M, et al. Revealing the root causes of digital health data quality issues: a qualitative investigation of the Odigos framework. In: Proceedings of the 56th Hawaii International Conference on System Sciences. 2023. Presented at: HICSS 2023; January 3-6, 2023; Maui, HI. [CrossRef]
  49. Sockolow PS, Bowles KH, Adelsberger MC, Chittams JL, Liao C. Impact of homecare electronic health record on timeliness of clinical documentation, reimbursement, and patient outcomes. Appl Clin Inform. Apr 30, 2014;5(2):445-462. [FREE Full text] [CrossRef] [Medline]
  50. Lee D, Jiang X, Yu H. Harmonized representation learning on dynamic EHR graphs. J Biomed Inform. Jun 2020;106:103426. [FREE Full text] [CrossRef] [Medline]
  51. von Lucadou M, Ganslandt T, Prokosch HU, Toddenroth D. Feasibility analysis of conducting observational studies with the electronic health record. BMC Med Inform Decis Mak. Oct 28, 2019;19(1):202. [FREE Full text] [CrossRef] [Medline]
  52. Zayas-Cabán T, Haque SN, Kemper N. Identifying opportunities for workflow automation in health care: lessons learned from other industries. Appl Clin Inform. May 2021;12(3):686-697. [FREE Full text] [CrossRef] [Medline]
  53. Tsou AY, Lehmann CU, Michel J, Solomon R, Possanza L, Gandhi T. Safe practices for copy and paste in the EHR. Systematic review, recommendations, and novel model for health IT collaboration. Appl Clin Inform. Jan 11, 2017;8(1):12-34. [FREE Full text] [CrossRef] [Medline]
  54. Quindroit P, Fruchart M, Degoul S, Perichon R, Martignène N, Soula J, et al. Definition of a practical taxonomy for referencing data quality problems in health care databases. Methods Inf Med. May 2023;62(1-02):19-30. [FREE Full text] [CrossRef] [Medline]
  55. Liaw ST, Guo JG, Ansari S, Jonnagaddala J, Godinho MA, Borelli AJ, et al. Quality assessment of real-world data repositories across the data life cycle: a literature review. J Am Med Inform Assoc. Jul 14, 2021;28(7):1591-1599. [FREE Full text] [CrossRef] [Medline]
  56. Shi X, Prins C, van Pottelbergh G, Mamouris P, Vaes B, de Moor B. An automated data cleaning method for electronic health records by incorporating clinical knowledge. BMC Med Inform Decis Mak. Sep 17, 2021;21(1):267. [FREE Full text] [CrossRef] [Medline]
  57. Phan HT, Borca F, Cable D, Batchelor J, Davies JH, Ennis S. Automated data cleaning of paediatric anthropometric data from longitudinal electronic health records: protocol and application to a large patient cohort. Sci Rep. Jun 23, 2020;10(1):10164. [FREE Full text] [CrossRef] [Medline]
  58. Dziadkowiec O, Callahan T, Ozkaynak M, Reeder B, Welton J. Using a data quality framework to clean data extracted from the electronic health record: a case study. EGEMS (Wash DC). Jun 24, 2016;4(1):1201. [FREE Full text] [CrossRef] [Medline]
  59. Backhouse A, Ogunlayi F. Quality improvement into practice. BMJ. Mar 31, 2020;368:m865. [FREE Full text] [CrossRef] [Medline]
  60. Taylor MJ, McNicholas C, Nicolay C, Darzi A, Bell D, Reed JE. Systematic review of the application of the plan-do-study-act method to improve quality in healthcare. BMJ Qual Saf. Apr 2014;23(4):290-298. [FREE Full text] [CrossRef] [Medline]
  61. Wang RY. A product perspective on total data quality management. Commun ACM. Feb 01, 1998;41(2):58-65. [CrossRef]
  62. D'Andreamatteo A, Ianni L, Lega F, Sargiacomo M. Lean in healthcare: a comprehensive review. Health Policy. Sep 2015;119(9):1197-1209. [CrossRef] [Medline]
  63. Sokovic M, Pavletic D, Pipan KK. Quality improvement methodologies – PDCA cycle, RADAR matrix, DMAIC and DFSS. J Achiev Mater Manuf Eng. 2010;43(1):476-483.
  64. Deming WE. Elementary Principles of the Statistical Control of Quality: A Series of Lectures. Tokyo, Japan. Nippon Kagaku Gijutsu Remmei; 1950.
  65. Moen R, Norman C. Evolution of the PDCA cycle. In: Proceedings of the 7th ANQ Congress. 2009. Presented at: ANQ 2009; September 15-18, 2009; Tokyo, Japan. URL: https://www.anforq.org/activities/congresses/index.html
  66. Burns L, Roux NL, Kalesnik-Orszulak R, Christian J, Hukkelhoven M, Rockhold F, et al. Real-world evidence for regulatory decision-making: guidance from around the world. Clin Ther. Mar 2022;44(3):420-437. [FREE Full text] [CrossRef] [Medline]
  67. Brouwer HJ, Bindels PJ, Weert HC. Data quality improvement in general practice. Fam Pract. Oct 2006;23(5):529-536. [CrossRef] [Medline]
  68. Wiebe N, Otero Varela L, Niven DJ, Ronksley PE, Iragorri N, Quan H. Evaluation of interventions to improve inpatient hospital documentation within electronic health records: a systematic review. J Am Med Inform Assoc. Nov 01, 2019;26(11):1389-1400. [FREE Full text] [CrossRef] [Medline]
  69. Lemma S, Janson A, Persson LÅ, Wickremasinghe D, Källestål C. Improving quality and use of routine health information system data in low- and middle-income countries: a scoping review. PLoS One. Oct 8, 2020;15(10):e0239683. [FREE Full text] [CrossRef] [Medline]
  70. Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. Mar 29, 2021;372:n71. [FREE Full text] [CrossRef] [Medline]
  71. Carrera-Rivera A, Larrinaga F, Lasa G. Context-awareness for the design of smart-product service systems: literature review. Comput Ind. Nov 2022;142:103730. [CrossRef]
  72. Yang H, Lee HJ. Research trend visualization by MeSH terms from PubMed. Int J Environ Res Public Health. May 30, 2018;15(6):1113. [FREE Full text] [CrossRef] [Medline]
  73. Rayyan homepage. Rayyan. URL: https://www.rayyan.ai/ [accessed 2023-06-30]
  74. Larrow A, Chong A, Robison T, Patel A, Kuelbs C, Fisher E, et al. A quality improvement initiative to improve discharge timeliness and documentation. Pediatr Qual Saf. Jul 28, 2021;6(4):e440. [FREE Full text] [CrossRef] [Medline]
  75. Dean NP, Cheng JJ, Crumbley I, DuVal J, Maldonado E, Ghebremariam E. Improving accuracy and timeliness of nursing documentation of pediatric early warning scores. Pediatr Qual Saf. Mar 25, 2020;5(2):e278. [FREE Full text] [CrossRef] [Medline]
  76. Tuti T, Aluvaala J, Chelangat D, Mbevi G, Wainaina J, Mumelo L, Clinical Information Network Group, et al. Improving in-patient neonatal data quality as a pre-requisite for monitoring and improving quality of care at scale: a multisite retrospective cohort study in Kenya. PLOS Glob Public Health. Oct 20, 2022;2(10):e0000673. [FREE Full text] [CrossRef] [Medline]
  77. Ma S, Lawpoolsri S, Soonthornworasiri N, Khamsiriwatchara A, Jandee K, Taweeseneepitch K, et al. Effectiveness of implementation of electronic malaria information system as the national malaria surveillance system in Thailand. JMIR Public Health Surveill. May 06, 2016;2(1):e20. [FREE Full text] [CrossRef] [Medline]
  78. Ng C, Payne AS, Patel AK, Thomas-Mohtat R, Maxwell A, Abo A. Improving point-of-care ultrasound documentation and billing accuracy in a pediatric emergency department. Pediatr Qual Saf. Jul 21, 2020;5(4):e315. [FREE Full text] [CrossRef] [Medline]
  79. Pfaff E, Girvin A, Gabriel D, Kostka K, Morris M, Palchuk MB, N3C Consortium, et al. Synergies between centralized and federated approaches to data quality: a report from the national COVID cohort collaborative. J Am Med Inform Assoc. Mar 15, 2022;29(4):609-618. [FREE Full text] [CrossRef] [Medline]
  80. Kiogou SD, Chi CL, Zhang R, Ma S, Adam TJ. Clinical data cohort quality improvement: the case of the medication data in the University of Minnesota's clinical data repository. AMIA Jt Summits Transl Sci Proc. May 23, 2022;2022:293-302. [FREE Full text] [Medline]
  81. Koo J, Moyer L, Castello MA, Arain Y. Improving accuracy of handoff by implementing an electronic health record-generated tool: an improvement project in an academic neonatal intensive care unit. Pediatr Qual Saf. Jul 10, 2020;5(4):e329. [FREE Full text] [CrossRef] [Medline]
  82. Bhattacharya AA, Allen E, Umar N, Audu A, Felix H, Schellenberg J, et al. Improving the quality of routine maternal and newborn data captured in primary health facilities in Gombe State, Northeastern Nigeria: a before-and-after study. BMJ Open. Dec 02, 2020;10(12):e038174. [FREE Full text] [CrossRef] [Medline]
  83. Manesen R, Mekler KA, Molobi TR, Tyiki AA, Madlavu MJ, Velen K, et al. Data quality assessment of a South African electronic registry for drug-resistant TB, 2015-2016. Public Health Action. Mar 21, 2021;11(1):33-39. [FREE Full text] [CrossRef] [Medline]
  84. Olagundoye O, van Boven K, Daramola O, Njoku K, Omosun A. Improving the accuracy of ICD-10 coding of morbidity/mortality data through the introduction of an electronic diagnostic terminology tool at the general hospitals in Lagos, Nigeria. BMJ Open Qual. Mar 2021;10(1):e000938. [FREE Full text] [CrossRef] [Medline]
  85. Moomba K, Williams A, Savory T, Lumpa M, Chilembo P, Tweya H, et al. Effects of real-time electronic data entry on HIV programme data quality in Lusaka, Zambia. Public Health Action. Mar 21, 2020;10(1):47-52. [FREE Full text] [CrossRef] [Medline]
  86. Tizifa TA, Nkhono W, Mtengula S, van Vugt M, Munn Z, Kabaghe AN. Leveraging phone-based mobile technology to improve data quality at health facilities in rural Malawi: a best practice project. Malar J. Apr 27, 2021;20(1):203. [FREE Full text] [CrossRef] [Medline]
  87. Sinaiko AD, Barnett ML, Gaye M, Soriano M, Mulvey T, Hochberg E. Association of peer comparison emails with electronic health record documentation of cancer stage by oncologists. JAMA Netw Open. Oct 01, 2020;3(10):e2015935. [FREE Full text] [CrossRef] [Medline]
  88. Daniel C, Serre P, Orlova N, Bréant S, Paris N, Griffon N. Initializing a hospital-wide data quality program. The AP-HP experience. Comput Methods Programs Biomed. Nov 2019;181:104804. [CrossRef] [Medline]
  89. Miyoshi NS, Azevedo-Marques JM, Alves D, Azevedo-Marques PM. An eHealth platform for the support of a Brazilian regional network of mental health care (eHealth-interop): development of an interoperability platform for mental care integration. JMIR Ment Health. Dec 07, 2018;5(4):e10129. [FREE Full text] [CrossRef] [Medline]
  90. Qualls LG, Phillips TA, Hammill BG, Topping J, Louzao DM, Brown JS, et al. Evaluating foundational data quality in the national patient-centered clinical research network (PCORnet®). EGEMS (Wash DC). Apr 13, 2018;6(1):3. [FREE Full text] [CrossRef] [Medline]
  91. Njuguna C, Vandi M, Mugagga M, Kanu J, Liyosi E, Chimbaru A, et al. Institutionalized data quality assessments: a critical pathway to improving the accuracy of integrated disease surveillance data in Sierra Leone. BMC Health Serv Res. Aug 07, 2020;20(1):724. [FREE Full text] [CrossRef] [Medline]
  92. Edgerton JR, Herbert MA, Hamman BL, Ring WS. Can use of an administrative database improve accuracy of hospital-reported readmission rates? J Thorac Cardiovasc Surg. May 2018;155(5):2043-2047. [FREE Full text] [CrossRef] [Medline]
  93. Qin Y, Zhou R, Wu Q, Huang X, Chen X, Wang W, et al. The effect of nursing participation in the design of a critical care information system: a case study in a Chinese hospital. BMC Med Inform Decis Mak. Dec 06, 2017;17(1):165. [FREE Full text] [CrossRef] [Medline]
  94. Muthee V, Bochner AF, Osterman A, Liku N, Akhwale W, Kwach J, et al. The impact of routine data quality assessments on electronic medical record data quality in Kenya. PLoS One. Apr 18, 2018;13(4):e0195362. [FREE Full text] [CrossRef] [Medline]
  95. Smith BL, Ramadan M, Corley B, Hablas A, Seifeldein IA, Soliman AS. Measuring the effect of improvement in methodological techniques on data collection in the Gharbiah population-based cancer registry in Egypt: implications for other low- and middle-income countries. Cancer Epidemiol. Dec 2015;39(6):1010-1014. [FREE Full text] [CrossRef] [Medline]
  96. Ahlbrandt J, Henrich M, Hartmann BA, Bundschuh BB, Schwarz J, Klasen J, et al. Small cause - big effect: improvement in interface design results in improved data quality - a multicenter crossover study. Stud Health Technol Inform. 2012;180:393-397. [Medline]
  97. Tuti T, Bitok M, Malla L, Paton C, Muinga N, Gathara D, et al. Improving documentation of clinical care within a clinical information network: an essential initial step in efforts to understand and improve care in Kenyan hospitals. BMJ Glob Health. May 24, 2016;1(1):e000028. [FREE Full text] [CrossRef] [Medline]
  98. Taggart J, Liaw ST, Yu H. Structured data quality reports to improve EHR data quality. Int J Med Inform. Dec 2015;84(12):1094-1098. [CrossRef] [Medline]
  99. Haskew J, Rø G, Saito K, Turner K, Odhiambo G, Wamae A, et al. Implementation of a cloud-based electronic medical record for maternal and child health in rural Kenya. Int J Med Inform. May 2015;84(5):349-354. [FREE Full text] [CrossRef] [Medline]
  100. Siegel EM, Jacobsen PB, Lee JH, Malafa M, Fulp W, Fletcher M, et al. Florida Initiative for Quality Cancer Care: improvements on colorectal cancer quality of care indicators during a 3-year interval. J Am Coll Surg. Jan 2014;218(1):16-25.e1. [FREE Full text] [CrossRef] [Medline]
  101. Knight AW, Szucs C, Dhillon M, Lembke T, Mitchell C. The eCollaborative: using a quality improvement collaborative to implement the National eHealth Record System in Australian primary care practices. Int J Qual Health Care. Aug 2014;26(4):411-417. [FREE Full text] [CrossRef] [Medline]
  102. Soto M, Capurro D, Catalán S. Evaluating the data completeness in the electronic health record after the implementation of an outpatient electronic health record. Stud Health Technol Inform. 2015;216:885. [Medline]
  103. Ewing M, Funk GA, Warren AM, Rapier N, Reynolds M, Bennett M, et al. Improving National Trauma Data Bank® coding data reliability for traumatic injury using a prospective systems approach. Health Informatics J. Dec 2016;22(4):1076-1082. [FREE Full text] [CrossRef] [Medline]
  104. Genet IC, Firestone KS, Volsko TA. Neonatal respiratory therapist-led rounds can improve staff satisfaction and timeliness of respiratory interventions. Respir Care. Mar 2015;60(3):321-327. [CrossRef] [Medline]
  105. Nassaralla CL, Naessens JM, Hunt VL, Bhagra A, Chaudhry R, Hansen MA, et al. Medication reconciliation in ambulatory care: attempts at improvement. Qual Saf Health Care. Oct 2009;18(5):402-407. [CrossRef] [Medline]
  106. Rahbar MH, Gonzales NR, Ardjomand-Hessabi M, Tahanan A, Sline MR, Peng H, et al. The University of Texas Houston Stroke Registry (UTHSR): implementation of enhanced data quality assurance procedures improves data quality. BMC Neurol. Jun 15, 2013;13:61. [FREE Full text] [CrossRef] [Medline]
  107. Greiver CD, Barnsley J, Aliarzadeh B, Krueger P, Moineddin R, Butt DA, et al. Using a data entry clerk to improve data quality in primary care electronic medical records: a pilot study. Inform Prim Care. 2011;19(4):241-250. [FREE Full text] [CrossRef] [Medline]
  108. Wallace CJ, Stansfield D, Gibb Ellis KA, Clemmer TP. Implementation of an electronic logbook for intensive care units. Proc AMIA Symp. 2002:840-844. [FREE Full text] [Medline]
  109. de Lusignan S, Stephens PN, Adal N, Majeed A. Does feedback improve the quality of computerized medical records in primary care? J Am Med Inform Assoc. 2002;9(4):395-401. [FREE Full text] [CrossRef] [Medline]
  110. Mphatswe W, Mate KS, Bennett B, Ngidi H, Reddy J, Barker PM, et al. Improving public health information: a data quality intervention in KwaZulu-Natal, South Africa. Bull World Health Organ. Mar 01, 2012;90(3):176-182. [FREE Full text] [CrossRef] [Medline]
  111. Benard M, Bouaud J, Marsault C, Boudghene F, Carette MF, Séroussi B. Diagnostic imaging requisition quality when using an electronic medical record: a before-after study. Stud Health Technol Inform. 2015;210:364-368. [Medline]
  112. Amoroso CL, Akimana B, Wise B, Fraser HS. Using electronic medical records for HIV care in rural Rwanda. Stud Health Technol Inform. 2010;160(Pt 1):337-341. [Medline]
  113. Johnson KB, Neuss MJ, Detmer DE. Electronic health records and clinician burnout: a story of three eras. J Am Med Inform Assoc. Apr 23, 2021;28(5):967-973. [FREE Full text] [CrossRef] [Medline]
  114. Christoff P. Running PDSA cycles. Curr Probl Pediatr Adolesc Health Care. Aug 2018;48(8):198-201. [CrossRef] [Medline]


DAMA: Data Management Association
DMAIC: define-measure-analyze-improve-control
DQ: data quality
EHR: electronic health record
EMA: European Medicines Agency
EMR: electronic medical record
FDA: Food and Drug Administration
HIS: hospital information system
HL7: Health Language 7
ICD: International Classification of Diseases
ICPM: International Classification of Procedures in Medicine
LHS: learning health care system
MeSH: Medical Subject Headings
MHRA: Medicines and Healthcare Products Regulatory Agency
NICE: National Institute for Health and Care Excellence
OPS: Operationen und Prozedurenschlüssel
PACS: picture archiving and communication system
PDQI-9: Physician Documentation Quality Instrument
PDQI-9: Physician Documentation Quality Instrument
PDSA: plan-do-study-act
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses
RWD: real-world data
RWE: real-world evidence
SQL: structured query language
TDQM: total data quality management
WHO: World Health Organization


Edited by A Mavragani; submitted 11.03.24; peer-reviewed by A Hassan, T Khinvasara, A Wani; comments to author 26.04.24; revised version received 10.05.24; accepted 30.05.24; published 22.08.24.

Copyright

©Anthony Lighterness, Michael Adcock, Lauren Abigail Scanlon, Gareth Price. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 22.08.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.