Abstract

On September 29, 2021, the U.S. Food and Drug Administration (FDA) released a draft guidance discussing “considerations when proposing to use electronic health records (EHR) or medical claims data in clinical studies to support regulatory decisions on effectiveness and safety.”1 The guidance is a result of years of FDA consideration of use of real-world data (RWD), including many stakeholder discussions. It addresses the adequacy and relevance of RWD to address specific research questions and validation and data quality/provenance of study question elements.1 In addition, the guidance considers whether follow-up in RWD sources is sufficient to ascertain outcomes and if missing data limit interpretation of results.1-4 Notably, FDA recommendations do not appear to be limited to new product indications or labeling changes, nor to non-interventional studies. Study design and analytic techniques, including handling of confounding will be addressed in forthcoming FDA documents. While this draft guidance moves the pharmaceutical industry closer to the use of RWD for regulatory decisions, it appears to incorporate inefficiencies common in randomized clinical trials (RCTs) into studies using RWD and attempts to be broadly applicable to studies in EHR and claims- and to RCTs and observational studies- without considering important differences between them. Here, we evaluate and provide commentary on the draft guidance with respect to sponsors aiming to use RWD sources to support regulatory decisions and labeling. We focus on key components of the guidance: adequacy and relevance of the data source, missing data, data linkage, and the ascertainment and validation of study variables along with quality assurance (QA). For illustrative purposes, we contrast the recommendations in this guidance for RWD studies to existing requirements for RCTs. While all scientific research should pre-specify objectives, study design, and statistical analysis plans in a (preferably registered) protocol,5 we agree with FDA that studies using claims or EHR should also detail the conceptual and operational definitions of all elements of the research question, possibly using the PICOTS framework as previously suggested.3, 4 As FDA stated, evaluation of adequacy of a data source must be in the context of the research question and regulatory decision (e.g., approval, labeling change, new indication) to be made.2, 3, 6, 7 The FDA could acknowledge that no RWD source is entirely complete with all variables that might be of interest, but RWD can be useful despite some missing data or variables. Lack of less important covariates may not matter and quantitative bias analyses (QBA) and analyses assessing sensitivity of findings to varying operational definitions and assumptions regarding missing data should be encouraged and pre-specified.8-10 For example, acceptance of RWD research findings might depend on whether sensitivity analyses and QBA suggest minimal change when accounting for bias, missing data or varying operational definitions, particularly if the effect size is moderately large. In Section V.C.2 and V.C.3 of the guidance, the FDA appropriately expresses concerns about missing data for prescriptions (e.g., samples, out-of-pocket purchase) and outcomes (e.g., out-of-network care), noting that linkage (discussed here in Section 4) to data sources may enhance capture of these elements.1 The guidance should have noted that additional sensitivity analyses, akin to those in RCTs for missing exposure or outcome data, should be encouraged to understand whether results would have changed if such missing data were available. Section IV.A. of the draft FDA guidance appropriately highlights critical information not in EHR or claims data that pharmacoepidemiologists routinely consider when assessing data sources for specific research questions, such as relevant formulary restrictions, tiering or stepped therapy, and prior authorizations that limit medication use.1 Variation in medical practices, diagnostic criteria, and treatment patterns across health care systems are also acknowledged as important considerations. Because U.S. residents often switch medical insurance with change in employer, continuity of care is critical to allow sufficient longitudinal follow-up (person-time) and capture of study outcomes. Such longitudinal follow-up might not be achievable without linkage to other data sources. To supplement EHR or claims data when key data elements are missing or poorly measured, the FDA guidance recommends linkage to other data sources (see FDA guidance Section IV.B.2). While important to consider, linkage may not always be an option due to lack of linking variables, privacy issues, legal considerations for sharing data, and/or data integrity. Even with linkage, the resulting cohort can be limited due to loss of adequate sample size and potential loss of generalizability. The FDA guidance acknowledges that information for some data elements may be contained in unstructured fields within EHR data (e.g., clinical notes), and recommends thoroughly describing the process for extracting and verifying such data, whether manually or using automated technology. However, this process is not detailed in the guidance. In Section V.C.1, the FDA guidance indicates that the medication exposure definition should include dose, formulation, strength, route of administration, timing, frequency, duration studied (if relevant), and potentially manufacturer, some of which may not be available in EHR or claims databases. FDA should encourage consideration of just how crucial each of these exposure attributes is to the specific research question. FDA pointed out that algorithms to identify exposures, outcomes, and important covariates should be operationally defined by data such as medical diagnoses, procedural codes, or dispensed medications. As with clinical assessments in RCTs, algorithms may not always be accurate, which can lead to misclassification. However, misclassification may or may not lead to biased estimates of treatment effect, and FDA should encourage proactive sensitivity analyses and QBA to inform the potential impact of such biases,8-10 particularly when effect size is modest. The guidance outlines the need to have documentation on all QA procedures during data accrual, curation, and transformation to the final analytic dataset (see FDA guidance Section VI.B). While data curation and provenance are important, it is extremely challenging for sponsors to obtain detailed documentation and verify data against original source records when using commercially licensed databases. Guidance for industry should focus on QA that sponsors can readily influence and document from the time of the receipt of data, and QA procedures should be driven by how potential misclassification, errors and bias might affect the conclusions of the study. Stringent QA, auditing and data provision requirements for all covariates, regardless of their association with outcomes, significantly increase inefficiencies, in contrast to the mandate to FDA to assess ways to use RWD to increase efficiencies of clinical research. Instead, FDA should recommend risk-based monitoring11 and work with data providers to encourage transparency of QA procedures and to replicate analyses. In general, the FDA draft guidance describes QA processes for data accrual and curation that are similar to or even beyond the requirements for primary data collection in phase 3 RCTs supporting product approval (see Table 1). For example, throughout the guidance, FDA refers to complete (100%) verification of all study variables (PICO and covariates) as the “most rigorous approach” to using RWD, acknowledging that it may not be feasible. The use of the term ‘verification’ may be conflated with ‘validation.’ Data verification typically involves confirming the correct data point for each patient (or subset), which is impossible for all variables in any RWD study. Such an approach is not expected for RCTs nor is it performed by the FDA in the Sentinel System to evaluate safety signals. In contrast, validation assesses the accuracy of an operational definition compared to a reference standard to yield test characteristics such as positive and negative predictive value, sensitivity, and specificity, contributing to understanding of potential misclassification. If the reference standard is the medical record, the guidance appropriately recommends blinding data abstractors and adjudicators to medication exposure to reduce bias and using standardized and reproducible abstraction and adjudication processes to minimize intra- and inter-rater error. However, the guidance should clarify the appropriate validation approaches to consider when EHR data are used for the study and the same EHR is the medical record reference standard. Of note is that these same EHR data are often used as the source of data collected in RCTs. The guidance fails to distinguish between or prioritize critical variables in recommending validation. In RCTs, data are typically collected at each clinical site through direct assessment of a patient or from medical records. While procedures are available to ensure the accuracy of these data, risk-based quality assurance procedures drive focus on more critical variables,11 and 100% verification is rare. Achieving 100% accuracy (no misclassification) takes extraordinary effort and expense and may never be achievable. Selecting only patients whose outcome can be ‘validated’ could result in a biased sample of patients with more complete records, decreasing generalizability. Validating each covariate in a study is inefficient, costly, and will yield little that might change the overall interpretation of the study. Instead, sponsors should justify the adequacy of critical study elements (exposure, primary outcomes, and key confounders) and perform analyses to assess the sensitivity of results to varying degrees of potential misclassification. In Section IV.D, the FDA guidance suggests that the performance of algorithms or operational definitions for key variables should be demonstrated using sufficiently large samples, appropriate sampling techniques, and reasonable reference standards, but gives little guidance on how to define a ‘sufficiently large’ sample or the appropriate sampling methods. Appropriate reference standards are also not recommended. FDA states that studies to validate health outcomes of interest can be performed, or a prior study or publication can be used if conducted in a “similar population or data source.” The accuracy of case-identifying algorithms can vary across populations, healthcare settings, coding systems, or calendar time periods, and each of these factors can affect an algorithm's performance and transportability to other health databases. Thus, when considering applying algorithms in a different setting or database, researchers should consider whether the outcome prevalence differs, and whether restrictions on formulary access to medications are comparable between the validation population and the study population. The level of documentation on these points could be clarified. FDA states that validation may be performed in cases when only false-positives are of concern, and in cases and non-cases when false-negatives are also of concern. However, validating non-cases in very large studies of rare outcomes may be impractical; a random sample of non-cases may be more practical.12 The draft guidance outlining specific issues of concern to FDA is a welcome leap forward to carve a path for use of RWD to support regulatory decision-making. Throughout the draft guidance, FDA recommends that Sponsors discuss specific issues with the relevant review division. However, the process and timing for such discussions need explicit definition. While providing a potential path forward, the path is arduous and unlikely to address the current inefficiencies of RCTs. The guidance recommends practices that would significantly limit the ability to use RWD to generate reliable results of effectiveness and safety for regulatory purposes and imposes numerous processes for RCTs on observational studies. The guidance could be enhanced by implementing a risk-based approach to prioritize operational definition validation and QA as well as advising on the types of sensitivity analyses and QBA that would give FDA more confidence in the findings. The pharmaceutical industry and regulators should advocate for more transparency in metrics and data curation procedures from data providers, and for routine agreement to provide patient-level data to FDA to support submissions from around the world. Resources available to assist in the identification of decision-grade, fit for purpose data and methods for RWE studies should be referenced.2, 4, 7, 13-15 Similarly, recommendations for justifying adequacy of unstructured data in RWE studies would be helpful. These additions would better address the congressional charge to FDA to identify ways that RWD can be used to more efficiently and more rapidly generate evidence on product effectiveness and safety. The authors would like to thank Molly Aldridge, MPH for her assistance with careful editing of the Commentary before and after response to reviewer comments, and the reviewers for their thoughtful comments to improve the Commentary. Authors acknowledge additional comments from review by the RWE & Regulatory Decisions workgroup of the ISPE RWE Task Force and ISPE members, particularly Drs. Kim Brodovicz, Cathy Anne Pinto, Magdalene Assimon, Helga Gardarsdottir and Rachel Sobel.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call