Statistical methods for reducing bias in comparative effectiveness research when using patient data from doctor visits

LIMITATIONS: While our work developed extensive theory to assess the effects of outcome-dependent visits, many of our results were based on simulation studies in which we introduced known degrees of bias. As with any simulation-based approach, the results cannot be known to apply in full generality....

Full description

Bibliographic Details
Main Authors: McCulloch, Charles E., Neuhaus, John M. (Author)
Corporate Author: Patient-Centered Outcomes Research Institute (U.S.)
Format: eBook
Language:English
Published: Washington (DC) Patient-Centered Outcomes Research Institute [2019], 2019
Series:Final research report
Online Access:
Collection: National Center for Biotechnology Information - Collection details see MPG.ReNa
Description
Summary:LIMITATIONS: While our work developed extensive theory to assess the effects of outcome-dependent visits, many of our results were based on simulation studies in which we introduced known degrees of bias. As with any simulation-based approach, the results cannot be known to apply in full generality. However, the breadth of our simulations, the approximate calculations, and sensitivity analyses all support the primary conclusions
BACKGROUND: Clinical databases, such as those collected in electronic health records systems, are increasingly being used for patient-centered outcomes research. Because the number and timing of visits in such data are often driven by patient characteristics and may be related to the outcomes being measured (which we call an outcome-dependent visit process), the danger is that this will result in biased and misleading conclusions compared with analyses from designed, prospective studies. Most, if not all, of the extant statistical methodology relies on unrealistic models for the outcome-dependent visit process.
We also developed methods to diagnose the outcome dependence and showed that these diagnostic methods have high power to detect outcome-dependent visit processes. Furthermore, high power was achieved using these methods before maximum likelihood-based statistical analysis methods exhibited significant bias. CONCLUSIONS: The results of our research give practical guidance in the validity of inferences from outcome-dependent visit processes. When data are subject to outcome dependence, bias is restricted to a subset of the covariates (those with associated random effects in the outcome model). Standard, maximum likelihood-based methods such as mixed-model regression often exhibited little bias, even for parameters with associated random effects. Generalized estimating equations methods, especially those based on an independence working correlation, were more susceptible to bias. The diagnostic methods we developed have high power to detect outcome-dependent visit processes.
OBJECTIVES: Our goals were to (1) develop realistic models for outcome-dependent visit time models, (2) use theoretical calculations and simulation models to assess bias and efficiency in longitudinal statistical analyses applied to outcome-dependent visit databases, (3) provide guidance as to which types of statistical inferences are accurate and exhibit little bias when using databases collected with outcome-dependent visit times vs those that are likely to be inaccurate, and (4) make recommendations of how to deal with outcome-dependent visit processes in clinical research. METHODS: We used semistructured interviews with clinician-scientists and analysis of their visit pattern data to develop realistic models for the connections between the likelihood of a visit and the outcome process. Using these realistic models, we assessed the performance of standard statistical methods, such as mixed-model regression, as well as analysis methods that purported to deal with outcome dependence.
We used theory and simulations to evaluate the bias as well as the impact of including a small number of regularly scheduled visits. We used a wide variety of sensitivity analyses to determine the generality of the results. Using theory and simulation we also developed and evaluated the performance of methods to diagnose the outcome dependence. RESULTS: Analysis approaches designed to deal with outcome dependence fared no better and, for some methods, worse than ignoring the outcome dependence and using standard, mixed-model analysis methods. The bias using standard methods under outcome-dependent visit processes is mostly confined to covariates that have associated random effects. A wide variety of sensitivity analyses confirm the generality of the results. To address this bias, we showed that inclusion of a few, non-outcome dependent observations can significantly reduce the bias when using maximum likelihood fitting methods.
Physical Description:1 PDF file (123 pages) illustrations