(C) PLOS One
This story was originally published by PLOS One and is unaltered.
. . . . . . . . . .



A retrospective cohort analysis leveraging augmented intelligence to characterize long COVID in the electronic health record: A precision medicine framework [1]

['Zachary H. Strasser', 'Department Of Medicine', 'Massachusetts General Hospital', 'Boston', 'Massachusetts', 'United States Of America', 'Arianna Dagliati', 'Department Of Electrical Computer', 'Biomedical Engineering', 'University Of Pavia']

Date: 2023-08

Abstract Physical and psychological symptoms lasting months following an acute COVID-19 infection are now recognized as post-acute sequelae of COVID-19 (PASC). Accurate tools for identifying such patients could enhance screening capabilities for the recruitment for clinical trials, improve the reliability of disease estimates, and allow for more accurate downstream cohort analysis. In this retrospective cohort study, we analyzed the EHR of hospitalized COVID-19 patients across three healthcare systems to develop a pipeline for better identifying patients with persistent PASC symptoms (dyspnea, fatigue, or joint pain) after their SARS-CoV-2 infection. We implemented distributed representation learning powered by the Machine Learning for modeling Health Outcomes (MLHO) to identify novel EHR features that could suggest PASC symptoms outside of typical diagnosis codes. MLHO applies an entropy-based feature selection and boosting algorithms for representation mining. These improved definitions were then used for estimating PASC among hospitalized patients. 30,422 hospitalized patients were diagnosed with COVID-19 across three healthcare systems between March 13, 2020 and February 28, 2021. The mean age of the population was 62.3 years (SD, 21.0 years) and 15,124 (49.7%) were female. We implemented the distributed representation learning technique to augment PASC definitions. These definitions were found to have positive predictive values of 0.73, 0.74, and 0.91 for dyspnea, fatigue, and joint pain, respectively. We estimated that 25 percent (CI 95%: 6–48), 11 percent (CI 95%: 6–15), and 13 percent (CI 95%: 8–17) of hospitalized COVID-19 patients will have dyspnea, fatigue, and joint pain, respectively, 3 months or longer after a COVID-19 diagnosis. We present a validated framework for screening and identifying patients with PASC in the EHR and then use the tool to estimate its prevalence among hospitalized COVID-19 patients.

Author summary Analyzing long COVID using the healthcare system’s electronic health records presents unique challenges due to variable coding practices by healthcare providers and medical coders. For instance, different providers may emphasize different aspects of a patient’s condition such as shortness of breath versus the underlying cause of the symptom (i.e., COVID-19, congestive heart failure or chronic obstructive pulmonary disease). Additionally, some health records may only hint at new or persistent symptoms through a new prescription, a procedure, or a laboratory order. This complexity was heightened prior to the introduction of the long COVID billing code since there was not a clear consensus for how to code patients with ongoing symptoms. Our study utilized a novel representation learning approach to navigate these challenges. We built models using diverse electronic health record data (diagnosis, medication, procedure, and laboratory orders) gathered from several hospital systems to better identify patients showing potential signs of long COVID. We validated the accuracy of our models by manual patient chart reviews. Using this method, we obtained estimates of hospitalized COVID-19 patients exhibiting dyspnea, fatigue, or joint pain three months post-hospitalization. Our augmented definitions can be used to identify potential long COVID patients from the structured data in the electronic health records.

Citation: Strasser ZH, Dagliati A, Shakeri Hossein Abad Z, Klann JG, Wagholikar KB, Mesa R, et al. (2023) A retrospective cohort analysis leveraging augmented intelligence to characterize long COVID in the electronic health record: A precision medicine framework. PLOS Digit Health 2(7): e0000301. https://doi.org/10.1371/journal.pdig.0000301 Editor: Yuan Lai, Tsinghua University, CHINA Received: December 1, 2022; Accepted: June 16, 2023; Published: July 25, 2023 Copyright: © 2023 Strasser et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Data Availability: The computer scripts used for distributed learning representation and for subsequent analysis are available at https://github.com/rebeccamesa/pascPhen.git. The full patient-level dataset contains sensitive and potentially re-identifiable data. Therefore, it cannot currently be made available directly. Data may be made available to affiliated researchers given the MGB IRB approval. For access to the data, please email [email protected]. Funding: ZSH is supported by National Institutes of Health (NIH) National Library of Medicine (NLM) T15 LM007092. JK is supported by National Center for Advancing Translational Sciences (NCATS) UL1TR001857. KBW is supported by NIH R01 HL151643. YL is supported by NCATS U01TR003528 and NLM 1R01LM013337. DWH is supported by NCATS UL1 TR001998. GSO is supported by NIH grants P30ES017885 and U24CA210967. ZX is supported by National Institute of Neurological Disorders and Stroke (NINDS) R01NS098023 and NINDS R01NS124882. JHH is supported by NCATS UL1-TR001878. HE is supported by National Institute of Allergy and Infectious Diseases (NIAID) R01AI165535 and National Institute on Aging (NIA) RF1AG074372. SM is supported by NCATS UL1TR001857. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Competing interests: The authors have declared that no competing interests exist.

Introduction Persistent physical symptoms lasting months following an acute COVID-19 infection are well known and now widely documented [1–5]. Psychological or cognitive complaints have also been reported during recovery from the SARS-CoV-2 infection [6–8]. These patients have been collectively referred to as having long COVID, post-acute COVID-19 syndrome (PACS), or post-acute sequelae of SARS-CoV-2 (PASC). While the exact definition continues to evolve, there is general agreement that it refers to symptoms that persist or relapse at least 3 months from the onset of acute infection, have an impact on the patient’s life, and are not explained by an alternative cause [9,10]. Several large epidemiology studies have now been published that attempt to quantify the prevalence of long COVID and characterize the etiology [11–16]. While valuable, the insights from these studies rely primarily on analyses of the diagnosis codes from the electronic health records. There are several limitations when exclusively using diagnosis codes for identifying specific patients. This is because the diagnosis codes are not meant to be research quality data, but are instead assigned through the transactional interaction between the healthcare system and patient. The diagnoses only indirectly represent an individual’s actual health [17]. Previous studies have found variable rates of sensitivity and specificity for diagnosis codes to accurately describe symptoms and disease [18–21]. This makes using the diagnosis codes from electronic health records challenging for studying long COVID. Many patients that have a specific symptom may not have it documented as researchers expect, and those that have the diagnosis may not actually have the symptom. Adding to this complexity is that the “U09.9 long COVID” code itself was not introduced until late in the pandemic. If only the “U09.9 long COVID” is used for identification, patients with onset of long COVID early in the pandemic would be missed. There is also growing evidence that there is a spectrum of long COVID [22–24]. The long COVID diagnosis code does not differentiate among the long COVID symptom types. For each of these reasons, a data driven process for selecting codes that is validated for identifying long COVID is needed. This study proposes a framework for developing enhanced definitions for detecting long COVID. We focused on three common and well-known symptoms of long COVID: dyspnea, fatigue, and joint pain [1–5,25,26]. We implemented an augmented intelligence strategy that combines machine learning methodology with clinical knowledge to improve the groups of diagnosis codes representing a specific symptom with additional multi-modal data from the EHR. Then we assessed the quality of the enhanced definition by reviewing clinical notes to estimate the positive predictive value of the new definition. Based on this assessment we then estimated the number of patients previously hospitalized with COVID-19 who are likely to develop long COVID symptoms.

Methods In order to develop our PASC definitions, we utilized EHR data from three academic healthcare systems in the United States that participate in the 4CE consortium [27–29]. Each contributing institution received institutional review board approval for aggregate data sharing. No patient level data were shared outside of the respective institution. We employed a validated machine learning framework for modeling evolving phenotypes, MLHO [30], with proven utility in studying long COVID [31], to enrich an expert-curated definition for each of the three PASC problems (dyspnea, fatigue, joint pain), through a distributed representation learning process (Fig 1). We then evaluated the MLHO-produced representation based on clinical expertise to develop and validate the framework for providing population-level estimates. Our study followed the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) reporting guideline. PPT PowerPoint slide

PNG larger image

TIFF original image Download: Fig 1. The augmented intelligence framework for identifying long COVID patients. https://doi.org/10.1371/journal.pdig.0000301.g001 Data set We leveraged the data and network in the international Consortium for Clinical Characterization of COVID-19 by EHR (4CE) [27]. Members of the consortium use Integrating Biology and the Bedside (i2b2) or Observational Medical Outcomes Partnership (OMOP) platforms to map their data to a common data model. The data are harmonized locally and then shared in an aggregated form for analysis and visualization. Three hospital systems (Mass General Brigham, the University of Kentucky, and the University of Pittsburgh Medical Center) collaborated through the 4CE network to create local data sets for analyzing long COVID. The inclusion criteria were that patients needed a positive SARS-CoV-2 polymerase chase reaction (PCR) test for the first time 7 days before to 14 days after the time of hospitalization. Additionally, the COVID-19 hospitalization needed to occur between March 13, 2020 and February 28, 2021. Both adults and children were considered. There were no exclusion criteria. Each center then extracted its own EHR data elements, including ICD-10 diagnosis codes and pre-specified laboratory tests, medications, and procedure codes (S1 Table) for local analysis. ICD-10 codes from 1 year to 14 days before the COVID admission were grouped with Elixhauser Comorbidity Software Refined using the R package “comorbidity” to determine patient comorbidities. [32] Several of the Elixhauser comorbidities were further consolidated based on clinical similarity (S2 Table). Initial clinical symptom definition and cohort identification First, we grouped International Classification of Diseases, Tenth Revision (ICD-10) codes that best match the symptom of interest. For example, in the case of dyspnea, the definition included all diagnosis codes within the R06 group, which represents “Abnormalities of Breathing”. Previous COVID-19 studies used similar groupings of ICD-10 codes to represent specific PASC symptoms [11]. These initial data elements used to define the symptoms of interest are referred to as the core data elements throughout the manuscript. A complete list of the core data elements for each of the three symptoms can be found in S3 Table. Distributed representation learning and clinical augmentation Next, we used a Machine Learning (ML) approach to identify additional structured data elements that could signal the presence of one of the three symptoms, but were not included as one of the original core features. These newly identified features are referred to as the augmented features. To determine these augmented features each of the three academic sites used their hospitalized COVID-19 patients to develop three different training sets. Each training set focused on a specific symptom. Patients were labeled as having the symptom if the core data element appeared for the first time at least 90 days after a COVID-19 hospitalization, with a look back period of one year before hospitalization. Patients were labeled as a negative case if they had a follow-up appointment at least 90 days after hospitalization and did not have a new core data element. This created three training sets per hospital with positive and negative cases. We then implemented a previously described pipeline for identifying additional features that discriminate between the positive and negative patients in each training sample [31]. First, the core features used to identify the positive cases were removed from the training set. All other diagnosis codes, laboratory orders, medication orders, and procedure codes were considered possible features for discriminating between positive and negative patients. Then features were selected using a sparsity screen (required 2% prevalence), computation of joint mutual information, and boosting to identify those with highest association. Then a 5-fold cross validation is performed (80–20 train-test split) to develop a confidence score based on the number of times a particular feature is used in the model to identify a positive case (S2 Text). The clinical team then reviewed the EHR features identified at all three sites (S4 Table) to assess if they were clinically meaningful (and therefore likely generalizable) and whether their incorporation into the original definition for each PASC phenotype would potentially result in an enrichment of the definition. To standardize and facilitate this process, we developed categories that could explain the underlying association between the identified data elements and the PASC symptom. All of the categories deemed likely to enhance the initial definitions were incorporated into a new definition, which were referred to as the augmented definition. This augmented definition includes both the original core features and the new augmented features. In contrast, the original definitions only have the core features. Validation of proposed model The augmented definitions were then implemented in one of the three healthcare systems to identify the SARS-CoV-2 hospitalized patients most likely to have the specified persistent symptoms three months after the index date (positive PCR test). These patients were subdivided into four distinct groups (S1 Fig) for chart review and validation. Group 1 included those patients with both the core and augmented features. Group 2 included those with exclusively the core features but not the augmented features. Group 3 included those exclusively with the augmented features but not the core features. Group 4 was identified without any of the core or augmented features. Charts were then randomly sampled from each of the groups for review. The charts were examined by clinicians for descriptive language in the clinical note that would confirm the presence of the specific symptom after the COVID-19 diagnosis. Analysis of long COVID subgroups Based on the sampled charts for each of the groups a positive predictive value for each of the feature sets was determined. Then a confidence interval was applied to the local population with the standard formula: where p is the sample proportion, 1.96 is the critical value of the normal distribution for a confidence level of 95%, n is the sample size and N is the population size. The groups were assumed to be independent of each other and were then summed to determine the estimated proportion of hospitalized COVID-19 patients with the specific persistent symptoms at 3 months with 95% confidence intervals. Finally, the augmented definition was used to identify patients with likely long COVID. The identified patients in each of the symptom clusters were examined based on a variety of characteristics including age and comorbidities prior to COVID-19 diagnosis.

Discussion Previous studies have leveraged different techniques for identifying long COVID patients. Pfaff et al. used visits to the long COVID clinic as a proxy for long COVID [33]. However, access to a long COVID clinic remains uneven, and there may be specific patient interactions within a highly specialized healthcare system that enable such access. It is unclear whether the EHR features for identifying these patients would be generalizable to the broader population. As with the diagnosis code, this approach does not differentiate among the types of symptoms that one could develop with long COVID. The Global Burden of Disease Long COVID Collaborators used a variety of different data sources including primary literature and relied heavily on the claims data from two separate networks [11]. However, the claims data analysis still fundamentally relied on a group of experts determining a priori a group of ICD codes that would define a symptom cluster of interest. As with any group using prior knowledge, their identified labels could miss certain patients that didn’t get that specific symptom code. The value in our approach is that we enhance our initial definitions with a distributed learning approach to improve definitions so that they have a higher positive predictive value. These new definitions can then be used for understanding the true prevalence of the disease and analyzing the group of patients most likely suffering from the disease. Accurate tools for identifying such patients could enhance screening capabilities for the recruitment for clinical trials, improve the reliability of disease estimates, and allow for more accurate downstream cohort analysis. This includes the potential to detect rare associations and understand complex non-linear relationships. The results of our analysis identify several important findings. Previous studies have shown that underlying comorbidities increase your likelihood of severe acute COVID-19 [34]. Pfaff et al. also suggested that greater comorbidities before acute COVID-19 contribute to greater likelihood of Long COVID [33]. Hanson et al. did not specifically look at comorbidities but they also identified that more severe COVID-19 infections were likely to lead to long COVID [11]. Our manuscript shows that underlying diseases was associated with increased likelihood of developing long COVID across all three symptom clusters studied. Even in the case of relatively less devastating chronic illness, like hypertension, there is an increased prevalence among those who went on to develop long COVID. This finding has important implications for clinical care. Providers should consider increased vigilance when evaluating patients with multiple underlying comorbidities for the possibility they may have an increased likelihood of suffering from long COVID. It has long been known that patients with comorbidities are at increased risk of severe COVID [35]. Clinicians will need to continue their increased vigilance of patients with co-morbid conditions after the initial COVID-19 infection. Our approach provides a robust and scalable framework for identifying patients with specific PASC subtypes in the EHR. Robustness of this study stems from its integration of clinical knowledge and data-driven discovery using distributed representation learning across multiple health systems, and validation through chart reviews by clinician experts. Scalability of this framework is based on its utilization of widely accessible structured EHR data, which can identify patients with the three PASC subtypes with reasonable accuracy. In the future, these augmented definitions could be applied in other healthcare systems to quickly ascertain persistent symptoms. Limitations As with any EHR study, patients may be missed who go to a provider outside of the specific healthcare system and their ongoing symptoms might not be recorded in the EHR. However, each of the sites are large healthcare networks that include both primary care and tertiary academic medical centers. Additionally, since this study focused on patients hospitalized in three academic medical networks, there may be unique coding techniques among such centers compared to for-profit hospital systems. However, this is unlikely to be a significant limitation as the networks still include both large, specialized academic centers as well as smaller, community hospitals. The patients were still likely treated from a diverse array of providers. This study did only include hospitalized patients and therefore it is unknown how these findings would affect patients treated for their acute infection at home or those who were asymptomatic. Since only hospitalized patients were included, the population is older and sicker than the general population. Additionally, the chart review process was carried out by a single clinician, rather than multiple clinicians, which could introduce some bias. However, despite this limitation, the authors are unaware of other large epidemiological COVID-19 manuscripts that have used structured data and validated this with chart reviews.

[END]
---
[1] Url: https://journals.plos.org/digitalhealth/article?id=10.1371/journal.pdig.0000301

Published and (C) by PLOS One
Content appears here under this condition or license: Creative Commons - Attribution BY 4.0.

via Magical.Fish Gopher News Feeds:
gopher://magical.fish/1/feeds/news/plosone/