(C) PLOS One
This story was originally published by PLOS One and is unaltered.
. . . . . . . . . .



Frequency of multiple changes to prespecified primary outcomes of clinical trials completed between 2009 and 2017 in German university medical centers: A meta-research study [1]

['Martin Holst', 'Quest Center For Responsible Research', 'Berlin Institute Of Health At Charité Universitätsmedizin Berlin', 'Berlin', 'Institute For Ethics', 'History', 'Philosophy Of Medicine', 'Medizinische Hochschule Hannover', 'Hannover', 'Martin Haslberger']

Date: 2023-11

In this study, we observed that changes to primary outcomes occur in 55% of trials, with 23% trials having major changes. They are rarely transparently reported in the results publication and often not visible in the latest registry entry version. More transparency is needed, supported by deeper analysis of registry entries to make these changes more easily recognizable.

We analyzed the complete history of changes of registry entries in all 1746 randomized controlled trials completed at German university medical centers between 2009 and 2017, with published results up to 2022, that were registered in ClinicalTrials.gov or the German WHO primary registry (German Clinical Trials Register; DRKS). Data were retrieved on 24 January 2022. We assessed deviations between registry entries and publications in a random subsample of 292 trials. We determined changes of primary outcomes (1) between different versions of registry entries at key trial milestones, (2) between the latest registry entry version and the results publication, and (3) changes that occurred after trial start with no change between latest registry entry version and publication (so that assessing the full history of changes is required for detection of changes). We categorized changes as major if primary outcomes were added, dropped, changed to secondary outcomes, or secondary outcomes were turned into primary outcomes. We also assessed (4) the proportion of publications transparently reporting changes and (5) characteristics associated with changes. Of all 1746 trials, 23% (n = 393) had a primary outcome change between trial start and latest registry entry version, with 8% (n = 142) being major changes, that is, primary outcomes were added, dropped, changed to secondary outcomes, or secondary outcomes were turned into primary outcomes. Primary outcomes in publications were different from the latest registry entry version in 41% of trials (120 of the 292 sampled trials; 95% confidence interval (CI) [35%, 47%]), with major changes in 18% (54 of 292; 95% CI [14%, 23%]). Overall, 55% of trials (161 of 292; 95% CI [49%, 61%]) had primary outcome changes at any timepoint over the course of a trial, with 23% of trials (67 of 292; 95% CI [18%, 28%]) having major changes. Changes only within registry records, with no apparent discrepancy between latest registry entry version and publication, were observed in 14% of trials (41 of 292; 95% CI [10%, 19%]), with 4% (13 of 292; 95% CI [2%, 7%]) being major changes. One percent of trials with a change reported this in their publication (2 of 161 trials; 95% CI [0%, 4%]). An exploratory logistic regression analysis indicated that trials were less likely to have a discrepant registry entry if they were registered more recently (odds ratio (OR) 0.74; 95% CI [0.69, 0.80]; p<0.001), were not registered on ClinicalTrials.gov (OR 0.41; 95% CI [0.23, 0.70]; p = 0.002), or were not industry-sponsored (OR 0.29; 95% CI [0.21, 0.41]; p<0.001). Key limitations include some degree of subjectivity in the categorization of outcome changes and inclusion of a single geographic region.

Clinical trial registries allow assessment of deviations of published trials from their protocol, which may indicate a considerable risk of bias. However, since entries in many registries can be updated at any time, deviations may go unnoticed. We aimed to assess the frequency of changes to primary outcomes in different historical versions of registry entries, and how often they would go unnoticed if only deviations between published trial reports and the most recent registry entry are assessed.

In this study, we aimed to provide an in-depth analysis of changes to primary outcomes using a comprehensive nationwide sample over 9 years that represents the current practice of trial registration in academic research. Specifically, we aimed to assess how often primary outcomes of clinical trials were changed (1) across all clinical trial registry entry versions and (2) between the latest registry entry and the results publication; how often (3) changes are not detectable when only the latest registry entry version would be inspected, and (4) how often they are transparently reported in the results publications; and (5) which trial characteristics are associated with these outcome changes.

We systematically searched studies that have assessed the frequency of outcome changes between registries and publications [ 7 ]. We identified several analyses which widely agree that this practice is common, with a median estimated prevalence of 31% (interquartile range (IQR): 17–45%) of clinical trials affected [ 8 ]. However, most studies did not report which registry entry version they used [ 9 – 16 ], or reported that they only assessed changes between the publication and one single registry entry version (i.e., the latest available entry [ 17 – 25 ], the latest entry before trial completion [ 26 ], the first available entry [ 27 ], or the first entry during the active phase [ 28 ]). We are aware of only two studies that examined changes to primary outcomes over the course of the entire registry version history. Both used relatively small and selective samples (either trials published in ICMJE journals or trials registered on ClinicalTrials.gov covered by the Food and Drug Administration Amendments Act) over a four-year time period [ 29 , 30 ]. We are not aware of studies quantifying the frequency of changes that may go potentially unnoticed because the latest registry entry and the results publication match.

Many clinical trial registries, such as ClinicalTrials.gov, allow entries to be updated after initial registration. Updating a registry entry is a useful tool, as parts of a trial protocol often change and it is desirable for an entry to reflect the most current information. It still constitutes another possible source of reporting bias if entries are not properly maintained and larger changes go unreported. While a number of ethical and reporting guidance documents recommend prospective trial registration [ 2 , 4 , 5 ] and some recommend the transparent reporting of protocol changes [ 2 , 5 ], this is only checked in a minority of cases by peer-reviewers [ 6 ]. But even if trial registries were considered during review, looking only at the latest registry entry version may lead to an incomplete assessment, as changes over the course of the trial would remain unapparent if the latest registry entry matches with the results publication. These changes would only be discovered by closely investigating the history of changes in trial registries, which some trial registry platforms maintain. These histories, however, are often not easily accessible.

Clinical trial registries are tools to enhance the transparency and trustworthiness of clinical trial evidence by openly reporting key study parameters in a rapidly accessible manner [ 1 ]. They make it possible to detect selective reporting biases that pose a threat to validity, by allowing reviewers to assess possible deviations from an original protocol and analysis plan. One of the most critical deviations is a change to prespecified outcome measures (or ‘outcomes’, for short). Consequently, outcomes and their pre-specification are among the most important reporting items of clinical trials, with the CONSORT statement asking authors to describe ‘completely defined pre-specified primary and secondary outcome measures, including how and when they were assessed’ as well as ‘any changes to trial outcomes after the trial commenced, with reasons’ [ 2 ]. If readers do not know whether there have been any changes to the outcomes, e.g., by adding or dropping primary outcomes, they cannot assess the risk of bias due to selective reporting of only statistically significant results or multiple hypothesis testing [ 3 ].

We report descriptive statistics for the frequency of different types of within-registry outcome changes (between key trial milestones) and registry-publication outcome changes. For percentages within the random sample of 292 trials with manual assessment of results publications, we report 95% confidence intervals (CI). For the analysis of determinants of within-registry outcome changes or registry-publication outcome changes, we used logistic regressions, with either any within-registry outcome change, or any registry-publication outcome change, as output variables, and a set of 9 candidate input variables (study phase, sponsor, publication year, registration year, medical field, registry, multicenter trial, enrollment, intervention). These variables were prespecified before the start of regression analyses (details in S1 and S2 Tables). We used the PRISMA statement [ 42 ] to structure our manuscript ( S1 PRISMA Checklist).

The overall inter-rater agreement before discussion, over all timepoints and items (measured by Cohen’s kappa), was κ = 0.81 for the within-registry change ratings and κ = 0.46 for the registry-publication change ratings. Disagreement resulted partly from cases without clear identification of the primary outcomes in results publications (if only the explicitly mentioned outcomes were used for the registry-publication change ratings, inter-rater agreement was κ = 0.51). In a post-hoc sensitivity analysis we restricted our analysis to primary outcomes that were explicitly named as such in the results publications or had been used for sample size calculation.

For feasibility, we drew a random sample of 300 trials with accompanying results publications, from which three reviewers (MH, MH, SY) extracted primary outcomes. Eight articles were excluded from the sample because they were not the main results publications (i.e., the final sample size was 292). We considered outcomes primary if they were explicitly named as such using a list of keywords (‘primary’, ‘main’, ‘outcome’, ‘endpoint’, ‘end point’). Otherwise, we used the outcome used for sample size calculation, if reported. In all other cases, we used the first reported outcome in the abstract and results section (with priority given to the abstract). For each entry, reviewers compared the extracted primary outcome(s) to the latest available registry entry version using the developed rating system ( Fig 1 and Table 1 ). Reviewers also recorded whether there were any mentions of outcome changes in the assessed results publications. Conflicts were resolved through discussion between reviewers, with one additional reviewer (LGH) resolving conflicts, where necessary.

Registry entry versions at four subsequent key trial timepoints were assessed (blue box). Changes can occur between registry entry versions (within-registry changes; red arrows) and between the latest registry entry and results publication (blue arrow). In the lower part, examples of primary outcome descriptions are given (pink boxes), with corresponding categorizations of changes in grey next to the arrows (see Table 1 for more details on categorization).

Two reviewers (MH, MH) assessed within-registry changes in duplicate and resolved conflicts through discussion, using the Numbat Systematic Review Manager [ 39 ]. Where necessary, a third reviewer (LGH) resolved conflicts. Reviewers assessed changes to the primary outcomes between the four key trial milestones within the registry ( Fig 1 ), using a classification system that categorizes changes to primary outcomes as either major changes (based on an existing classification system [ 40 ]) or minor changes (adapted from [ 20 , 41 ]). Major changes were defined as adding or dropping primary outcomes, changing them to secondary outcomes, or turning secondary outcomes into primary outcomes. Minor changes were defined as changes to the outcome’s type of measurement, metric or method of aggregation, or timing of assessment, or as either added specificity or omissions to these (for more detail see Table 1 ).

Trial identifiers (NCT or DRKS number) and basic information about the trials and their accompanying results publications were included in the dataset [ 31 , 32 ]. We then retrieved all historical registry entry versions of these trials, using R 4.1.3 [ 35 ] and the cthist R package [ 36 ], on 24 January 2022 (DRKS) and 12 April 2022 (ClinicalTrials.gov). We extracted data from the fields listed in the codebook [ 37 ]. For our classification of within-registry changes, we extracted primary and secondary outcomes at four key trial milestones: (1) the first entry after study start (start date; inclusion of the first patient), (2) the latest entry before the end of active status (completion date; determined by ‘completed’ status in the registry), (3) the latest entry before results publication (publication date), and (4) latest available entry.

We included any study that fulfilled all of the following criteria: 1) has a registry entry in either the ClinicalTrials.gov or the DRKS database, 2) was completed between 2009 and 2017 according to the registry trial status described as ’Completed’, ’Unknown status’, ’Terminated’, or ’Suspended’ (ClinicalTrials.gov), or ’Recruiting complete, follow-up complete’, ’Recruiting stopped after recruiting started’, or ’Recruiting suspended on temporary hold’ (DRKS), 3) reported in the registry that a German university medical center was involved (i.e., mentioned as responsible party, lead/primary sponsor, principal investigator, study chair, study director, facility, collaborator, or recruitment location; for definitions see [ 31 , 32 ]), 4) has published results, i.e., a full-text publication of the trial results was found by the search methods described in the protocol for the underlying datasets [ 31 , 32 ], 5) is registered as a randomized trial.

We based our analyses on previous work of our group that provided the highly granular and large-scale data on both registry entries and accompanying results publications required to address our aims. Two large datasets [ 31 , 32 ] of randomized controlled trials covering a period of 9 years of trial completion plus a period of 5 years of publication tracking were used for this project. The trials were completed at German university medical centers between 2009 and 2017 and had been registered in either ClinicalTrials.gov or the German World Health Organization primary registry (German Clinical Trials Register or Deutsches Register Klinischer Studien; DRKS). Both registries have an accessible history of changes. The corresponding results publications were identified in manual searches up to 2022, ensuring an at least five-year period after trial completion and complementing the often inconsistent references in the trial registry [ 33 ]. We retrieved a combined dataset for both projects from a GitHub repository [ 34 ] on 24 January 2022.

Of all 161 trials with any change of a primary outcome (i.e., 120 trials with easily detectable changes between the latest registry entry and the publication and 41 trials with more unapparent changes before), 2 trials reported this change in the results publication (1%, 95% CI [0%, 4%]). One of the trials, which exhibited a major change both within the registry (between completion and publication date) and between the latest registry entry version and the results publication, stated that some of the primary outcomes were going to be reported in a separate manuscript [ 43 ]. The other publication, which had a major change between the latest registry entry version and the results publication, stated that the changed outcome was the ‘more sensitive and direct measurement’ [ 44 ].

Registry entry versions were assessed at four subsequent key trial timepoints (within-registry changes; red arrows) and between the latest registry entry and results publication (blue arrow). Rows A to K show each observed combination of outcome change at the different timepoints across the 292 trials. For example, Row A indicates that 131 of 292 trials had no outcome change (45%). Row B indicates that 95 of 292 trials only had a change between the latest registry entry version and the results publication (blue circle; 33%). Changes in rows C, D, H, I and J only occurred within registry entries (red arrow, red circle) while there was agreement of the latest registry entry and the publication (no blue circle), i.e., these changes would not be detected by comparing the results publication to the latest available registry entry (41 of 292 trials, 14%). Row E, F, G and K indicates trials that had a change between the latest entry and the published paper, but also a change at earlier timepoints.

Among the 292 trials, 41 trials had changes only within the registry histories while the latest registry entry matched the results publication (14%; 95% CI [10%, 19%]). Of these changes, 13 were major (4%; 95% CI [2%, 7%]) and 35 minor (12%; 95% CI [8%, 16%]). Changes between the latest registry entry and the results publication, but also within the registry were found in 25 trials (9%; 95% CI [6%, 12%]), of which 12 (4%; 95% CI [2%, 7%]) had major and 14 (5%; 95% CI [2%, 7%]) had minor changes. Overall, 161 trials had any change to their primary outcomes after study start (55%; 95% CI [49%, 61%]), with 67 having major changes (23%; 95% CI [18%, 28%]) and 110 having minor changes (38%; 95% CI [32%, 44%]) (Figs 3 , S2 , and S3 ).

Of 1746 trials, 393 (23%) had a primary outcome change between trial start and latest registry entry version, i.e., within registry entry versions between key trial milestones ( Table 3 and S1 Fig ). Minor changes were found in 318 (18%) trials. 142 trials (8%) had major changes, i.e., primary outcomes were added, dropped, changed to secondary outcomes, or secondary outcomes were turned into primary outcomes. Of these major changes, 66 (4%) happened between first patient inclusion (start date) and trial completion date, 49 (3%) between completion and publication date, and 36 (2%) between the publication date and the latest registry entry.

Discussion

Our assessment of all 1746 published clinical trials conducted at German university medical centers over a period of 9 years shows that changes to primary outcomes are common. About one in four trials (23%) have outcome changes within the registry after study start; a major change occurs within one in twelve trials (8%). Of the 292 assessed results publications, 41% have changes between the latest registry entry and the outcomes presented to readers, and 14% of the trials have no easily detectable change, as the changes occurred between different historical versions of the trials’ registry entries while the latest entries match the results publications. These changes were minor in 12% and major in 4% of trials. An inspection of only the latest registry entry version alongside the results publication, without assessment of the entire registration history, would miss a primary outcome change in these trials.

We used a comparatively large sample of German clinical trials registered between 2009 and 2017, with a validated set of accompanying results publications for registry entries, determined by an extensive manual screening process [31, 32]. Our methodology allowed us to detect changes to registered primary outcomes that might have gone unnoticed by other studies, by retrieving all historical versions for trials, and detecting changes using a streamlined workflow. This is different from previous studies, which instead used manual searches [29, 30]. Many trials in our sample had changes to the registry after study completion (9%) or after publication of the first results publication (8%). The changes, for which there may be very plausible reasons, were almost never reported in the results publications. We revealed potential indicators for these practices, with trials that were registered later, trials not registered on ClinicalTrials.gov, and non-industry-sponsored trials having lower odds of within-registry outcome changes. The observation that the frequency of outcome changes declined over time may be explained by recent discussion on clinical trial reporting practices. Regarding the type of registry, the association might be explained by better usability of the ClinicalTrials.gov platform, which might facilitate more frequent changes to registry entries. Regarding sponsorship, our results could indicate better reporting and higher transparency in industry-sponsored trials (which change their registry entries more often than non-industry-sponsored trials). We currently have no sufficient data to quantify the role of selective reporting bias in this sample.

Assessments of outcome changes that rely on a single version of registry entries might underestimate the true prevalence of changes. The ICMJE policy states that any changes to the registration should be explained by authors in a publication [5]. Both the observed lack of reporting on outcome changes in results publications, as well as survey results among manuscript reviewers [6], indicate that reviewers check the registry entry of a clinical trial only in the minority of cases. While registries make change histories available and offer tools to compare different registration versions, and journal editors and reviewers might in some cases use these tools to check the registration histories, the low number of reporting of changes to primary outcomes in the results publications suggests that this is not properly enforced. At the same time, in their current form, historical versions of registry entries are not easily accessible and readable, which makes checking them a time-consuming task. Journals could implement editorial policies that require specialized personnel to assess the historical versions of submitted clinical trials. Clinical trial registries, on the other hand, could implement solutions to identify and mark trials with major changes to their outcomes.

Still, some trials change their outcomes at two or even three different timepoints over the course of the trial. Interestingly, we found not only cases in which primary outcomes gained more detail, but also cases in which primary outcomes were described with less detail than before. Overall, we found 55% of trials to have some form of outcome change over the course of the study. This is somewhat higher compared to other studies, which exhibit a large range for the frequency, with a median of 31% (IQR: 17–45% [8]). This might be due to different methodology, samples, timepoints, or definition of outcome changes. Chan et al. [45], for example, used a more narrow definition of outcome changes (roughly corresponding to our ‘major’ category). We used a relatively broad definition of primary outcome changes, defining changes to the specificity of primary outcomes as ‘minor’ changes. Our broader definition and inclusion of more ‘minor’ outcome changes, some of which might not constitute a large threat to a trial’s validity, might have led to our higher estimates.

This study has several limitations. First, we rely on the reporting quality of published studies. In 48 out of 292 publications, the primary outcome was not explicitly described in the results publication, so we had to select the outcome used for sample size calculation or the first reported outcome instead. A sensitivity analysis excluding these 48 studies supported our main analysis. Second, we only assessed reporting of outcome changes in the main results publications of the respective trials. These changes could also have been published in a protocol or methods paper. We assessed the results publications as they have higher impact, and because current reporting guidelines such as CONSORT state that deviations from the protocol should be reported in the results publication. Third, while we have information about changes that occurred post-completion, we have no data regarding changes in relation to the clinical trial’s data lock, that is, the timepoint at which no further information is added to the dataset (which might lie after the completion date). Fourth, the categorization of outcome changes is to some degree subjective, and while the agreement was very high for the within-registry changes (κ = 0.81), it was moderate (κ = 0.46) for the registration-publication ratings due to the added complexity of identifying the correct outcome as ‘primary’ and different wording and descriptions between registry and publication. However, conflicting primary outcome definitions and conflicting ratings were resolved in discussions between the reviewers (and, where necessary, a third reviewer), and excluding unclear situations in the above-mentioned sensitivity analysis did not change the main findings. Still, this could become an issue in practice if, for example, different peer-reviewers disagree in their assessment of whether an outcome change has happened or not. Fifth, although we used high quality underlying data reflecting German clinical research output, we only assess a single geographic area, partly reflecting EU/German registration policies. Still, many registration policies are international, and our results point to the possibility that technical details of the registries rather determine whether outcome changes occur. Finally, we only looked at changes between registry entry versions but did not assess how many outcome measures remain underspecified (e.g., no information on the type of measurement, timing, or method of aggregation).

Our analysis also demonstrated the feasibility of an efficient workflow [46] that can be used for future projects and overcome previously described challenges to incorporate historical registry data [47].

To conclude, primary outcome changes are not uncommon in clinical research. Such changes are, in one of four cases, not directly visible upon first inspection of the trial registries and almost never reported in publications. While changes to primary outcomes are not unusual in clinical trials, may be well justified and sometimes desirable, undisclosed changes are a problem. While the problem seems to be ubiquitous, it may be more common in older trials, trials registered in ClinicalTrials.gov, and industry-sponsored trials. Even though trial registries can be an important tool to detect outcome changes, review procedures employed by journals seem to rarely make use of them. Our approach provides a feasible workflow to further investigate this issue. Overall, it is not sufficient to only assess the publication and the latest registry entry–a careful assessment of the full research conduct from conception to publication is needed to ensure trustworthy evidence for clinical decision-making.

[END]
---
[1] Url: https://journals.plos.org/plosmedicine/article?id=10.1371/journal.pmed.1004306

Published and (C) by PLOS One
Content appears here under this condition or license: Creative Commons - Attribution BY 4.0.

via Magical.Fish Gopher News Feeds:
gopher://magical.fish/1/feeds/news/plosone/