Advanced search
7,998
Views
6
CrossRef citations to date
0
Altmetric
Editorial

Five good reasons to be disappointed with randomized trials

&

Background

Randomized controlled trials (RCTs) are recognized to exhibit very high levels of evidence, representing a coveted position near the top of the evidence-based pyramid [1]. Both authors of this editorial have been part of small to large-scale RCTs and support the need for this form of research design. Yet, few things annoy us more than the deification that clinicians and selected researchers have given to randomize controlled trials. Yes, RCTs are useful in testing the efficacy and effectiveness of interventions between groups; essentially, identifying which treatment intervention is superior between two or more unique groups [2]. Moreover, RCTs are necessary to reduce bias and confounding and are perceived to yield causal inferences [3]. However (and we can’t emphasize this enough), it is our impression that few understand the noteworthy limitations of RCTs, and even fewer are able to extrapolate how these limitations influence clinical practice. Our experiences with these misunderstandings have prompted us to outline some (trust us, there are more) of the limitations of RCTs, specifically those that might influence clinical practice in an orthopedic setting.

Limitations

Reason One: Right Question-Wrong Design: A common response we hear is the belittling of a given study finding because it didn’t involve an RCT. It is imperative to understand that RCTs are a form of research design and this design is not appropriate for all forms of research needs. For example, diagnostic accuracy studies are best analyzed using a case-based, case-control design. Rare diseases are best studied using case-control designs. If one is looking at predictive analytics then a prospective cohort design is the design of choice [2]. Looking for patterns and effects across different data sources?; a systematic review or a meta-analysis is the design of choice. And although an influential paper from 2004 called for better reporting of harms in RCTs [4,5], an RCT is not the most appropriate study design to truly understand the prevalence of these adverse events [6]. An observational case-cohort design will better reflect the population, prevalence and downstream influence of harms associated with dedicated care processes [7].

Reason Two: The Marginal Patient: Perhaps the most well-known limitation of an RCT is external validity. External validity is the degree to which the conclusions in your study would hold for other persons in other places and at other times. In RCTs, there are unavoidable disparities between the study conditions and populations in comparison to the conditions and populations in which the finding will be inferred [8]. A common assumption is that the findings would be transferable to all patient populations, treatment environments, and cultures. This ‘It-works-somewhere’[9] concept is defined as: projected realism.

In an effort to ‘control’ for confounding variables and increase study power, a homogenous sample of diagnostically uniform patients are included that may not represent the actual demographics and complexity in the clinic. These less simple patients are termed ‘the marginal patients’ because the average patient may or may not respond to a given treatment [1012]. Unfortunately, many of the requirements needed in an RCT to improve internal validity (and control for confounding bias) result in an artificial-like setting that does not closely match a real-world environment [13]. Despite the notable juxtaposition between external and internal validity, many RCTs and observational designs involving similar interventions and participants find similar results [14]. Because RCTs are often exceptionally expensive, authors have recommended different designs, alternative data sources, and unique methodological approaches to identify similar findings (at a reduced cost) [15].

Reason Three: Mixed Treatment Effect- Just because one group reports better outcomes than another group in an RCT, it does not mean that the intervention in the group with better outcomes works for all individuals in that group or future groups [13]. Yes, if one finds differences between two groups, the intervention that is associated with an improved outcome may indeed have higher efficacy (for the group tested). Nevertheless, as most studies demonstrate, some individuals in both groups improve whereas some individuals in both groups do not. An RCT only functions to show whether more people improved in one group versus the other, or ‘who’ (which group) benefits. Why someone improved is not a property of a RCT.

To determine ‘why’ someone improves requires a causal mediation design. Causal mediation analysis identifies potential pathways that could explain why the outcomes were more effective with that intervention [16]. Causal mediation analysis allows an understanding of the roles of intermediate variables that lie in the causal path between the treatment and outcome variables, and allows the clinician to focus on both the mediating and primary (intervention) variables with targeted applications. Additionally, not all patients may be appropriate to a given mix of interventions with similar conditions. Thus, determining an effective treatment mix may provide more clinically useful information as opposed to a single treatment approach that demonstrates an effective average treatment effect [1719]. Sadly, although causal mediation designs are often secondary analyses within an RCT, an RCT in isolation does not provide that information.

Reason Four: Treatment Fidelity: Intervention fidelity refers to the reliability and validity of the clinical interventions that are used in the randomized trial [20]. In other words, fidelity reflects the applicability of the interventions for the condition of interest, whether the interventions are appropriately performed (application, dosage, and intensity) and whether the interventions adequately represent how the intervention is performed in clinical practice. Interestingly, past studies have found that intervention fidelity is consistently either poorly performed, poorly reported or both [21]. Unfortunately, because of the costs associated with RCTs, fidelity is commonly sacrificed. Even pragmatic randomized trials (trials designed to test the effectiveness of the intervention in a broad routine clinical practice) are guilty of limited fidelity in the application of behavioral or exercise-based interventions [20].

Reason Five: Unmeasured Bias: The post-randomization experience is the period that immediately follows individuals’ consent and randomization to one of the treatment groups [22]. Randomization is used to reduce errors, differences in groups, and confounding properties that are unforeseen. The post-randomization experience (‘what happens after the randomization’) can also be a period in which bias may play a notable role. Outside of fidelity and some of the aforementioned items, there are five major considerations involving the post-randomization experience. The Hawthorn effect is a change in behavior of the research subjects, administrators, and clinicians in experimental or observational studies [23]. Patients hold certain beliefs and expectations regarding a treatment that have been shown to influence the outcomes [24]. If the allocated treatment group does not match the patients’ beliefs and expectations then the treatment effect is likely subdued. Personal equipoise exists when a clinician has no good basis for a choice between two or more care options or when one is truly uncertain about the overall benefit or harm offered by the treatment to his/her patient [25]. Mode of administration bias exists when the method of outcomes collection (how outcomes were collected from the research participant) is tainted between clinician and research subject [26]. Lastly, contamination bias occurs when the members of one group in a trial receive the treatment or are exposed to the intervention that is provided to the other group.

To reinforce the influence of the Hawthorne effect and personal equipoise, we provide the following examples. First, provider, health services patterns, and comparison of profession are study foci that are particularly pre-disposed to the Hawthorn effect. Although the studies involve randomizing to control biases, clinician behaviors are likely to change since they know they are being evaluated in a formal study. For example, if you are the prescribing physician in a trial that is examining the negative effects of opioids, you are likely going to prescribe fewer opioids. Personal equipoise toward a particular intervention will unconsciously cause an improve outcome for the treatment of preference. For example, in randomized trials where clinicians preferred a particular treatment approach (despite being randomized between two groups), the preference influenced outcomes in a way that supported their preference [27,28].

Summary

Randomized controlled trials are useful in testing the efficacy and effectiveness of interventions between groups [2]. Understanding their limitations is essential before extrapolation to clinical practice. Other research designs are needed to understand the diagnosis, validity of outcomes, and other important research issues. Participants enrolled in RCTs may or may not adequately represent the full population in which the study is designed to represent. Randomized controlled trials evaluate the effects of treatment at population levels and do not explain why the outcomes were more effective with that intervention [9]. The care provided may or may not reflect what is appropriately provided in clinical practice. And lastly, a biased post-randomization experience is not protected by the initial randomization. Careful controls are necessary at this phase of the trial as well.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

Alternative formats

 

Related research

People also read lists articles that other readers of this article have read.

Recommended articles lists articles that we recommend and is powered by our AI driven recommendation engine.

Cited by lists all citing articles based on Crossref citations.
Articles with the Crossref icon will open in a new tab.