Beyond critical appraisal
- Michael P Lowe, Bradleigh D Hayhow
- Aust Prescr 2006;29:122-4
- 1 October 2006
- DOI: 10.18773/austprescr.2006.076
All doctors require skills to critically appraise medical research. Critical appraisal is important but limited by its focus on the internal logic of research publications. A broader knowledge of the context in which studies are generated is sometimes necessary to understand their conclusions and their implications for clinical practice.
A key foundation of evidence-based medicine (EBM) is that clinicians with appropriate training can critically appraise research papers. Techniques of critical appraisal are taught to students and have been explained in several publications.1,2
Critical appraisal has at least one major limitation - it suggests that by examining the content of publications alone one can assess the truth of their conclusions. The difficulty with this lies in a fundamental distinction between 'validity' and 'soundness'. Validity relates to the methodology used in a study, whereas soundness relates to the truth of the original data and its interpretation. Critical appraisal examines the validity of scientific studies to determine whether the evidence that is cited supports the conclusions, but it is unable to vouch for the soundness of those conclusions.
We tend to rely on researchers' assertions that their data are true. However, medical research occurs in an environment in which there are many conflicts of interest and powerful influences on researchers. We therefore need to know about the context in which evidence is generated before a true picture of the research findings can be made.
Deliberate academic fraud represents one situation in which published data may be misleading. In one study almost 5% of medical authors reported fabrication or misrepresentation of results within the previous 10 years, and 17% of authors personally knew about a case of fraud in the previous 10 years.3
For example, in 1990 Werner Bezwoda appeared to begin trials of high-dose chemotherapy with bone marrow transplantation in high-risk breast cancer patients. His published results showed markedly improved outcomes with this technique and therefore exerted a substantial influence on clinical practice worldwide. In 2000, a site visit to his laboratory revealed that the original results could only have been obtained by fraud.4
Academic fraud cannot necessarily be detected by critical appraisal. It is occasionally revealed by whistleblowers or by the discrepant results of subsequent studies but may take many years to come to light. Researchers do not often wish to repeat previous studies and ethics committees may say it is unethical to do so. Yet our vulnerability to academic fraud can only be reduced by independent corroboration of findings.
When evaluating clinical results one must be careful about the results of inappropriate sub-group analysis. Comparisons of multiple sub-groups can easily result in exaggeration of differences that are found. To avoid such problems, researchers are urged to clearly state their major hypotheses before their study begins.5This requires researchers to be honest about their intentions as well as their data.
An example of inappropriate sub-group analysis occurred in the reporting of the CLASS trial.6This trial was reported as a three-arm trial comparing the effects of celecoxib with two older non-steroidal anti-inflammatory drugs (NSAIDs) over a time period of six months. It showed a decrease in gastrointestinal complications for people treated with celecoxib. These results led to a marked rise in celecoxib prescribing around the world.
One year after the CLASS publication, it was revealed that the original intention of the trial had been very different, with a planned follow-up of 12-15 months, not six months.7The trial had shown no difference in gastrointestinal adverse effects over the longer period, but when results had been restricted to six months a difference had emerged. To the original readers of the CLASS trial, none of this was evident and critical appraisal of the original article could only conclude that celecoxib was beneficial.
Leading medical journals now require all major trials to be registered at their onset and all Australian trials must be registered with the Australian Clinical Trials Registry (www.actr.org.au). However, doctors will continue to be bombarded with information from poorer quality trials in which problems of inappropriate analysis will be undetectable.
When analysing clinical study data all relevant outcomes should be considered, however it may not be clear which outcomes are important. Often clinical trials do not have the statistical power to detect important adverse events.
Rofecoxib was withdrawn after showing an increase in cardiovascular deaths with sustained use. Trials of rofecoxib (such as the VIGOR trial8) had noted but not emphasised this outcome, and attempted to explain it away. As a result, approval of the drug in world markets was based purely on equivalence of pain-relieving effects and decreased gastrointestinal adverse effects. Yet the possibility of adverse cardiac outcomes was apparent to experts soon after the drug's release.9
When trials are stopped early it may also be difficult to assess all relevant outcomes. Rules for stopping trials tend to rely on only one outcome (such as improvements in mortality) and may lead to other outcomes being ignored.
Trials need to ask the right question. Testing a drug against an inappropriate comparator or an inappropriate dose of a comparator can mislead practitioners. While there continues to be a place for placebo-controlled trials, there is no justification for use of 'semi-placebos' such as an inappropriately small dose of a competitor's drug.
A 1994 article entitled 'The continuing unethical use of placebo controls' suggested that wherever an established treatment existed, it should be used in trials in place of a placebo.10Avoidance of placebos began to be seen as an important ethical principle and led to increasing numbers of so-called 'equivalence trials' in which new drugs were shown to be equivalent to older drugs rather than superior to placebos. Such trials may not always be clinically useful, and they assume that the established treatment has previously been shown to be significantly superior to a placebo.11Critical appraisal of any drug trial that is not placebo-controlled must therefore rely on expert knowledge of the evidence for the comparator drug.
An excellent summary of the problems encountered in critical appraisal warns about the issues that arise from 'conflating' trials.12It uses the example of the PROGRESS trial - which purported to show the benefits of ACE inhibitors after stroke.13
In fact, the PROGRESS trial actually shows a benefit from indapamide as a second-line agent, or from combinations of antihypertensives, rather than from an ACE inhibitor alone. Although the problem was noted by the editorial that accompanied the trial14, the result was so obscured within the paper that we believe only expert epidemiologists could come to the correct conclusion.
Evidence-based medicine downplays the role of experts, suggesting that we can all undertake critical appraisal. Yet an expert view of trials such as VIGOR would have differed from that of a general medical reader, not because of differing skills in critical appraisal, but because of a different knowledge of background issues. High levels of expertise in critical appraisal are also required for the interpretation of some trials in which key features may be deliberately hidden.
Until 2003, the Medical Journal of Australia published a series called 'EBM in action' in which the authors attempted to answer clinical questions by using techniques of critical appraisal. At the end of the series the authors appeared somewhat bemused by the reactions they had received:
There was a side effect that we did not anticipate. Content experts often disagreed with the evidence that we found - a collision between the findings of evidence expertise and content expertise. This often spilled over into the columns of the Journal's 'Letters to the Editor', generating about two letters for each 'EBM in action' article.15
This should not have been surprising. The content of the medical literature can really only be interpreted within the context of clinical medicine. Specialists in the field are 'content experts' who are ideally placed to assess the value of trials within this context. For this reason we believe that it is important to continue to emphasise the role of the content expert in augmenting the process of critical appraisal. However, we must be aware that experts may have conflicts of interest or be subject to influences that affect their views.
We believe that clinicians, in addition to paying attention to the method and results sections of a paper, should take note of editorials and any non-biased expert commentary that is available.
Critical appraisal uses techniques for analysing the validity of published evidence, however it is far less attuned to the soundness of that evidence. A solution to this problem is to pay greater attention to the context in which data are generated, but it seems unlikely that this will fall within the scope of most busy practising clinicians.
We believe that some simple rules can help prevent general medical readers from being misled by unreliable evidence. These include:
Clinical Dean, Northern Territory Clinical School, Royal Darwin Hospital, Darwin, Northern Territory
Resident Medical Officer, Kimberley Population Health Unit, Broome, Western Australia