Dear Robert Courtney
Thank you for your ongoing and detailed scrutiny of our review. We have the greatest respect for your right to comment on and disagree with our work, but in the spirit of openness, transparency and mutual respect we must politely agree to disagree.
Presenting health statistics in a way that makes sense to the reader is a challenge. Statistical illiteracy is – according to Girgerenzer and co-workers – common in patients, journalists, and physicians (1). With this in mind we have presented the results as mean difference (MD) related to the relevant measurement scales, for example Chalder Fatigue Scale, as well as standardised mean difference (SMD). The use of MD enables the reader to transfer the results to the relevant measurement scale directly and judge the effect in relation to the scale. We disagree that presenting MD and SMD rather than SMD and MD is an important change, and we disagree with the claim that the analysis based on MD and SMD are inconsistent. This has been discussed as part of the peer-review process. Confidence intervals are probably a better way to interpret data that P values when borderline results are found (2). Interpreting the confidence intervals, we find it likely that exercise with its SMD on -0.63 (95% CI -1.32 to 0.06) is associated with a positive effect. Moreover, one should also keep in mind that the confidence interval of the SMD analysis are inflated by the inclusion of two studies that we recognize as outliers throughout our review. Absence of statistical significance does not directly imply that no difference exists.
All the included studies reported results after the intervention period and this is the main results. The results at different follow-up times are presented in the text, but we have only included data available at the last search date, 9 may 2014. When the review is updated, a new search will be conducted to find new, relevant follow up data and new studies. As a general comment, it is often challenging to analyse follow-up data gathered after the formal end of a trial period. There is always a chance that participants may receive other treatments following the end of the trial period, a behaviour that will lead to contamination of the original treatment arms and challenge the analysis.
Cochrane reviews aim to report the review process in a transparent way, which enables the reader to agree or disagree with the choices made. We do not agree that the presentation of the results should be changed. We note that you read this differently.
Regards,
Lillebeth Larun
1. Girgerenzer G, Gaissmaier W, Kurtz-Milcke E, Schwartz LM, Woloshin S. Helping Doctors and Patients Make Sense of Health Statistics. Pyschological Science in the Public Interest, 2008;82):53-96. http://www.psychologicalscience.org/journals/pspi/pspi_8_2_article.pdf.
2. Hackshaw A and Kirkwood A. Interpreting and reporting clinical trials with results of borderline significance. BMJ 2011;343:d3340 doi: 10.1136/bmj.d3340
http://www.bmj.com/content/343/bmj.d3340.longNew interventions used to be compared with minimal or no treatment, so researchers were looking for and finding large treatment effects. Clear recommendations were made because the P values were usually small (eg, P<0.001). However, modern interventions are usually compared with the existing standard treatment, so that the effects are often expected to be smaller than before, and it is no longer as easy to get small P values. The cut-off used to indicate a real effect is widely taken as P=0.05 (called statistically significant).
“Since 2011, PACE trial data has been shared with many independent scientists as part of normal research collaboration, including the internationally respected research organisation Cochrane, which independently validated the findings.”
Do any of our more statistically skilled members think that I'm missing anything of substance here?:
"We disagree that presenting MD and SMD rather than SMD and MD is an important change, and we disagree with the claim that the analysis based on MD and SMD are inconsistent. This has been discussed as part of the peer-review process. Confidence intervals are probably a better way to interpret data that P values when borderline results are found (2). Interpreting the confidence intervals, we find it likely that exercise with its SMD on -0.63 (95% CI -1.32 to 0.06) is associated with a positive effect. Moreover, one should also keep in mind that the confidence interval of the SMD analysis are inflated by the inclusion of two studies that we recognize as outliers throughout our review. Absence of statistical significance does not directly imply that no difference exists."
Do we know the effect of excluding the outliers?bump.
Do we know the effect of excluding the outliers?
IMO, we need Cochrane to do what AHRQ did - reevaluate the evidence after eliminating the Oxford studies because Oxford includes patients with other conditions.
I think it needs to be made clear to Cochrane that they have to have competent assessors. I have tried to do that but have had no feedback. I am not that optimistic that even people like Iain Chalmers understand the problem. The phoney nature of Cochrane Mental Health board needs to be exposed but it may take time to get that into the public consciousness.
How can this be done? I.e. what can we and/or our charities and/or our researchers do?
I think a much more fundamental change is needed. There is nothing wrong with Oxford criteria studies per se. The scientific problem with Oxford studies of exercise therapy like PACE is more subtle and relates to the fact that the criteria will skew the recruitment of patients who have been informed of the nature of the treatment arms.
The more fundamental problem is that the people who have been assessing these trials simply have no understanding of basic trial methodology and reliability of evidence. The reviews need to be done by people who understand trials. The current situation seems to relate to the fact that the Mental Health section of Cochrane was set up by people who do not understand.
The terms "RCT" and "gold standard" seem to get bandied about whilly nilly by people who should know much better, and people who should know much better seem to get duped by it.I think a much more fundamental change is needed. There is nothing wrong with Oxford criteria studies per se. The scientific problem with Oxford studies of exercise therapy like PACE is more subtle and relates to the fact that the criteria will skew the recruitment of patients who have been informed of the nature of the treatment arms.
The more fundamental problem is that the people who have been assessing these trials simply have no understanding of basic trial methodology and reliability of evidence. The reviews need to be done by people who understand trials. The current situation seems to relate to the fact that the Mental Health section of Cochrane was set up by people who do not understand.
If the reviewers understood then ALL the exercise therapy trials would be rejected because none of them are controlled trials and Cochrane reviews require controlled trials. The current reviewers do not understand what a controlled trial is.
I think it needs to be made clear to Cochrane that they have to have competent assessors. I have tried to do that but have had no feedback. I am not that optimistic that even people like Iain Chalmers understand the problem. The phoney nature of Cochrane Mental Health board needs to be exposed but it may take time to get that into the public consciousness.
RCTs are NOT the gold standard. Properly designed double blinded RCTs with objective outcome measures are the "gold standard" (for clinical trials), ignoring for now the "platinum standard" of meta-analyses.The terms "RCT" and "gold standard" seem to get bandied about whilly nilly by people who should know much better, and people who should know much better seem to get duped by it.
I look forward to reading more about this at the appropriate time.There are various initiatives going on but that is all I can say at present.
Quite so. Errors compounding errors. Not so different from the principle of measuring off a set of marks on a piece of wood, say. If you need to make 12 cuts an inch apart, do you measure the next mark from the previous one? Or from the original baseline one? As we all know, if you do the former then preceding errors accumulate into the next, as it implicitly assumes the previous marks are error-free.In short, a metastudy of poor quality data results in a poor quality metastudy.
Cochrane and other guidelines have methods to do this. However they are most notable in the breach, not the adherence to these guidelines. Furthermore they are checklist guidelines, so that if a problem falls outside the checklist it wont be identified. Typically in an evidence based review an investigator (and there are often many) might have hundreds or even thousands of studies to investigate. So they run a fast checklist test. They typically do not do any deep investigating. This is one of the problems with EBM.I've no expertise on such things, but I would have thought a fundamental principle of any metastudy should be reassessing the integrity of the underlying trials (additional independent peer reviewing maybe), and not just blindly trotting out what the original authors reported.
I've no expertise on such things, but I would have thought a fundamental principle of any metastudy should be reassessing the integrity of the underlying trials (additional independent peer reviewing maybe), and not just blindly trotting out what the original authors reported.
That is what I see too.As far as I can see there is a problem with a standardised 'tool' Cochran uses for evidence quality - it is not fit for purpose. Either that or it is not applied.
Prof. Gundersen tweeted about a fresh Cochrane review on CFS and CBT/GET with link to this article, but I can't find the publication date.
Is this brand new?
Larun et al Exercise therapy for chronic fatigue syndrome
AUTHORS' CONCLUSIONS: Patients with CFS may generally benefit and feel less fatigued following exercise therapy, and no evidence suggests that exercise therapy may worsen outcomes. A positive effect with respect to sleep, physical function and self-perceived general health has been observed, but no conclusions for the outcomes of pain, quality of life, anxiety, depression, drop-out rate and health service resources were possible. The effectiveness of exercise therapy seems greater than that of pacing but similar to that of CBT. Randomised trials with low risk of bias are needed to investigate the type, duration and intensity of the most beneficial exercise intervention.