FPIN Journal Club RANDOMIZED CONTROLLED TRIAL SPEAKER NOTES Title: Surgery for persistent knee pain not so fast. Journal Club Author: Anne Mounsey, MD- University of North Carolina Journal Club Editor: Kate Rowland, MD, MS – Rush-Copley Medical Center PURL Citation: Brown M M, Mounsey A. Surgery for persistent knee pain not so fast. J Fam Prac. 2014;63:534-535. Original Article: Arthroscopic Partial Meniscectomy versus Sham Surgery for a Degenerative Meniscal Tear. N Engl J Med. 2013;369:2515-2524. 1. What question did the study attempt to answer? Patients – aged 35-65 yrs with 3 months of knee pain and a degenerative medial meniscus tear but no arthritis. Intervention – Arthroscopic partial medial meniscectomy with preservation of as much of the normal meniscus as possible. Comparison – Sham meniscectomy using same instruments and suction and took same length of time as real surgery. Outcome – Knee pain after exercise, measured on 11 point (0-10) scale , Knee pain on the Lysholm scale and the Western Ontario meniscal evaluation tool (WOMET), both 100 point scales where lower scores indicate worse pain Did the study address an appropriate and clearly focused question 2. Determining Relevance: a. Did the authors study a clinically meaningful Yes and/or a patient oriented outcome? b. The patients covered by the review similar to your population Although the population is very specific Yes No No Yes No 3. Determining Validity: Study design; a. Was it a controlled trial? Yes No b. Were patients randomly allocated to comparison groups? Yes No Unclear c. Were groups similar at the start of a trial? Yes No Unclear d. Were patients and study personnel “blind” to treatment? Yes No Unclear The surgeon performing the surgery was not blinded but did not participate in the trial or follow up after the surgery. e. Aside from allocated treatment, were groups treated equally? Yes No Unclear f. Were all patients who entered the trial properly accounted for at it’s conclusion Yes No Unclear follow up was 100%! 4. What are the results? a. What are the overall results of the study? There were no between the groups from baseline to 12 months in change in any of the 3 scores. Both groups improved. Between group difference in Lysholm was -1.6 (95% confidence interval [CI] -7.2 to 4.0). WOMET -2.5 (95% CI -9.2 to 4.1) and knee pain -0.1 (95% CI -0.9 to 0.7). No significant difference in sham and real surgery groups in number of patients who thought they had sham surgery, 47% vs 38% (p=0.39). 7% (5) patients in the sham surgery group ended up having surgery. b. Are the results statistically significant? No difference Yes No c. Are the results clinically significant? Yes No d. Were there other factors that might have affected the outcome? Yes No It could be that arthroscopic lavage which both groups received shows a benefit. However, a previous study that evaluated lavage as a treatment for knee pain showed no benefit. 5. Applying the evidence: a. If the findings are valid and relevant, will this change your current practice? Yes No b. Is the change in practice something that can be done in a medical care setting of a family physician? Do not refer for arthroscopy Yes No c. Can the results be implemented? Yes No d. Are there any barriers to immediate implementation? Yes No e. How was this study funded? Privately, no drug industry funding 6. Teaching Points. Power analysis. This was a negative study (i.e.; the results showed no difference in the outcomes between the actual surgery and sham surgery groups). There may have been no difference between the 2 groups due to insufficient participants. When this happens, it is called a type II error (i.e.; the study shows no difference when there really is one). To assess the risk of type II error, we need to ensure the study had enough participants by looking at the power analysis. A negative study with enough people (an “adequately powered trial”) is just a negative study, but a negative study that did not enroll enough people may have committed a type two error, and the results are at risk being invalid. You need to refer back to the section in the paper usually headed “statistical analysis” to search for the power analysis. At the beginning of this section the authors should state, and did in this paper, what the minimal meaningful clinical difference is in the primary outcomes. In this study differences of over 11.5, 15.5 or 2.0 in the Lysholm, WOMET and knee pain scores respectively were considered clinically relevant. A calculation is then done using level of significance, the power and the effect size to estimate how many patients would need to be in each of the intervention and control groups to show a difference for each of the 3 outcomes (i.e.; the sample size). Level of significance: this is the probability cut-off (usually 0.05 or 5%). In this case it means we are 95% certain that there is no true difference between the 2 treatments. A power level of 80% is nearly always used and means there is an eight in ten chance of detecting a difference of the specified effect size (in this case, the 11.5, 15.5, or 2.0 on the knee pain scales above) The effect size is a measure of the difference in the outcomes of the experimental and control groups. It’s a measure of the effectiveness of the treatment, which is calculated from the standard deviation and the absolute difference in outcomes. It is often derived from previous studies looking at similar outcomes, as it was in this case. The values for the clinically meaningful changes on the knee pain scales came from previous studies that asked patients how much their pain had to change before they felt it was really significantly better. In other cases, where no previous information exists, pilot data or the investigators’ best guesses may have to suffice. As it states in this section for the study to be 80% sure that there was a difference between the 2 groups they need 40,54 and 40 in each group for the 3 outcomes. The samples sizes for both groups were over 69 so we know that there were enough participants to show a difference if one had been present. Of course if there had been a difference found it would not be so important for us to refer to the power analysis as it would be inferred that there were enough participants to show the difference in outcomes. Power calculations are usually only calculated for the primary outcome which is why the results from secondary outcomes need to be interpreted with more caution particularly if they show no difference in outcomes. A type I error—a discussion for another PURLs journal club—happens when the study incorrectly detects a difference when one does not really exist.