The case for confidence intervals in controlled clinical trials

https://doi.org/10.1016/0197-2456(94)90036-1Get rights and content

Abstract

A statistical wit once remarked that researchers often pose the wrong question and then proceed to answer that question incorrectly. The question that researchers intend to ask is whether or not a treatment effect is clinically significant. The question that is typically asked, however, is whether or not the treatment effect is statistically significant—a question that may be only marginally related to the issue of clinical impact. Similarly, the response, in the form of a p value, is typically assumed to reflect clinical significance but in fact reflects statistical significance. In an attempt to address this problem the medical literature over the past decade has been moving away from tests of significance and toward the use of confidence intervals. Concretely, study reports are moving away from “the difference was significant with a p value under 0.01” and toward “the one-year survival rate was increased by 20 percentage points with a 95% confidence interval of 15 to 24 percentage points.” By focusing on what the effect is rather than on what the effect is not confidence intervals offer an appropriate framework for reporting the results of clinical trials. This paper offers a non-technical introduction to confidence intervals, shows how the confidence intervals framework offers advantages over hypothesis testing, and highlights some of the controversy that has developed around the application of this method. Additionally, we make the argument that studies which will be reported in terms of confidence intervals should similarly be planned with reference to confidence intervals. The sample size should be set to ensure that the estimates of effect size will be reported not only with adequate power but also with appropriate precision.

References (81)

  • KR Murphy

    If the null hypothesis is impossible, why test it?

    Am Psychol

    (1990)
  • J Cohen

    Some statistical issues in psychological research

  • AR Feinstein

    Clinical biostatistics: Stochastic sugnificance, apposite data, and some remedies for the intellectual pollutants of statistical vocabulary

    Clin Pharm Ther

    (1978)
  • SB Friedman et al.

    What's the difference? Pediatric residents and their inaccurate concepts regarding statistics

    (1981)
  • L Kish

    Some statistical problems in research design

    Am Soc Rev

    (1959)
  • N Nelson et al.

    Interpretation of significance levels and effect sizes by psychological researchers

    Am Psychol

    (1986)
  • KJ Rothman

    Significance questing (editorial)

    Ann Intern Med

    (1986)
  • A Tversky et al.

    Belief in the law of small numbers

    Psychol Bull

    (1971)
  • Wonnacott

    Statistically significant

    Can Med Assoc J

    (1985)
  • JA Freiman et al.

    The importance of beta, the type II error and sample size in the design and interpretation of the randomized control trials. Survey of “negative” trials

    N Engl J Med

    (1978)
  • JA Freiman et al.

    The importance of beta, the type II error, and sample size in the design and interpretation of the randomized controlled trial: survey of two sets of “negative” trials

  • DG Altman

    Statistics and ethics in medical research: How large a sample?

    Br Med J

    (1980)
  • J Hartung et al.

    Absence of evidence is not evidence of absence

    Anesthesiology

    (1983)
  • J Kane et al.

    Compliance in the long-term treatment of schizophrenia

    Psychopharmacol Bull

    (1985)
  • WC Phillips et al.

    The significance of “no significance”: What a negative statistical test really means

    Am J Radiol

    (1983)
  • JF Reed et al.

    Statistical proof in inconclusive “negative” trials

    Arch Intern Med

    (1981)
  • M Borenstein et al.

    A visual approach to statistical power analysis on the microcomputer

    Behav Res Meth Instr Comp

    (1992)
  • JK Brewer

    On the power of statistical tests in the American Educational Research Journal

    Am Education Res J

    (1972)
  • LJ Chase et al.

    A statistical power analysis of applied psychological research

    J Appl Psychol

    (1976)
  • J Cohen

    The statistical power of abnormal-social psychological research: A review

    J Abnormal Soc Psychol

    (1962)
  • SN Goodman et al.

    Evidence and scientific research

    Am J Public Health

    (1988)
  • P Sedlmeier et al.

    Do studies of statistical power have an effect on the power of studies?

    Psychol Bull

    (1989)
  • J Cornfield

    A statistical problem arising from retrospective studies

  • J Fleiss

    Statistical Methods in Rates and Proportions

    (1981)
  • C Poole

    Beyond the confidence interval

    Am J Public Health

    (1987)
  • IJ Selikoff et al.

    Asbestos exposure, smoking and neoplasia

    JAMA

    (1968)
  • D Hoffman et al.

    Smoking and occupational cancers

    Prev Med

    (1976)
  • P Cole et al.

    Occupation

  • WR Grace

    Construction Products Division. There Are Some Things You Should Know

    Employee Pamphlet

    (1993)
  • AS Detsky

    Establishing therapeutic equivalency: What is clinically significant difference?

    Ann Intern Med

    (1986)
  • Cited by (71)

    View all citing articles on Scopus

    This work was supported in part by the following grants: NIMH/SBIR, 1-R43/44-MH-43083, and NIMH MH-41960.

    View full text