Televised medical talk shows—what they recommend and the evidence to support their recommendations: a prospective observational study
BMJ 2014; 349 doi: https://doi.org/10.1136/bmj.g7346 (Published 17 December 2014) Cite this as: BMJ 2014;349:g7346All rapid responses
Rapid responses are electronic comments to the editor. They enable our users to debate issues raised in articles published on bmj.com. A rapid response is first posted online. If you need the URL (web address) of an individual response, simply click on the response headline and copy the URL from the browser window. A proportion of responses will, after editing, be published online and in the print journal as letters, which are indexed in PubMed. Rapid responses are not indexed in PubMed and they are not journal articles. The BMJ reserves the right to remove responses which are being wilfully misrepresented as published articles or when it is brought to our attention that a response spreads misinformation.
From March 2022, the word limit for rapid responses will be 600 words not including references and author details. We will no longer post responses that exceed this limit.
The word limit for letters selected from posted responses remains 300 words.
I thank Dr. Korownyk for her effort in addressing my concerns raised in this platform and correcting my misunderstanding the objectives of your study. I have had no communications with any of these shows - now or ever, and do not appreciate the accusatory tone in Dr. Korownyk's response - which seems to be her general style throughout each one of her responses. I, as a concerned scientist was simply stating my thoughts about the dangers of disseminating flimsy research. I was not defending any of these shows and am glad if they are harming the public, it should be exposed. I main objective was not directed to the intention, but to the misleading scientific validity of the study itself.. In her defensiveness, Dr. Korownyk seemed to have misunderstood my motivations.
We have already seen how a poorly done vaccine study help to fuel an anti-vaccine movement. When a study receives so much public attention, it should be expected that it will receive microscopic inspection. I believe I was quite unbiased in my comments. Dr. Korownyk's team's concerns may be well founded, but that does not mean that we should rush to publishing our beliefs as sound scientific evidence. After all, we mismeasured men's head and attributed its large size to his intelligence. History has shown many times how easy to design studies confirming our biases. Dr. Korownyk herself suggests that reanalyzing their data can lead to conforming others' own biases!!! How can an objectively obtained data be analyzed differently by another group of scientists? Isn't reproducibility one of the cornerstones of doing good science?
Sincerely,
Competing interests: No competing interests
I appreciate the response from Dr. Korownyk and, while I continue to disagree on the science, I agree with her advice to not belabor this debate on these pages. My concerns with her methods and conclusions, which persist, are documented sufficiently in my lengthy Rapid Responses. However, I feel compelled to defend myself against her suggestion that I am not a “trusted professional” and her unfounded claim that I had a conflict of interest in reviewing her study. I am an elected member of the Institute of Medicine and have published more than 170 peer-reviewed papers. My publications (especially in my early career) reveal extensive writing on the proper methods of critical appraisal of scientific evidence and evidence-based practice guidelines, but I have never written about talk shows or television doctors to justify a claim of a conflict of interest. Throughout my entire career I have steadfastly refused funding from pharmaceutical companies or any other industry, nor did I receive a penny to review Dr. Korownyk’s article. Her claim that I have an “association with the Dr. Oz show” is false and asserted without evidence. I know Dr. Oz professionally through a mutual colleague and he shared Dr. Korownyk’s data file after asking me to conduct an independent review of the study. I was informed that Dr. Korownyk had provided the data file for independent review, a customary practice among the researchers I associate with, but I was never informed of Dr. Korownyk’s request that it “not be disseminated further without permission.” I agree with Dr. Korownyk that there should be room for open scientific discourse, but I see no need for the pages of the BMJ to host ad hominum accusations, whether it be the use of arguable analytic methods to assail television doctors or the use of speculation to challenge the integrity of those who question the science behind those analytic methods.
Competing interests: No competing interests
In the June 9 rebuttal by Dr. Korowynk, substantive criticisms of her work made by objective reviewers continue to be ignored.
The accuracy of a survey’s conclusions are dictated by the quality of its raw data and methodology. To that end, The Dr. Oz Show, doctors, and journalists officially requested the 80 recommendations and searchable questions from which the authors’ assumptions were drawn. The questions were ultimately sent to members of The Dr. Oz Show, who performed an internal analysis.
After identifying concerns during the Show’s initial review, we sought expert advice and shared the data privately with four well-respected researchers. We provided Dr. Korowynk with the names of all four experts but felt no obligation to ask permission to complete our review process since this was the original purpose of obtaining the raw data. We hoped to foster a private and honest dialogue. In that spirit, the reviews were sent privately to the BMJ editorial leadership; however, the editors directed that the reviews be posted publicly in the response forum.
Unfortunately, the discourse has become enflamed with unwarranted personal attacks that have been launched against these four experts and several additional reviewers. The Doctor Oz Show has never met, nor had interaction with many of the contributors to this BMJ discussion, including Dr. Tezgin who was singled out by Dr. Korowynk. Dr. Woolf, one of the reviewers – an expert in critical appraisal and former North American editor for the BMJ – informed the editorial leadership in writing that he was approached for an objective analysis of the review by Dr. Oz. Dr. Woolf had no knowledge of raw data privacy concerns, no prior interaction or association with the show, and voluntarily reviewed the work to provide his expert opinion on the matter.
Intriguingly, Dr. Korowynk’s latest note claims that she "offered full disclosure of our data to the BMJ to verify any aspect of our research,” implying that the BMJ’s original reviewers of this manuscript never saw the raw data. The subjective nature of the analysis would demand a review of the original statements made on the show in order to assess the validity of the effort. Ironically, the Sackett paper, which was used by Dr Korowynk to criticize Dr. Woolf, lists all of the conditions and interventions analyzed so that readers can judge for themselves.
As requested, the 80 questions have not been shared publicly, although it is the belief of many that academic integrity would dictate this transparency. After all, the only four reviewers who did have access to the raw data do not believe this paper worthy of the attention it has generated in the media. Due to what we believe to be misleading statements that have raised substantive concerns, it is our belief that an independent analysis should be performed immediately. Specifically, there are erroneous statements in the original manuscript that have been raised by reviewers and need to be addressed:
First, there were never any undisclosed conflicts of interest for any of The Dr. Oz Show content analyzed, despite a claim to the contrary. The United States Federal Trade Commission mandates that all broadcast network television shows, including The Dr. Oz Show and The Doctors, disclose conflicts if they exist. It is illegal to do otherwise.
Second, Dr. Hernandez reported that 22 of 80 show statements were abstracted incorrectly, therefore unable to be accurately judged. Other reviewers had similar concerns and the original BMJ reviewers might have had similar observations if they had been allowed access to this raw data.
Third, only three recommendations were felt to be contradicted by the literature as opposed to the author’s report of 12. This four-fold discrepancy between experts skilled in this process further raises concerns with the methodology used.
Finally, the methods chosen for this study demand a more humble conclusion than the one that currently exists. This sentiment was acknowledged by the authors in earlier comments expressing surprise at the anger aimed at The Dr. Oz Show based on the paper.
The aforementioned issues warrant a re-review and probable revision to the original manuscript. An additional point has been echoed by others in peer reviewed editorials who believe that the authors should amend their conclusion to respect the reality of what these television shows are accomplishing – making complex health insights accessible to a large audience1. Health advice on television often emanates from guidelines provided by government and professional organizations who are trying to close the gap between where peer-reviewed data ends and the needs of our citizens start. These “best practice” recommendations should not be dismissed completely nor always viewed “skeptically” by the viewers, as these suggestions can empower individuals to achieve a healthier daily lifestyle with more informed decision-making.
1 Katz DL Doctors, TV, and Truth: Evidence in the Realm of Edutainment. J Public Health Management Practice, 2015, Published online ahead of print. doi: 10.1097/PHH.0000000000000239
Competing interests: No competing interests
In February 2015, in response to a request from The Dr. Oz Show, we provided the show’s producer with the recommendations and the clinical questions created from our evidence review. With this data they had enough information to perform their own systematic evidence review of the shows’ recommendations. This information was provided with the following statement that “As per standard protocols in the sharing of research data, we would trust that only your office would use this information and it may not be disseminated further without our permission.”
The show then shared our data (without permission) with “4 trusted professionals”. We contacted The Dr Oz show and they wrote to us saying these individuals included Dr. Steven Woolf, Dr. Adrian Hernandez Diaz, Dr. Cornelius Dyke and Dr. William Li (names published at the request of the BMJ). We do not know if these “trusted professionals” were made aware of the fact this was confidential research information and should not be published without our permission.
All four of these “trusted professionals" proceeded to publish critiques of our research in the BMJ electronic letters section; and some of them actually included some of our research data in their response. At the time of our original submission of this response, (May 19), none of these “trusted professionals” declared their association with The Dr. Oz Show as a potential conflict of interest nor that they attained our data from The Dr. Oz Show. Furthermore, a few of their responses seem not intended for scientific debate or constructive suggestions for improvement. Dr. Hernandez has just (June 3) submitted a second critique, this time acknowledging his receipt of confidential data although does not declare from where he obtained this data. From these named “trusted professionals” 7 negative responses have been submitted. Ironically, all accuse us of inadequate transparency while none have declared their conflict of interest, and all have had contact with The Dr. Oz Show. It is not clear if any other letter writers have any affiliation with The Dr. Oz Show.
Nonetheless, in the spirit of collegial debate, we will address the key issues raised in the most recent rapid responses, some of whom have received our data from The Dr. Oz Show.
With respect to Dr. Woolf’s review, readers should bear in mind the aforementioned undisclosed interaction with the Dr. Oz Show. We note that he focuses solely on The Dr. Oz Show and never once mentions the other show we reviewed. Dr. Woolf’s review of our study is extensive, as he comments on anything potentially related no matter how peripheral, in his attempt to discredit the study.
His concerns do not relate to the findings of our study but rather focus on every aspect of our methods, mostly suggesting we did not provide enough information (even for results not presented in our study). If we had followed Dr. Woolf’s suggestions, our methods would have been even longer than his letter. Interestingly, the peer reviewers of our study had suggested we shorten and simplify the methods. Such is the reality of peer-review publication.
Dr. Woolf’s first fundamental argument is we are being unfair in assuming the “standards of evidence we use to critique physician guidance… are appropriate for a television show. ” Similarly Ayse Tezcan suggests the comparator standard should not be “evidence but what standard doctors do.” (PhD candidate Tezcan was not listed by the Dr. Oz Show as one of the “trusted professionals” with whom they shared data and therefore, to our knowledge, has no association with that show.) Dr. Woolf suggests our data contributes to “denigrating the reputation of a spokesman.” It is imperative for us to stress that in no way did we imply any value judgment to the findings we reported. We have also repeated this in previous responses. We have many times stated that sometimes guidelines and other recommendations have similar levels of evidence to what we found on these shows. We find it baffling that simply reporting on the evidence available to support these shows is considered denigrating. Other studies have compared clinical practice and guidelines to evidence (research) and we felt a similar comparison was the most reasonable. While some of our critics suggest the bar should be lowered for television, given the impressive viewership for these shows we strongly disagree.
Their second concern is that we should have compared these results to practicing clinicians. No one has so much as described the recommendations for medical shows before let alone completed an evidence review for a sample of the recommendations. This is an example of Dr Woolf’s faulty argument. Simply stating he feels we should have done more or done something differently does not invalidate nor change what we observed in our review.
Dr. Woolf found our sampling via “internet search” inadequate to determine which medical talk shows to review but failed to indicate the programs we missed.
Certainly his suggestion that we should have used qualitative tools, is one way to analyze this type of data. However, ours was a quantitative approach, another reasonable option. Our approach makes his suggestions regarding specific qualitative tools somewhat nonsensical.
Dr. Woolf’s comment that the data was difficult to obtain is interesting as he never approached us to request more data. He raises a number of captious issues: our title (which we addressed previously); the classification (which we acknowledged was tricky but did not in anyway affect the analysis of the evidence nor our assessment of the reporting of both benefits and harms); that we broke one recommendation into two (there were 160 recommendations, thus effecting 0.6% of our data); etc. Dr. Woolf hunts for faults while ignoring strengths like our standard data extraction techniques. He does not mention our dual independent review but implies our research assistants lacked skills to perform the task without ever having met them or assessing their skills. He states that David Sackett estimated only 10% of practice is evidence based. David Sackett and colleagues have indeed published on this topic, and found 82% of interventions offered to patients by a general medical team were evidence based.1 We would commend Dr. Woolf to a re-reading of Sackett’s paper as he would find that Sackett’s methods of evidence classification were not dissimilar to ours and not nearly as robust as he suggests they must be.
Dr. Sackner-Bernstein and Dr. Woolf were correct in stating we used unvalidated tools in our assessment. Given that this is the first evaluation of the evidence pertaining to televised medical shows, it is not surprising there were no validated tools to use. However, it is important to realize that even “validated” tools in evidence-based medicine, are not that reliable. In fact, there is poor agreement and inconsistency (often Kappa <0.5) in tools like GRADE and Cochrane Risk of Bias.2,3 We noted this in the original dual review of our study and attempted to improve this by using a committee of four experienced evidence-based reviewers and requiring agreement in three of four reviewers. We decided to use the term “believable” as we felt it would be more understandable to non-scientific readers than other categorizations that were never developed to assess medical recommendations on a TV show. Ideally a full systematic review for each of the evidence questions would have been ideal but the time commitment to complete 160 systematic reviews is simply prohibitive.
In the first Tezcan seems unclear on the objective of the study. The objective of our study, as with most research papers, is at the end of the introduction. Tezcan is concerned that our sampling was from a previous year of shows. A project of this magnitude cannot be accomplished and published in a few months. Clearly, only future research would be able to determine whether current shows are similar to past shows. With regard to what “no evidence” meant: it meant there was no evidence. That means research has not been published on that specific clinical question. However, it does not mean there is conflicting evidence. When there was conflicting evidence pertaining to a particular recommendation we deferred to the highest level of evidence we could find. If conflict occurred at a similar level of evidence, we gave the benefit of the doubt to the shows and concluded the evidence supported the recommendation. In the second posting 10 days later, Tezcan states she has now watched more episodes of The Dr. Oz Show than our team did (40 episodes). She then suggests another idea for a study of these shows, which we encourage her to pursue.
We have been exceptionally clear about the limits of our study including the challenges of researching this previously unexamined area, particularly the non-specific and subjective nature of the recommendations. Complaints, such as from Dr. Woolf, about how we generated different questions compared to how he would, simply reinforces our findings regarding the subjective nature of the recommendations. It should be noted we also used other information from the show to try to more accurately define the questions and represent the context of each recommendation more fairly. Regardless, the exercise by our critics demonstrates why this research should be replicated from the start. Simply jumping in at the end does not provide Dr. Oz’s “trusted professionals” with adequate context of these recommendations to develop adequate questions, nor a complete understanding of the process.
We would now like to address the request for the public release of our data. In addition to the normal peer review process that occurred before the study was published, we have offered full disclosure of our data to the BMJ to verify any aspect of our research.
We have concerns about full public release of our data. First, a number of the requests for release of data have not been related to scientific purposes. For example, we have had requests to use our data for online games, website development and to further particular agendas. This, as was pointed out in our first rapid response, was neither the purpose nor would it meaningfully add to the scientific discussion around our paper.
Second, we believe people with a stance for or against the shows may seek to repeat our specific evaluation with the result of verifying their predisposition bias. Indeed, the most caustic responses have come as a result of our data being shared (without our permission) by The Dr. Oz show, without any declaration of potential competing interests. We have strong concerns that any further analysis of our data will likely be biased to assert a specific agenda.
Third, these data were very time consuming to collect and process. We are still considering further evaluations and possible future publications. Complete data release could potentially negate this.
Fourth, some of the criticisms about our study arise from our methodology and possible subjectivity in our interpretations. We have been clear about our methodology and its limitations and have encouraged others to repeat the research in its entirety including: independent dual review of episodes, identifying recommendations, delineating stronger recommendations, elucidating the details around the recommendations, generating clinical questions from the recommendations, searching evidence for the recommendations, and finally assigning the final results of the evidence assessment. We would encourage authors such as Dr. Woolf to develop a framework for this type of study. Replication from the start is the most reliable way to verify and improve upon research. We would be more than happy to assist other truly independent academic researchers to replicate our study, which could include a consideration for sharing data with independent scientists who present a research and analysis plan.
Finally, everyone should be reminded that these shows are public and the data is readily available for people willing to watch and systematically record the data as we did. To assist in the evaluation of these shows, we would encourage all televised medical shows to publish transcripts and make them readily available.
As one last note, we hope others can soon move on from this debate. Most of the critics of our paper have raised similar issues and we have addressed them. We recognize that some people will still be dissatisfied but further debate is unlikely to change that. Furthermore, several of the critics have communicated with at least one of the shows and failed to disclose this, leaving us uncertain if the motivation for correspondence stems from interest in the research or enthusiasm to defend the shows. These debates are time consuming and have little further scientific merit. We hope those keen on future letters will direct that effort to repeating and improving on our study.
Interestingly, some of the key findings of our study, such as the relative lack of information on the magnitude of the benefit, potential harms, and costs provided with the recommendations, have been completely ignored by our critics. We hope future health recommendations in all media avenues will consistently include the:
• specific details of the proposed benefit
• magnitude of the proposed benefit
• potential harms and cost(s) of the intervention
• evidence supporting the recommendation, and
• disclosure of potential conflicts of interest.
Only then will the viewers/readers (our patients) be able to decide whether they should consider adopting televised medical shows recommendations.
References
1. Ellis J, Mulligan I, Rowe J, Sackett DL. Inpatient general medicine is evidence based. A-Team, Nuffield Department of Clinical Medicine. Lancet. 1995;346(8972):407-10.
2. Hartling L, Fernandes RM, Seida J, Vandermeer B, Dryden DM. From the trenches: a cross-sectional study applying the GRADE tool in systematic reviews of healthcare interventions. PLoS One. 2012;7(4):e34697.
3. Hartling L, Hamm MP, Milne A, Vandermeer B, Santaguida PL, Ansari M, et al. Testing the risk of bias tool showed low reliability between individual reviewers and across consensus assessments of reviewer pairs. J Clin Epidemiol. 2013;66(9):973-81.
Competing interests: No competing interests
I received several comments from colleagues about my rapid review published on May 2nd 2015. I would like to clarify some of the points I raised about the Korownyk et al. BMJ 2014;349:g7346 paper. I must mention I have extensive experience in searching for evidence and conducting systematic reviews in clinical medicine.
My main concerns on this paper are related to the lack of proper documentation that had to be available for reviewers and readers, and methodological deficiencies of the study. The two concerns harm the conclusions of the study.
Several evaluated recommendations were vague and the authors did not say how exactly they extracted the recommendations from the shows. Sharing the transcripts around such recommendations is essential to helping the reader to judge whether a given extracted recommendation was the right one for a show comment.
Some of the recommendations given in the two TV shows were backed by scientific organizations or invited experts only; this situation had to be disclosed in the shows as this type of evidence is not the best available. The strength of recommendations is based not only on efficacy or effectiveness evidence from studies, but also on costs and patient values and preferences (1). I think strong recommendations should provide a formal GRADE assessment, for example by using Summary of Findings tables (2); weak recommendations should mention that evidence does not exist or is poor. Also, some recommendations were removed by authors from the original list and some others were retained; the process on how to decide it is not well explained and is highly subjective.
I had confidential access to the authors’ spreadsheet. Developing searchable questions from vague recommendations made in the context of a talk show is a difficult process with a high risk of bias. Authors of the study did not explain in detail how they did this entire process, and reviewers/readers therefore cannot judge its appropriateness. The supplement file did not provide enough information. Some of the searchable questions, for instance, were focused on the first part of the recommendation, so an incomplete or inappropriate question was developed for such recommendation. In particular 22 questions from the total of 80 examined may have been erroneous, and 7 seemed very difficult to elaborate because of their complexity.
From the 44 recommendations that authors mentioned did not have evidence (n=41) or evidence was weak (n=3), I agreed with 42 of them. However, even though the authors report a 15% incidence of recommendations that are “contradicted” by literature, I found a 4% incidence so there does not appear to be a pattern of harmful recommendations. For my conclusion, I ran non-systematic searches for every recommendation in pubmed, scopus and google scholar in limited time; in line with my comments above, recommendations coming from scientific organizations were not considered good evidence.
Other identified methodological problems of the study include the authors’ modification of the number of reviewers of evidence for recommendations from two to four in the middle of the study period; again the process is not well detailed. Some terms were vaguely defined such as believability and consistency of the evidence; according with the description I cannot understand what authors exactly mean.
The lack of transparency of authors on details of their study and the methodological deficiencies are the main problems of the Korownyk et al. paper. A full disclosure of study details would definitely improve this paper, and judging medical shows without placing the study deficiencies in context is risky at the least.
Finally, although I understand the authors' desire to strictly define high quality information as not including government and medical society recommendations, I am disappointed in how the data resulting from this approach has been interpreted by the authors and subsequent press. This was a very difficult study to conduct and abstracting testable hypotheses offers risks, so the author’s conclusions should have been more modest. For example, if a TV show includes government or medical society data (like advising young men to examine their testicles for common cancers), does it make TV health shows so unreliable that viewers should be dismissive or skeptical? This accusation does not follow the self imposed methodological limitations of this study.
References:
1. Andrews JC et al. GRADE guidelines: 15. Going from evidence to recommendation-determinants of a recommendation’s direction and strength. J Clin Epidemiol 2013;66:726-35.
2. Guyatt GH et al. GRADE guidelines: 12. Preparing summary of findings tables-binary outcomes. J Clin Epidemiol 2013;66:158-72.
Adrian V. Hernandez, MD, PhD
Research Professor, School of Medicine, Universidad Peruana de Ciencias Aplicadas, Lima, Peru.
Adjunct Faculty, Health Outcomes and Clinical Epidemiology, Dept. of Quantitative Health Sciences, Cleveland Clinic, OH, USA.
adrianhernandezdiaz@gmail.com
Competing interests: No competing interests
I have read with interest the article authored by C. Korownyk, et al, and appreciate their attempt to take on this controversial topic. I am Boarded in Diagnostic Radiology, Nuclear Medicine as well as Integrative Medicine and, am familiar with rigorous scientific inquiry as well as the challenges of publishing in the Alternative Medicine realm. I, therefore, have questions regarding to the data used for this investigation.
Without clear understanding of the exact topics and the specific questions posed to investigate each topic, it is impossible to determine whether the conclusions drawn are appropriate. More accurate interpretation of this article would by facilitated by the inclusion of a table listing the searched topics as well as the exact search terms used. Was this information available to the reviewers of this article, as I cannot comprehend how a decision to publish could be made without access to the raw data.
Competing interests: No competing interests
Knowing all the design flaws of this study, hearing more and more of this article being provided as evidence to discredit the physicians in these shows makes me wince.
BMJ's role in disseminating biases to overcome some TV shows' (emphasis on 'shows') perceived lack of EBM recommendations does not correlate with BMJ’s respected journalistic reputation. Is this the idea: to prevent a bigger harm, we are justified to do little harm? With that logic, these shows may be providing a lot of good advice in return for little harm. Then, how about considering this: these shows may be providing a reasonable and healthier advice to people who would, otherwise, have no access to any health advice or be following true quack advice from completely unqualified individuals or maybe quietly practicing board-certified physicians. An intriguing study would have been to find out how much misinformation these shows dispel, give healthy life-style recommendations and perhaps direct people to seek proper advice. (I finally watched some episodes of the Dr. Oz show; in fact, more episodes than these authors did.) Additionally, an analysis of context-based recommendations comparing these shows to the traditional medical practice model could be more enlightening for the public than this currently published paper.
Then, how could a prestigious journal like BMJ let this kind of poor quality study be published in such short time and allow the public/health care providers to be misled while so many well-qualified papers get rejected (disclaimer: I have never submitted a paper to BMJ, so I am not a disgruntled scientist here)? Perhaps the publicity and impact score are more important for the journal than scientific rigor. I realize that there has been a growing concern among some physicians and scientists about the possible dangers to the public of the products and views expressed in these shows. These concerns might be well-placed; however, that does not give anyone to tout mere opinion as a study - a letter of concern could have done the job. The personal biases should not get in the way of doing and setting a good example for good science. We do not compromise good science practices to expose charlatans; that’s the tabloids’ job.
BMJ has a responsibility to respond to the concerns of and excellent points made by the commentators here. Hopefully, they are not determining the reliability and validity of a study by the popularity vote.
Thank you in advance.
Yours truly,
Ayse Tezcan
Competing interests: No competing interests
Following is Part 3 of my critique of Korownyk et al:
Examination and search of the evidence: The authors provide few details on their search methods. They say little about precisely how they converted statements made on air into “searchable questions,” saying only “we developed a search question.” There are grounds to question the validity of their choices. For example, the supplement mentions that a statement—“you want your CRP numbers to be less than 1.0 mg/dL”—implies that “you have to change your CRP,” arguing that this justified a literature search for evidence of benefit from lowering CRP levels. It could be argued, however, that the statement is actually a description of the normal range of CRP, and the normal range derives not from intervention studies but from observational data on population risks associated with values outside the normal distribution (e.g., the 95% confidence interval). Normal distribution data are the basis for the Recommended Daily Allowances (RDA) for nutrients on every food package in the United States, along with the normal ranges for electrolytes, blood counts, urine sediment, and any other reference range on laboratory reports; on all of these, “you want your numbers to be less than” the elevated values associated with increased risk. The searchable question for such statements ought to be “is there evidence that values above the threshold are associated with significantly increased risk of disease?”
I have obtained the author’s spreadsheet, which raises further doubts about the formation of the search questions and an apparent tendency to create straw men by setting more narrow boundaries for the search than the reported statement covered. Consider the first 10 statements in their list of 80 recommendations, in which I found 5 instances in which the question was narrowly reframed:
• Reported broadcast statement: “Give yourself permission to have your [emotional] pain, not to be afraid of it.” The research question should be a variant of “Does acceptance of emotional pain improve wellbeing?” The authors constructed a narrower question: “Does allowing yourself to grieve prevent physical pain?”
• Reported broadcast statement: “Start your day green: wheat grass, vegetables etc...” The research question should address the spectrum of outcomes associated with fresh produce intake. The authors framed a narrower question: “Does eating green food oxygenate your body/blood?”
• Reported broadcast statement: “Look for the sneaky words on the menu. Want to avoid - cheesy, creamy, buttery. But look out for Homemade, legendary, luscious, mama's.” The research question should be whether foods with these labels contain higher levels of unhealthy nutrients. The authors’ reframed the question as if the statement alluded to weight gain: “Does food in restaurants with the word homemade, legendary, luscious, or mama's cause weight gain?”
• The broadcast statement reportedly outlined six symptoms of toxocariasis (abdominal pain, headache, coughing bouts, SOB, wheezing, some chest pain). The authors narrowed the question: “Is abdominal pain a symptom of toxocariasis infection?”
• Reported broadcast statement: “Don't start with laxatives, they destroy you inside, hair, teeth, family and home.” The authors narrowed the question to: “Does chronic laxative use negatively impact the GI tract?”
The authors looked only at bibliographic sources to find corroborating evidence but seem to have ignored evidence-based practice guidelines and evidence-based policies issued by government panels and agencies. A U.S. physician who advises women to get a mammogram—as does Dr. Oz and thousands of other physicians—is often basing this advice on the guidance of panels like the U.S. Preventive Services Task Force or the American Cancer Society. Turning to PubMed for the answer buries the physician in a literature beset with controversy and debate about the dozens of RCTs and observational studies of mammography. By ignoring this larger literature, the authors seem to chide a physician for reflecting the recommendations of panels that have studied this evidence and issued national guidelines rather than limiting their advice to what they can find in individual papers.
Even for the latter, the authors are vague in defining what they mean by “believable” evidence, other than some generalities: “the quality, quantity, and type of evidence available.” Without supplying metrics for defining what constitutes good quality, adequate quantity, etc., one has to assume the judgment was highly subjective and idiosyncratic. A voting process is mentioned, but the details of that voting process were not provided.
The supplement says decisions were made about the “ideal study design” for each question but offers no details on how that choice was made (or which study design was chosen for each of the 80 statements). The supplement alludes parenthetically to “cohort or RCT,” but more classes of evidence are pertinent to the breadth of remarks uttered in a broadcast. For example, neither a cohort study nor a RCT is the ideal standard for test performance characteristics (e.g., sensitivity or specificity). The same is true for the nutrient content of foods, a topic covered, for example, in a broadcast statement about the glucose content of ripe bananas, information intended for patients with diabetes. Population studies are an inappropriate source on this question, yet the authors sought prospective studies of “whether ripe bananas affect blood glucose or insulin levels.”
In conclusion, the weaknesses in this study raise serious questions about the validity of the percentages that drew attention in the abstract and that went viral on social media. It would do well for the BMJ do be transparent about the questions surrounding the accuracy of these results and to move the direction of future research on this topic away from ad homonym frameworks. The journal should encourage a more systematic examination of this medium by investigators who understand the topic and are willing to hold their own research to a modicum of the evidence standards they seem willing to apply to the work of others.
Sincerely,
Steven H. Woolf, MD, MPH
Professor, Department of Family Medicine and Population Health
Virginia Commonwealth University
Competing interests: No competing interests
Following is Part 2 of my critique of the Korownyk et al. paper:
The authors describe themselves as “experienced evidence reviewers” but their writing suggests otherwise: they did not use current standards or terminology expected of the critical appraisal of evidence. For example, they say they searched for “believable” evidence, a term without citation that they appear to have devised themselves rather than using validated instruments such as the GRADE criteria or their equivalents (see http://handbook.cochrane.org/chapter_12/12_2_1_the_grade_approach.htm). Leaders in critical appraisal differentiate between the strength of evidence and the magnitude of effect, the latter including nuanced assessments that recognize that the tradeoff between benefits and harms is a continuous variable, not dichotomous (see methodology of the U.S. Preventive Services Task Force). The authors applied a blunt, binary approach that experts no longer recognize: the online supplement that accompanied the paper says “the evidence was reported as either ‘evidence of effect,” ‘evidence of no effect,’ ‘evidence of harm,’ and ‘no evidence found.’” The supplement adds, “the group determined whether the recommendation was supported by evidence with ‘yes’ or ‘no.’” This dichotomous approach to judging evidence ignores decades of work in evidence-based medicine, much of it published in the BMJ, to grade recommendations (e.g., A, B, C, etc.) to capture the relative strength of the evidence.
Other characteristics of this study lack the rigor one expects of a BMJ study, such as transparency of methods. The poorly worded supplement provides some additional information but lacks full details. It mentions a spreadsheet that lists each of the 80 recommendations and how they were evaluated, but the spreadsheet is difficult to obtain (and I suspect was unavailable to the BMJ reviewers, who might have otherwise discovered flaws in the authors’ coding scheme). Other features of the paper reflect inexperience with the science of critical appraisal. The very title of the paper belies a misunderstanding of the definition of a prospective observational study (there is nothing prospective about reviewing previously recorded broadcasts). The authors used odd terms, outside the mainstream of clinical epidemiology, such as “consistency” for what is generally known as internal and external reliability. Their power calculation omits α or β assumptions and they claim having “precision of +10%” without specifying a unit of analysis. More specific problems follow.
Design flaws in the study relate to the sampling of recommendations, the auditing technique for content analysis, the subjective definition of outcome measures, and the standards for examining the evidence.
Sampling: The paper’s introduction states the goal to sample the “most popular medical talk shows on television” but only an “internet search” is cited as the basis for selecting two talk shows. Although the authors make much of their use of randomization to select the 80 statements, the supplement omits to selective addition and removal of recommendations post-randomization based on criteria that seem arbitrary and subjective. For example, the supplement admits to having removed a recommendation they considered vague and without a clear outcome (“Don’t give up the foods you love, just prepare them in the healthiest way possible”). However, they retained other forms of advice that was equally generic (e.g., “Give yourself permission to have your [emotional] pain, not to be afraid of it”; “To get healthy with spouse: rule #3: Get advice - need to seek help before it happens - they see a counselor”).
Auditing technique: The authors gave no details on their auditing technique and seem to be inexperienced with the methods of conversational analysis. They admit to being unfamiliar with broadcast health information, seemingly making them poor choices to conduct this analysis. The supplement mentions that the authors could not obtain transcripts or DVDs from the shows’ producers but provides no documentation of what steps, if any, were taken to transcribe the 80 programs or to use qualitative tools (e.g., ATLAS.ti software) to code the content. Rather than a proper content analysis, they relied on family physicians to review the prior year’s episodes to “design a spreadsheet.” Without further details, it is difficult to know whether the broadcast statements entered into that spreadsheet were verbatim comments uttered on air or the authors’ paraphrasing. The supplement lists variables that were abstracted from the broadcasts (e.g., “key points made on the show,” “specific recommendations,” “who the recommendations were for,” discussion of “provocative topics,” and use of “medical jargon), but precisely how they were classified is not explained. Judgments in the second round about how the shows described benefits, harms, etc. were also not disclosed.
Outcome measures: The study hinges on which statements the authors labeled as “recommendations,” but the methods of this highly subjective task are poorly documented. They did not share the excerpts or conversational context from which the statements were drawn to know whether the chosen statements were taken out of context. Physician advice—at the bedside or on television—does not abide by a binary framework (recommendation/no recommendation) but spans multiple domains and varies in explicitness. The authors claim to have limited their focus to “stronger” recommendations, but no definition for “stronger” is provided. They said that they “attempted to delineate the more definitive recommendations” but say nothing about how they did this.
This lack of transparency is problematic given the inevitable imprecision that is inherent in these judgments, which (according to the paper) were left in the hands of research assistants (VL, KO), whose clinical experience for making these judgments is unknown. The research assistants were expected to “document the topics discussed and the details of recommendations” and the family physicians became involved only to adjudicate discrepancies. It is apparent from the supplement that the authors obviously struggled clumsily with this complexity, finding it necessary to retry classification strategies and modify criteria midstream—ultimately leaving behind many clues that their classification scheme was both subjective and confusing. One recommendation, according to the supplement, was “felt to be two recommendations.” Another statement (the need to get help for binge eating) was removed because it was “felt to be more of a definition of a psychiatric disorder.” Other statements that described psychiatric disorders (e.g., laxative abuse) were retained.
Imprecision in other outcomes also suggests inexperience with the topic. For example, the authors measured the frequency of disclosures of conflicts of interest, apparently unaware of the contextual influence of U.S. laws governing the subject. The U.S. Federal Trade Commission regulates "unfair, deceptive, or misleading" endorsements and makes it illegal for a broadcast employee to not disclose a benefit. Most hosts therefore do not bother mentioning the subject unless there is a disclosure to share, giving little meaning to the < 1% figure reported in the paper.
Further concerns follow in Part 3.
Sincerely,
Steven H. Woolf, MD, MPH
Professor, Department of Family Medicine and Population Health
Virginia Commonwealth University
Competing interests: No competing interests
No British shows in Trinidad and Tobago? Is there a ban?
In Trinidad and Tobago the predominance of American Television has blinded the people to the existence of the British Model and British way of doing things which is predominant in politics ,law health education and sports .There is no legislation banning British Television but American Television Dramas are used to blind citizens that the islands still is a British Model of civilisation.Citizens falsely believe we have moved on because they don't know what is the system in Britain to recognize it locally
Competing interests: No competing interests