The Quality of Evidence of and Engagement With Video Medical Claims, 2025, Kang et al

rvallee

Senior Member (Voting Rights)
The Quality of Evidence of and Engagement With Video Medical Claims
Journal of the American Medical Association
EunKyo Kang, MD1,2,3; HyeWon Lee, RN-BSN1; Juyoung Choi, RN-BSN1, HyoRim Ju, MD4,5

Key Points

Question

Are medical claims in health care professional–created online videos supported by strong scientific evidence, and are evidence levels associated with engagement metrics or traditional video quality assessment tools?

Findings
In this quality improvement study of 309 video claims, 62% of claims relied on very little or no evidence, while less than 20% were supported by high-quality evidence. Grade D videos (ie, those with very low or no certainty of evidence) were associated with a 35% higher view count than grade A videos (ie, those with very high certainty of evidence), and traditional quality tools showed weak correlations with evidence levels, thus failing to detect important qualitative differences.

Meaning
These findings suggest that to maintain scientific integrity in digital health communication, evidence-based content standards and improved science communication training for health care professionals are needed.


Abstract

Importance

The unexplored quality of evidence supporting online video claims by medical professionals creates a credibility-evidence gap that threatens the principles of evidence-based medicine.

Objective
To systematically evaluate the evidence hierarchy supporting medical claims in health care professional–created online videos using a novel evidence classification framework.

Design, Setting, and Participants
In this quality improvement study using a cross-sectional analysis, YouTube was searched using cancer- and diabetes-related terms. A total of 309 videos met the inclusion criteria. The video search, data extraction, and archiving were conducted between June 20 and 21, 2025, to create a static dataset. Videos were assessed using the newly developed Evidence-GRADE (E-GRADE [Grading of Recommendations Assessment, Development and Evaluation]) framework, categorizing evidence into 4 levels: grade A (high certainty from systematic reviews and/or guidelines), grade B (moderate certainty from randomized clinical trials, cohort studies, and high-quality observational studies with clear citations), grade C (low certainty from limited observational studies, physiological mechanisms, or case series without critical appraisal), and grade D (very low or no certainty from anecdotal evidence).

Exposure
Videos that had a minimum of 10 000 views, were created by health care professionals, had a minimum duration of 1 minute, and contained specific health claims.

Main Outcomes and Measures
Primary outcomes included the distribution of evidence grades (A-D) supporting medical claims. Secondary outcomes included correlations between evidence quality and engagement metrics (views and likes) and traditional quality scores (DISCERN, JAMA benchmark criteria, and Global Quality Scale).

Results
Among the 309 videos included, which had a median of 164 454 (IQR, 58 909-477 075) views, most medical claims (193 [62.5%]) were supported by very low or no evidence (grade D), while only 61 claims (19.7%) were supported by high-quality evidence (grade A). Moderate (grade B) and low (grade C) evidence levels were found in 45 (14.6%) and 10 (3.2%) videos, respectively. The correlation with view counts was statistically significant for grade D videos, which were associated with a 34.6% higher view count (incidence rate ratio, 1.35; 95% CI, 1.00-1.81; P = .047) than grade A videos. Traditional quality tools showed only weak correlations (range of coefficients, 0.11-0.23) with evidence levels, thus failing to detect important qualitative differences.

Conclusions and Relevance
In this quality improvement study, a substantial credibility-evidence gap was found in physician-generated video-sharing content, where medical authorities often legitimized claims lacking robust empirical support. These findings emphasize the need for evidence-based content guidelines and enhanced science communication training for health care professionals to maintain scientific integrity in digital health information.
 
They only looked at videos for cancer and diabetes, where the clinical and scientific evidence are very solid, and still most were of low quality.

This is the grading they used:
  • Grade A (high certainty): Claims are explicitly supported by systematic reviews, meta-analyses, or established clinical practice guidelines from major health organizations (defined as national or international, government-recognized, or large nonprofit professional bodies, eg, the World Health Organization, the American Dental Association).
  • Grade B (moderate certainty): Claims are backed by specific randomized clinical trials or high-quality observational studies with clear citations (operationalized as large-scale prospective cohorts or well-designed case-control studies).
  • Grade C (low certainty): Claims are supported by limited observational studies (eg, smaller retrospective studies, case series), physiological mechanisms, or case series without critical appraisal.
  • Grade D (very low or no certainty): Claims are based solely on anecdotal evidence, personal experience, or unsupported assertions.
Even the evidence used in textbooks, clinics and other official sources on chronic illnesses like ME/CFS and Long Covid mostly have a D grade and the rest barely have a C grade, so things are far worse than this because both diabetes and cancer pretty much have top tier respect from both the medical profession and the public, so it's about as good as it gets.

Except because of how screwed up things are, most evaluations of junk videos promoting psychosomatic models would likely be considered of moderate-to-high quality while accurate content from the patient community or organizations would be considered of lower grade. Because fundamentally this is a political process, change the people and you change the evaluation. The quality of the evidence doesn't change, but its evaluation does.
 
Back
Top Bottom