Assessment of the Quality and Reliability of YouTube Videos on Fuch’s Endothelial Corneal Dystrophy

DOI: https://doi.org/10.21203/rs.3.rs-943108/v1

Abstract

Purpose: The aim of the study was to assess the quality and reliability of videos on Fuch’s endothelial corneal dystrophy (FECD) on YouTube.

Methods: A search of YouTube was performed for the term “Fuch’s endothelial corneal dystrophy” without any changes to the website’s standard search preferences. The first 100 videos were recorded. A total of 71 videos were evaluated regarding the DISCERN score (min–max: 16–75), Journal of the American Medical Association score (min–max: 0–4), and Global Quality Scale score (min–max: 0–5).

Results: The mean DISCERN score was 40.1±15.6 (moderate), the mean JAMA score was 2.01±0.7 (poor), the mean Global Quality Scale score was 2.5±1.3 (moderate), and the mean video power index score was 106.8±135.7. Twenty-three of the videos (32.4%) were uploaded by physicians, 25 (35.2%) by universities/private hospitals, 21 (29.6%) by health channels. Thirty-six videos (50.7%) contained information about the disease, 24 (33.8%) discussed surgical techniques, and 11 (15.5%) were about patient experience.

Conclusion: YouTube provides only moderate-quality health information on FECD. Physicians and professional organizations should be aware of and embrace this evolving technology to raise awareness about FECD.

Introduction

The internet has evolved into a common information source, and 80% of users consult web sources for health information [1]. YouTube (www.youtube.com), which provides video sharing over the internet free of charge, is leading the way in information sharing [2]. Videos allow for the demonstration of difficult concepts using simulation, diagrams, dynamic illustrations, and real patients. Although anyone can access information via YouTube, it may not be possible for everyone to judge the quality, reliability and accuracy of that information. Biased or inaccurate information can have a negative effect on communication and the trust bond between a physician and patient, especially when it comes to discussing treatment options [3].

Fuchs endothelial corneal dystrophy (FECD) is a bilateral disease of the corneal endothelium characterized by the accelerated loss of corneal endothelial cells with changes in the Descemet membrane (DM), including the accumulation of the extracellular matrix and formation of the posterior focal excrescences, called guttae. Loss of vision from FECD can result from these DM changes as well as later-stage disruption of corneal endothelial pump-leak function causing corneal edema, bullae formation, and subepithelial fibrosis [4]. Medical therapy for FECD-related vision loss is limited to decreasing corneal edema with topical hyperosmotic drops and ointment. The management of FECD has dramatically evolved over the past 20 years. Currently, corneal transplantation represents the only definite treatment option, and FECD is the most common cause of corneal transplantation worldwide [5, 6]. The current preferred surgical management of FECD is endothelial keratoplasty (EK), which has shown excellent rates of corneal clearance. With EK, the risk of surgical intervention including graft rejection, prolonged visual rehabilitation, infection, and postoperative astigmatism has decreased significantly. Patients can now undergo transplantation earlier in the disease course, guided by visual symptoms and clinical findings [7].

In the literature, the quality and reliability of online videos have been investigated for different healthcare disciplines. However, to the best of our knowledge, no previous study has evaluated YouTube videos on FECD. Thus, the aim of our study was to assess the quality, reliability and popularity of YouTube videos on FECD.

Materials And Methods

This study did not require institutional review board approval. A search was performed on YouTube® (http://www. youtube.com) on July 16, 2021 using the keyword “Fuch’s endothelial corneal dystrophy”. The top 100 videos were selected for the review, as determined by ‘relevance’ according to YouTube's algorithm. Video searches were performed without any user login after clearing the entire search history of the browser. Videos presented in non-English-language, those without sound, those shorter than 60 seconds, those that were not related to FECD, and duplicate videos were excluded.

For each video included in the study, the title, number of views, video length, time since upload (age), number of likes, number of dislikes, and number of comments were recorded, and the like ratio, view ratio and video power index (VPI) were calculated. To assess the popularity of the videos, VPI and the view ratio were used. The view ratio was calculated by dividing the number of views by the time since upload. The like ratio was calculated as follows: number of likes × 100 / (number of likes + number of dislikes). To calculate VPI, the method described by Erdem et al. was used: like ratio × view ratio / 100.

The videos were also categorized into four groups according to their upload source: physicians, universities/private hospitals, health channels, and independent users. Categories classified according to content were 1) information about disease, 2) surgical technique, and 3) patient experience.

The quality of information presented in each video was assessed using the DISCERN score and the Journal of the American Medical Association (JAMA) benchmark. The DISCERN questionnaire consists of three sections including 16 questions, and a higher score indicates better quality. The first eight questions are related to reliability, and the next seven questions evaluate the specific details of treatments received. The last question addresses the overall quality of a publication. In the current study, according to the DISCERN scoring system, the videos were grouped into excellent quality (63–75 points), good quality (51–62 points), fair quality (39–50 points), poor quality (27–38 points), and very poor quality (16–26 points).

The JAMA benchmark evaluates the quality of online information based on four criteria: authorship, attribution, disclosure, and currency. One point is given for each criterion, and the highest quality is indicated by four points.

The Global Quality Scale (GQS) was used to assess each video in terms of its instructive aspects for patients. The GQS system allows users to evaluate the overall quality of a video’s content on a five-point Likert scale. While a score of one point indicates poorest quality, a score of five points indicates excellent quality.

Each video was independently analyzed by two experienced surgeons, and the obtained data were recorded. The reproducibility of the DISCERN, JAMA and GQS scores were tested before the primary analysis, and the inter-observer intra-class correlation coefficients revealed a Spearman correlation coefficient of greater than 0.90. The differences in the mean values of the scores between the observers were statistically analyzed.

Statistical analysis

The data obtained were analyzed using SPSS v. 16.0 (SPSS Inc., Chicago, IL, USA). Data distribution was tested using the Shapiro–Wilk test. The Kruskal-Wallis test was used to compare the parameters between the groups. The Tamhane test was used as a post-hoc method. The correlation statistics were obtained using the Spearman test. P values of less than 0.05 were considered statistically significant.

Results

A total of 100 videos were initially selected. After excluding videos that were not related to FECD (n = 6), duplicate videos (n = 6), videos shorter than 60 seconds (n = 7), those without sound (n = 7), those presented in any language other than English (n = 1), and those that could not be opened (n = 2), 71 videos were included in the sample. Table 1 summarizes the descriptive statistics of the videos.

Table 1

Descriptive Statistics of the Evaluated Videos

Descriptive statistics

Mean ± SD

Range

View count (n)

3242.54 ± 4845.22

15-34084

Length (min)

13.7 ± 18.9

0.5–65

Age (d)

44.1 ± 37.7

1-139

Likes (n)

33.1 ± 38.8

0-196

Dislikes (n)

1 ± 1.3

0–5

Comments (n)

2.61 ± 6.1

0–47

DISCERN score

40.1 ± 15.6

16–80

JAMA score

2.01 ± 0.7

1–4

GQS score

2.5 ± 1.3

1–5

Like ratio

95.3 ± 9.6

50–100

View ratio

105.8 ± 138.2

0.2–603

VPI

106.8 ± 135.7

0.2–597

SD, standard deviation; VPI, video power index; JAMA, Journal of the American Medical Association; GQS, Global Quality Scale;

Twenty-three of the videos (32.4%) were uploaded by physicians, 25 (35.2%) by universities/private hospitals, 21 (29.6%) by health channels, and two by independent users (2.8%). Thirty-six videos (50.7%) contained information about the disease, 24 (33.8%) discussed surgical techniques, and 11 (15.5%) were about patient experience.

The mean DISCERN score was 40.1 ± 15.6, the mean JAMA score was 2.01 ± 0.7, the mean GQS score was 2.5 ± 1.3, and the mean VPI was 106.8 ± 135.7. When we evaluated the videos according to the upload source, there was no significant difference in the DISCERN, JAMA, GQS and VPI scores (p = 0.18, p = 0.52, p = 0.68, and p = 0.76, respectively) (Table 2).

Table 2

Comparison of the DISCERN score, JAMA score, GQS score and VPI according to the video upload source

 

Physicians

Universities/private hospitals

Health channels

Independent users

p

DISCERN score

40.3 ± 14.7

37 ± 14.3

44.8 ± 17.1

37 ± 29.6

0.18

JAMA score

2.1 ± 0.46

1.91 ± 0.9

2.04 ± 0.8

 

0.52

GQS score

2.7 ± 1.2

2.3 ± 1.1

2.7 ± 1.5

2 ± 1.4

0.68

VPI

126 ± 141

84.1 ± 124.7

103.3 ± 136.4

192.9 ± 257.6

0.76

VPI, video power index; JAMA, Journal of the American Medical Association; GQS, Global Quality Scale

However, the DISCERN, JAMA and GQS scores statistically significantly differed according to video content (p = 0.032, p = 0.000, and p = 0.002, respectively) (Table 3). There was a significant difference between the videos on patient experience and those on surgical techniques in relation to the DISCERN score (p = 0.014). While there was no significant difference in the JAMA scores of the videos providing information about the disease and those discussing surgical techniques, the videos about patient experience had a significantly different JAMA score compared to the other two video content groups (p = 0.00 for both). A significant difference was also observed between the videos on patient experience and those on surgical techniques in terms of the GQS scores (p = 0.002). Lastly, VPI statistically significantly differed between the videos on patient experience and those providing information about the disease (p = 0.017).

Table 3

Comparison of the DISCERN score, JAMA score, GQS score and VPI according to video content

 

Information about Disease

Surgical technique

Patient experience

p*

DISCERN score

40.2 ± 15.1

45 ± 16.3

30.6 ± 15.2

0.032*

JAMA score

2.17 ± 0.67

2.16 ± 0.7

1.2 ± 0.4

0.000*

GQS score

2.38 ± 1.27

3.16 ± 1.2

1.7 ± 1.05

0.002*

VPI

138.9 ± 173.4

88.3 ± 77.05

42.1 ± 45.8

0.25

*p < 0.05; VPI, video power index; JAMA, Journal of the American Medical Association; GQS, Global Quality Scale

The DISCERN, JAMA and GQS scores showed significant correlations with each other. These scores were also correlated with video length, but they were not correlated with age, number of likes, number of dislikes, or number of comments (Table 4).

Table 4

Correlation Between the DISCERN Score, JAMA Score, GQS Score, VPI, and Video Parameters

 

DISCERN Score

JAMA Score

GQS Score

VPI Score

View Ratio

Age

Likes

Dislikes

Comments

Length

DISCERN Score

 

r = 0.357*

p = 0.002*

r = 0.803*

p = 0.000*

r = 0.114

p = 0.355

r = 0.119

p = 0.323

r = -0.1

p = 0.408

r = 0.056

p = 0.645

r = 0.002

p = 0.984

r = -0.019

p = 0.872

r = 0.346*

p = 0.03*

JAMA Score

r = 0.357*

p = 0.002*

 

r = 0.444*

p = 0.000*

r = 0.017

p = 0.888

r = 0.001

p = 0.992

r = -0.145

p = 0.228

r = 0.067

p = 0.579

r = -0.070

p = 0.561

r = 0.003

p = 0.977

r = 0.551 *

p = 0.000*

GQS Score

r = 0.803*

p = 0.000*

r = 0.444*

p = 0.000*

 

r = 0.173

p = 0.159

r = 0.212

p = 0.075

r = -0.018

p = 0.882

r = 0.154

p = 0.201

r = 0.081

p = 0.500

r = 0.137

p = 0.254

r = 0.399*

p = 0.001*

VPI

r = 0.114

p = 0.355

r = -0.017

p = 0.888

r = 0.173

p = 0.159

 

r = 0.995*

p = 0.000*

r = -0.196 p = 0.109

r = 0.748*

p = 0.000*

r = 0.418*

p = 0.000*

r = 0.430*

p = 0.000*

r = 0.195

p = 0.111

View Ratio

r = 0.119

p = 0.323

r = 0.001

p = 0.992

r = 0.212

p = 0.075

r = 0.995*

p = 0.000*

 

r = -0.194 p = 0.105

r = 0.764*

p = 0.000*

r = 0.478 *

p = 0.000*

r = 0.458*

p = 0.000*

r = 0.189

p = 0.114

Age

r = -0.1

p = 0.408

r = -0.145

p = 0.228

r = -0.018

p = 0.882

r = -0.196 p = 0.109

r = -0.194 p = 0.105

 

r = 0.208 p = 0.082

r = 0.195

p = 0.104

r = 0.209

p = 0.08

r = -0.319*

p = 0.007*

Likes

r = 0.056

p = 0.645

r = 0.067

p = 0.579

r = 0.154

p = 0.201

r = 0.748*

p = 0.000*

r = 0.764*

p = 0.000*

r = 0.208 p = 0.082

 

r = 0.632*

p = 0.000*

r = 0.673*

p = 0.000*

r = 0.164

p = 0.173

Dislikes

r = 0.002

p = 0.984

r = -0.070

p = 0.561

r = 0.081

p = 0.500

r = 0.418*

p = 0.000*

r = 0.478*

p = 0.000*

r = 0.195 p = 0.104

r = 0.632* p = 0.000*

 

r = 0.366*

p = 0.002*

r = -0.03

p = 0.802

Comments

r = -0.019

p = 0.872

r = 0.003

p = 0.977

r = 0.137

p = 0.254

r = 0.430*

p = 0.000*

r = 0.458*

p = 0.000*

r = 0.209 p = 0.08

r = 0.673* p = 0.000*

r = 0.366* p = 0.002*

 

r = 0.099

p = 0.414

*p < 0.05; VPI, video power index; JAMA, Journal of the American Medical Association; GQS, Global Quality Scale

Discussion

Patients who want to take a more active role in decision-making about their treatment are increasingly using the internet to seek information about their disease [8]. Achieving the right information from reliable sources increases patient satisfaction and can improve treatment results [911]. However, in our assessment of the quality of information presented in YouTube videos on FECD, we found that these videos provided only moderate-quality information about the disease.

Similar to our study, studies evaluating YouTube videos in the literature have reported that video content was generally of poor-moderate quality. In a YouTube study related to strabismus, Mangan et al. reported the mean DISCERN, JAMA, and GQS scores as 42.2 ± 15.3, 1.9 ± 1.2 and 2.7 ± 1.1, respectively and concluded that these videos provided only moderate-quality information about strabismus [3]. In a study by Küçük et al., in which Youtube videos were evaluated as an educational resource in refractive surgery, the DISCERN, JAMA and GQS scores were determined to be 33.25 ± 15.34 (poor), 0.74 ± 0.82 and 1.74 ± 0.81, respectively, and the authors reported that YouTube videos did not adequately inform patients about surgical procedures [12]. Sahin et al. found that 64% of YouTube videos on retinopathy of prematurity were useful, Gutrie et al. reported that 32% of videos on retinitis pigmentosa were useful, and Hickman et al. observed that only 23% of videos on neurological eye movements were of excellent quality [1315].

In our study, no significant difference was found between the video upload sources in terms of the DISCERN, JAMA benchmark, GQS and VPI scores. Unlike other studies, there were no patient-uploaded videos, and independent users had posted only 2.8% of the evaluated videos.3 Many studies have shown that the quality and reliability of videos uploaded by physicians are higher than those uploaded by non-physicians, but the former have a lower viewing rate [16]. In our study, no difference was found between the video sources in terms of the number of views.

When evaluated according to video content, there was no difference between the videos providing information about the disease and those discussing surgical techniques in terms of the DISCERN, JAMA, GQS and VPI values. In the videos presenting patient experience, the DISCERN, JAMA and GQS scores were lower compared to the videos discussing surgical techniques and VPI was lower than two groups. Unlike other studies, the popularity of the videos on patient experience was found to be lower [3, 17, 18]. In these videos, it is not possible to distinguish whether the patients describe their own experience and thoughts or they share this information for a fee for advertisement purposes. Our findings indicate that patients prefer more informative, accurate and reliable sources about FECD.

In our study, we did not separate the videos as useful or misleading, but there was no inaccurate, misleading, or non-evidence-based information in the FECD-related videos. One reason for the low DISCERN, JAMA and GQS scores ​​is that these parameters alone may not actually be sufficient to evaluate YouTube videos. The DISCERN questionnaire was developed in recognition of the need for a general set of quality criteria for written consumer health information [19].The JAMA benchmark was also created to assess online sources for the delivery of medical information [20]. Videos differ from web pages in several aspects. First, most videos focus on a single topic, such as genetics, symptoms, diagnostic methods, and treatment techniques rather than covering all aspects of the disease, and therefore they receive low scores in quality and reliability evaluations. Second, YouTube videos do not mention references or sources used for content, they do not declare the presence of any conflict of interest, and they are not regularly updated in contrast to online websites.

There are some limitations to the current study. The primary limitation of our study was analyzing only the first 100 videos on YouTube upon searching the keyword ‘Fuch’s endothelial corneal dystrophy’. A single time point was used to assess the quality of the videos, but search results may vary over time. In addition, the current study investigated only the videos in English language. Although this makes it hard to generalize the results of the study, English is accepted as the prevailing language among internet users. Lastly, we did not use a form to record the evaluated parameters for the FECD-related videos.

In conclusion, this was the first study to investigate the quality of FECD videos on a popular video-sharing site. The quality of information in YouTube videos on FECD was determined to be moderate. There was no difference in the quality and reliability of the videos according to the upload source or video content except patient experience. Ensuring that there are more videos providing healthcare information created by healthcare professionals and refined by a professional review process can increase public health awareness, and the internet can be a useful tool for delivering this high-quality and reliable information to the public.

Declarations

Acknowledgements: 

All authors contributed to the study conception and design. Material preparation, data collection and analysis were performed by RB, YO and EY. The first draft of the manuscript was written by RB and all authors commented on previous versions of the manuscript. 

 Funding: The authors declared that this study received no financial support. 

Compliance with ethical standard 

Conflict of interest: The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the paper. 

Ethical approval: All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards. 

Informed consent: Not applicable

References

  1. Wasserman M, Baxter NN, Rosen B, Burnstein M, Halverson AL (2014) Systematic review of internet patient information on colorectal cancer surgery. Dis Colon Rectum 57:64–69. https://doi.org/10.1097/DCR.0000000000000011
  2. Erdem MN, Karaca S (2018) Evaluating the accuracy and quality of the information in kyphosis videos shared on YouTube. Spine 43:1334–1339. https://doi.org/10.1097/BRS.0000000000002691
  3. Mangan MS, Cakir A, Yurttaser Ocak S, Tekcan H, Balci S, Ozcelik Kose A (2020) Analysis of the quality, reliability, and popularity of information on strabismus on YouTube. Strabismus 28:175–180. https://doi.org/10.1080/09273972.2020
  4. Matthaei M, Hribek A, Clahsen T, Bachmann B, Cursiefen C, Jun AS (2019) Fuchs Endothelial Corneal Dystrophy: Clinical, Genetic, Pathophysiologic, and Therapeutic Aspects. Annu Rev Vis Sci 5:151–175. https://doi.org/10.1146/annurev-vision-091718-014852
  5. Sarnicola C, Farooq AV, Colby K (2019) Fuchs endothelial corneal dystrophy: update on pathogenesis and future directions. Eye Contact Lens 45(1):1–10. https://doi.org/10.1097/ICL.0000000000000469
  6. Gain P, Jullienne R, He Z, Aldossary M, Acquart S, Cognasse F, Thuret G. Global Survey of Corneal Transplantation and Eye Banking. JAMA Ophthalmol. 2016;134:167–173. https://doi.org/ 10.1001/jamaophthalmol.2015.4776
  7. Blitzer AL, Colby KA (2020) Update on the Surgical Management of Fuchs Endothelial Corneal Dystrophy. Ophthalmol Ther 9:757–765. https://doi.org/10.1007/s40123-020-00293-3
  8. Singh AG, Singh S, Singh PP (2012) YouTube for information on rheumatoid arthritis–a wakeup call? J Rheumatol 39:899–903. https://doi.org/10.3899/jrheum.111114
  9. Schallhorn SC, Hannan SJ, Teenan D, Pelouskova M, Schallhorn JM (2018) Informed consent in refractive surgery: in-person vs telemedicine approach. Clin Ophthalmol 12:2459. https://doi.org/10.2147/OPTH.S183249
  10. Pager C (2005) Randomised controlled trial of preoperative information to improve satisfaction with cataract surgery. Br J Ophthalmol 89:10–13. https://doi.org/10.1136/bjo.2004.048637
  11. Shukla AN, Daly MK, Legutko P (2012) Informed consent for cataract surgery: patient understanding of verbal, written, and videotaped information. J Cataract Refract Surg 38:80–84. https://doi.org/10.1016/j.jcrs.2011.07.030
  12. Kuçuk B, Sirakaya E (2020 Apr) An Analysis of YouTube Videos as Educational Resources for Patients About Refractive Surgery. Cornea 39(4):491–494. https://doi.org/10.1097/ICO.0000000000002237
  13. Şahin A, Şahin M, Türkcü FM (2019 May) YouTube as a source of information in retinopathy of prematurity. Ir J Med Sci 188(2):613–617. https://doi.org/10.1007/s11845-018-1902-2
  14. Guthrie G, Davies R, Fleming C, Browning AC (2014) YouTube as a source of information about retinitis pigmentosa. Eye 28:499. https://doi.org/10.1038/eye.2013.312
  15. Hickman SJ (2016) An evaluation of educational neurological eye movement disorder videos posted on internet video sharing sites. J Neuroophthalmol 36:33–36. https://doi.org/10.1097/WNO.0000000000000292
  16. Azer SA. Are DISCERN and JAMA Suitable Instruments for Assessing YouTube Videos on Thyroid Cancer? Methodological Concerns. J Cancer Educ. 2020 Dec;35(6):1267–1277. https://doi.org/10.1007/s13187-020-01763-9
  17. Celik H, Polat O, Ozcan C, Camur S, Kilinc BE, Uzun M (2020) Assessment of the Quality and Reliability of the Information on Rotator Cuff Repair on YouTube. Orthop Traumatol Surg Res 106:31–34. https://doi.org/10.1016/j.otsr.2019.10.004
  18. Ferhatoglu MF, Kartal A, Ekici U, Gurkan A (2019) Evaluation of the Reliability, Utility, and Quality of the Information in Sleeve Gastrectomy Videos Shared on Open Access Video Sharing Platform YouTube. Obes Surg 29:1477–1484. https://doi.org/10.1007/s11695-019-03738-2
  19. Charnock D, Shepperd S, Needham G, Gann R (1999) DISCERN: an instrument for judging the quality of written consumer health information on treatment choices. J Epidemiol Community Health 53:105–111. https://doi.org/10.1136/jech.53.2.105
  20. Silberg WM, Lundberg GD, Musacchio RA (1997) Assessing, controlling, and assuring the quality of medical information on the Internet: Caveant lector et viewor–Let the reader and viewer beware. JAMA 277:1244–1245