Skip to main content

YouTube online videos as a source for patient education of cervical spondylosis—a reliability and quality analysis

A Correction to this article was published on 05 April 2024

This article has been updated

Abstract

Background

Given a prolonged course of Cervical spondylosis (CS) could cause irreversible neurological deficits, it is crucial to disseminate CS-related health information to the public to promote early diagnosis and treatment. YouTube has been widely used to search for medical information. However, the reliability and quality of videos on YouTube vary greatly. Thus, this study aimed to assess the reliability and educational quality of YouTube videos concerning CS and further explore strategies for optimization of patient education.

Methods

We searched YouTube online library for the keywords “cervical spondylosis”, “cervical radiculopathy” and “cervical myelopathy” on January 15, 2023. Ranked by “relevance”, the first 50 videos of each string were recorded. After exclusions, a total of 108 videos were included. All videos were extracted for characteristics and classified based on different sources or contents. Two raters independently evaluated the videos using Journal of American Medical Association (JAMA) benchmark criteria, Modified DISCERN (mDISCERN) tool, Global Quality Scale (GQS) and Cervical-Spondylosis-Specific Scale (CSSS), followed by statistical analyses. All continuous data were described as median (interquartile range).

Results

All videos had median values for JAMA, mDISCERN, GQS and CSSS scores of were 3.00 (1.00), 3.00 (2.00), 2.00 (1.00) and 7.00 (8.88), respectively. There were significant differences in VPI (P = 0.009) and JAMA (P = 0.001), mDISCERN (P < 0.001), GQS (P < 0.001) and CSSS (P < 0.001) scores among different sources. Videos from academic source had advantages in reliability and quality scores than other sources. VPI (P < 0.001), mDISCERN (P = 0.001), GQS (P < 0.001) and CSSS (P = 0.001) scores also significantly differed among videos of various contents. Spearman correlation analysis indicated VPI was not correlated with either reliability or quality. Multiple linear regression analysis showed a longer duration and an academic source were independent predictors of higher reliability and quality, while a clinical source also led to the higher video quality.

Conclusions

The reliability and educational quality of current CS-related videos on YouTube are unsatisfactory. Users face a high risk of encountering inaccurate and misleading information when searching for CS on YouTube. Longer duration, source of academic or clinician were closely correlated to higher video reliability and quality. Improving the holistic reliability and quality of online information requires the concerted effort from multiple parties, including uploaders, the platform and viewers.

Peer Review reports

Background

Cervical spondylosis (CS) is a chronic and progressive degenerative process of the cervical spine featuring the pathological change of vertebrae, joints, intervertebral discs, and other relevant structures. [1] As the common cause of neurological dysfunction in adults, CS may progress into cervical radiculopathy (CR) and cervical myelopathy (CM), with principal symptoms of pain, paresthesia and muscle weakness in the neck and/or extremities. [2, 3] In severe cases, additional manifestations such as bladder problems, ataxia, restricted motion, and even paralysis may ensue, portending poor outcomes. [4] Also, CS could be simultaneous with several comorbidities of depression, anxiety and sleep disorders etc. [5,6,7] Due to demographic aging and shifting lifestyle patterns, the prevalence of symptomatic cervical spondylosis is increasing, with up to 13.76% reported by a community-based study. Furthermore, onset is occurring at increasingly younger ages. [8] This has engendered a hefty burden on global economics and productivity, emerging as a salient public health challenge worldwide. [9, 10].

Given that the protracted course of CS could result in an irreversible detrimental outcome, [1, 4] early diagnosis and appropriate management are essential to obviate lifelong disability. However, the information asymmetry between doctors and general public perturbs patients aware of their conditions. Therefore, it is imperative to provide precise health education about CS to the public.

With the development of the internet and the exponential growth of digital information, an increasing number of netizens choose to consult and seek disease-related information online. [11, 12] YouTube, as the second most visited and the most popular video sharing site, has more than 22.8 billion visits per month and yields videos about specific medical information and tutorials that contain potential applicability for patient education and even management. [13] However, since the lack of the regulation, the quality of video information from YouTube is uneven [14, 15], especially concerning medical field. Substandard, inadequate or erroneous information diverges from professional clinical advice, engendering wrong perceptions in patients and leading to inappropriate self-assessment of their conditions. [16, 17] This may interfere with proper medical management for individuals and result in disease deterioration. For society as a whole, it could aggravate the doctor-patient relationship.

Hitherto, to the best of our knowledge, no research has quantified the reliability and quality of CS-related information available on YouTube. Herein, this study aims to assess the reliability and educational quality of YouTube videos regarding CS, and analyze relevant influencing factors, furtherly exploring strategies to optimize the quality of online health resources and guide patients to obtain valuable medical information related to CS.

Methods

Ethics approval and information consent

Since data from YouTube was all publicly available and no patients were involved in this study, the ethics committee approval and information consent were not required.

Search strategy and video characteristics

To evaluate the reliability and quality of online videos, we standardly searched for the keywords “cervical spondylosis”, “cervical radiculopathy” and “cervical myelopathy” on YouTube online library (https://www.youtube.com, January 15, 2023). Before conducting the search, we created an exclusive account and cleared browser cache and search history, also turning off data recording to avoid potential influence on the results. By default settings and ranking of “relevance”, we preliminarily recorded the first 50 videos in results of each string. This strategy could simulate the common browsing habit in most viewers and had been reported feasible in previous literatures. [18, 19] Furthermore, duplicate, irrelevant, non-English, audio-only, vision-only video, shorts or video with unacceptable audio/visual quality was excluded in this study. Finally, 108 videos were included for following analysis (Fig. 1).

Fig. 1
figure 1

Flowchart of videos selection on YouTube

The following data of video characteristics were extracted: (1) Title and URL, (2) uploader, (3) date of publication, (4) time since uploaded (till January 15, 2023), (5) video duration, (6) number of views, (7) number of likes, (8) number of dislikes, (9) number of referrers, (10) number of comments, (11) view ratio (views/day), (12) like ratio (like × 100/(like + dislike)), (13) uploader verification, (14) country/region of origin, (15) total views of host channel, (16) sum of subscribers of host channel and (17) Video Power Index (VPI). The specific calculation of VPI was as follows: view ratio × like ratio/100. VPI was an objective index to quantify the video popularity and had been applied in previous research. [20,21,22] Some of the indexes were collected through the reliable third-party browser plugin. [23, 24].

Classification of video sources and contents

In accordance to previous studies [12, 25] and the actual searching results, the videos were sorted into six categories based on the source: (1) academic (uploaders affiliated with universities, colleges or research groups); (2) clinician (individual clinician or clinician groups without affiliation of academic institutions); (3) non-clinician (allied health workers other than licensed clinicians: physiotherapists etc.); (4) trainer; (5) medical source (health-related channels or websites) or (6) commercial source (corporations or for-profit organizations).

The videos were also classified into the following categories based on the content: (1) exercise training (exercise related to CS); (2) CS-specific information (pathophysiology, examinations, diagnosis etc.); (3) surgical technique; (4) non-surgical management; (5) advertisement. A single video was limited to one theme. If a video involved several content topics, the content that occupied the largest proportion or which viewers gained most from the video would be determined.

Evaluation of video reliability and educational quality

Journal of American Medical Association (JAMA) benchmark criteria (Table 1), proposed by Silberg et al. [26], were used to evaluate the information reliability of included videos. Each of the four core standards (authorship, attribution, disclosure, and currency) is assigned one point, and the total JAMA score is calculated by summing up the fulfilled criteria. A maximal score of four represents the highest accuracy and reliability, whereas a score of zero indicates poor accuracy and reliability. Additionally, DISCERN tool, which was originally proposed by Charnock et al. and modified by singh et al., was adopted to verify the video reliability from another perspective (Table 2). [27, 28] Modified DISCERN (mDISCERN) tool is based on five binary yes/no questions, with every positive answer gaining one point and a maximal score of five indicating high reliability. Global Quality Scale (GQS) [29] was utilized for non-specific evaluation of videos’ educational quality (Table 3). GQS is a five-grade scale that ranges from one to five grades, with higher grades standing for higher quality.

Table 1 Journal of American Medical Association (JAMA) benchmark criteria
Table 2 Modified DISCERN (mDISCERN) criteria
Table 3 Global quality scale (GQS) criteria

There was no existing method to assess the educational content of CS videos specifically and comprehensively. Combining opinions from previous articles, reviews, guidelines and our clinical practice, we developed a novel scoring system entitled Cervical-Spondylosis-Specific Scale (CSSS). CSSS comprised four sections (information about CS, evaluation and diagnosis, treatment and postoperative course) and 19 sub-items in total. Diverse points were allocated to each item based on different priority and value. The total score was calculated by summing up the corresponding point(s) for all fulfilled items, with a maximum of 25 points indicating the highest educational quality for CS. Under this evaluation system, a high-quality CS-related video needs to elucidate the following information: the typical symptom of CS (neck pain, radiating pain, paresthesia etc.) and general nosogenesis (compression) and risk factors (age, poor postural habit, high loads); the main classification of CS (radiculopathy, myelopathy, etc.); diagnostic methods (physical examination, diagnostic imaging, differential diagnosis); treatment strategies (non-surgical and surgical options, highlight of the difference between treatments for CR and CM); posttreatment course (natural history, prognosis, complications). More specific items were shown in Table 4. Although CSSS was not validated, the similar structure for disease-specific scale had been broadly used for evaluation of video quality in peer-reviewed studies and been proven feasible. [30,31,32].

Table 4 Cervical-spondylosis-specific scale (CSSS) criteria

Two independent orthopedic doctors (H.W., C.Y.) assessed all the included videos using above scoring systems and repeated once after two weeks. For a single video, both the video itself and its description were taken into account. The original results assessed by two doctors were recorded separately. Any discrepancy between both was arbitrated by the third reviewer (H.L.) to achieve a unanimous result.

Statistical analysis

Video characteristics, reliability and educational quality were quantified by descriptive statistics. The missing data were processed using multiple Imputation. All continuous data in the study were described as median (interquartile range) as they didn’t comply normality. Qualitative data were expressed as fractions. Kruskal–Wallis test was utilized to evaluate intergroup differences of variables based on different video sources or contents, followed by post-hoc analyses of Bonferroni correction. Spearman correlation analysis was used to explore the correlation among VPI, JAMA, mDISCERN, GQS and CSSS scores. Then multiple linear regression analysis was applied to determine the independent predictor of the five above indexes from video characteristics, sources and contents. Variables that showed a univariate relationship (P < 0.20) with the target index or that were considered or reported relevant were incorporated into the regression model. Variables were carefully selected based on the number of events available to ensure the final model was parsimonious. [33] The categories of medical source and CS-specific information were set as the reference dummy variable of video sources and video contents, respectively, for the unified comparisons.

Intraobserver and interobserver agreements of four scoring systems (JAMA, mDISCERN, GQS, CSSS) were appraised by intraclass correlation coefficient (ICC), respectively. ICC values were interpreted referring to the guidelines: excellent (> 0.90), good (0.75–0.90), moderate (0.50–0.75) and poor (< 0.50). [34].

All the statistical analyses were performed by IBM SPSS Statistics v.26 (IBM Corp., Armonk, New York, USA). All reported P values were two-sided, and the difference was considered statistically significant when the P value < 0.05.

Results

Video baseline characteristics

Of the 150 videos screened, 108 were eligible and included in the study. The earliest video was published on April 13, 2010. Among all included videos, the number of each year generally increased by time, and more than a half (54.63%, 59/108) were released in 2019–2022 (Fig. 2A). Excluding videos with unknown original country, the top three countries that produced the most videos were USA (69), Indian (12) and UK (5) (Fig. 2B). Through analysis of all videos, the median value of time since uploaded, video duration and view count of each video were 1329.00 (1629.00) days, 263.00 (511.00) seconds and 18,551.50 (62,456.00), respectively. Every video received the median likes and dislikes count of 201.00 (846.75) and 5.00 (26.75) with the median referrers of 0.00 (1.00) and the median comments of 16.00 (62.00). The median view ratio, like ratio and VPI were calculated by established formulas to be 19.20 (67.21), 97.99 (4.24) and 18.49 (67.06), respectively. Nineteen videos were shared by officially verified uploaders. Every video host channel has the median total views of 7,000,000.00 (37,348,325.00) and the median subscribers sum of 49,550.00 (337,895.00) (Table 5).

Fig. 2
figure 2

Video counts of each year from 2010 to 2022 (A); Distribution of videos based on original countries (B)

Table 5 The baseline characteristics and evaluation results of involved YouTube videos

Video sources and contents

All the videos were sorted into six sources: Medical source led the largest share (30/108, 28%), followed by non-clinician (27/108, 25%), academic (18/108, 17%), clinician (17/108, 16%), commercial source (12/108, 11%), and trainer (4/108, 3%). (Fig. 3A).

Fig. 3
figure 3

Categorical distribution of videos based on sources (A); Categorical distribution of videos based on contents (B)

Among five categories based on video content, CS-specific information was the most frequently covered (47/108, 43%), followed by exercise training (20/108, 19%), surgical technique (18/108, 17%), non-surgical management (13/108, 12%), and advertisement (10/108, 9%) (Fig. 3B).

The heatmap provided an abstract representation of the number of videos with different content from different sources. Academics and clinicians focused more on CS-specific information and surgical techniques, while non-clinicians focused more on exercise training and CS-specific information. Trainers only produced videos about exercise training. Videos from medical source covered all topics, with a greater emphasis on CS-specific information. And videos from commercial source merely involved CS-specific information or non-surgical management (Fig. 4).

Fig. 4
figure 4

Heatmap of video counts concerning from different sources and contents

Video reliability and educational quality

The median values for JAMA, mDISCERN, GQS and CSSS scores of all videos were 3.00 (1.00), 3.00 (2.00), 2.00 (1.00) and 7.00 (8.88), respectively (Table 5). The Intraobserver reliability within each rater was excellent for all four scales. Two raters achieved excellent agreement in JAMA scores (ICC: 0.906, 95% confidence interval (CI): 0.860–0.937), GQS scores (ICC: 0.927, 95% CI: 0.895–0.949), CSSS scores (ICC: 0.939, 95% CI: 0.908–0.960); and good agreement in mDISCERN scores (ICC: 0.888, 95% CI: 0.831–0.925).

There were significant differences in VPI (P = 0.009) and JAMA (P = 0.001), mDISCERN (P < 0.001), GQS (P < 0.001) and CSSS (P < 0.001) scores among different video uploading sources. The only significant discrepancy in VPI existed between videos from non-clinician and medical source (28.06 (111.34) vs. 12.34 (19.04), P < 0.05). Videos from academic source showed significant (P < 0.05) higher scores in the four scales than any other source, except for clinician source in GQS and CSSS (P > 0.05). Meanwhile, VPI (P < 0.001), mDISCERN (P = 0.001), GQS (P < 0.001) and CSSS (P = 0.001) scores significantly differed among videos of various contents. Videos about exercise training had higher VPI than those covering CS-specific information, surgical technique and non-surgical management with significance (P < 0.05). In contrast, the contents of surgical technique and non-surgical management led to significantly (P < 0.05) higher scores in mDISCERN, GQS and CSSS scales than exercise training. More detailed analytical results are displayed in Table 6.

Table 6 The median VPI values, median scores of reliability and quality of different video source or content

Factors affecting video popularity, reliability and educational quality

Spearman correlation analysis revealed positive and significant correlations among every pair of JAMA, mDISCERN, GQS, and CSSS scores (P < 0.001 for each pair). However, none of these scores were significantly correlated with VPI (Table 7).

Table 7 Spearman correlation analysis between VPI, JAMA, mDISCERN, GQS and CSSS

Multiple linear regression analysis showed that a higher VPI was correlated with a higher number of comments (P < 0.001), a verified uploader (P = 0.034), fewer subscribers to the host channel (P = 0.011). And compared to the CS-specific information, the content about exercise was a independent predictor to higher VPI (P = 0.005); A higher JAMA score was associated with longer video duration (P < 0.001), greater like ratio (P < 0.001), a verified uploader (P = 0.002). The videos from academic source were correlated to higher JAMA scores than medical source (P = 0.003); A higher mDISCERN score was closely related to longer video duration (P < 0.001), greater like ratio (P < 0.001). In comparison to medical source, the sources of academic (P < 0.001) and trainer (P = 0.001) were associated with higher and lower mDISCERN scores, respectively; A higher GQS score was correlated to longer video duration (P < 0.001). The sources of academic (P = 0.001) and clinician (P = 0.002) were independent predictors of higher GQS scores compared to medical source, and the contents about exercise training (P = 0.021) and advertisement (P = 0.009) were related to lower GQS scores than the CS-specific information; A higher CSSS score was in correlation with longer video duration (P < 0.001). And Compared to the medical source, the sources of academic (P = 0.005) and clinician (P = 0.006) were associated with higher CSSS scores while the sources of non-clinician (P = 0.024) and trainer (P = 0.033) were related with lower CSSS scores.

Discussion

Given the neurological deficit caused by chronic CS course [1], it is crucial to promote early diagnosis and treatment. The internet provides another dimension to balance the information asymmetry between doctors and patients. Studies show that about 70–80% of netizens and 30% of orthopedic patients utilize the internet to acquire health information [35, 36], building their preliminary perception. With the advent of 5G technology, videos have emerged as a widely accepted medium for conveying information on the internet. This has led to the rise of numerous international visual websites, with YouTube being a prominent example. Unfortunately, the internet is replete with inaccurate and misleading information that can shape patients’ perspectives on their ailments in ways that are often at odds with professional recommendations, thereby reducing patient compliance. As clinicians, we may not be able to edit or correct all the public information shared on the web. However, we should at least understand the online information that patients receive, how it shapes their cognition, and how it can be optimized. This motivated the authors to undertake an exploration and evaluation of CS-related videos on YouTube.

The current results indicate that the reliability and educational quality of CS-related videos on YouTube are unsatisfactory, with the median JAMA, mDISCERN, GQS and CSSS scores of 3.00 (1.00), 3.00 (2.00), 2.00 (1.00) and 7.00 (8.88), respectively. This suggests that the common netizens or patients searching for information about CS on YouTube may be at a relatively high risk of encountering inadequate, inaccurate, or even misleading information. Our finding is consistent with previous research in the field of spinal health. Erdem et al. revealed the poor quality of videos covering kyphosis on YouTube, which had the mean JAMA, GQS and Kyphosis-Specific Scores (range: 0–32) of 1.36, 1.68, 3.02, respectively. [37] Similarly, Stogowski et al. studied 24 YouTube videos about anterior lumbar interbody fusion and concluded that the overall quality remains poor, with the mean DISCERN score of 38.21/75. [38] A detailed review about representative studies focusing on YouTube information of spine field was demonstrated in Table 8. In contrast, a few studies proposed a more positive attitude towards medical videos on YouTube. Unal-Ulutatar et al. searched for “systemic sclerosis” and “scleroderma” and determined 73% (84/115) of the videos were useful. [39] Ng et al. highlighted that there was an abundance of reliable and of high-quality YouTube videos with useful information on systemic lupus erythematosus. [40] The discrepancy in these conclusions may be attributed to the differences in studied fields and source proportions.

Table 8 Literature review of studies concerning YouTube information of spine fields

The low reliability and quality of medical videos on YouTube may be due to the absence of an access and censorship system. [30, 46] This allows unqualified individuals to publish videos on medical topics, potentially spreading unprofessional and unsubstantiated information. Of note that some videos were titled with phrases such as “treat with exercises”, “no surgery”, “best treatment”, which appeals to patients’ psychological needs. These videos contained considerably subjective statements with bias and may propagate inappropriate perception about the disease to viewers. Meanwhile, the recent fast-food culture (FFC) in video industry leads uploaders to create short, fast-paced video products or split videos into series, controlling the duration of less than 10 min to cater viewers’ preference [47]. In our study, the median duration of all involved videos was 263.00 (511.00) seconds (Table 5), and 73.15% (79/108) of videos were within 10 min long. This trend may inevitably fragment the intact information presented in a single video, leaving viewers with inadequate and unsystematic concepts.

In terms of specific content, the included videos were quite homogenous and shared several common issues. Nearly half (43%) of the videos discussed CS-specific information (Fig. 3B). Most videos covered the basic nosogenesis and typical symptoms of CS, but generally lacked deeper differentiation between CR and CM, which differ in severity, course and interfering methods. Videos should elucidate the divergence of their manifestations and managements precisely, and highlight the urgence and importance of early diagnosis and surgical intervention for CM, alerting viewers for accurate self-evaluations. In addition, the majority of videos pertaining to surgical techniques merely broached the fundamental concept, while few furnished detailed elucidations of the indications and advantages of various surgical approaches, which is a major concern for some viewers. In our outpatients, we often encountered patients who insisted on minimally invasive surgery without considering the objective fact that if the operative range is enough for the thorough decompression. It is challenging to coordinate with these patients who have preconceived notions and expectations. Therefore, it is necessary to inculcate them with comprehensive, objective and evidenced information about disease from the outset.

Despite the generally low quality of videos, there were still some of high caliber. For example, “cervical myelopathy and cervical radiculopathy- Everything You Need To Know—Dr. Nabil Ebraheim” from “nabil ebraheim”, “Cervical Radiculopathy—Why do you hurt and what is the plan to get you better?” from “Armaghani Spine” and “Exercises for pinched nerve in the neck (Cervical Radiculopathy) and neck pain relief” from “Dr. Andrea Furlan” were the top three videos that showed distinguished performance under our evaluation system. Our analysis revealed that the high-reliability and high-quality of videos tend to coexist with each other, and associated with certain video characteristics. Apparently, video duration is a crucial factor for reliability and quality, with a significant positive regression among them (Table 9). This finding was supported by other research. [48,49,50] Longer running time allows for more comprehensive coverage of topics and provide more educational information. Videos from academic source had significant advantages in both reliability and educational quality (Tables 6 and 9). These videos were produced by educational experts or groups in the spine field with high academic literacy, and were primarily aimed at clinicians or medical students, who demand higher breadth and accuracy of content. The source of clinicians could also be considered as an independent predictor of high quality (Table 9). Clinicians possess extensive clinical experience that could better meet the needs of users and patients. As increased licensed doctors participate in We-Media and share their clinical experience online, they achieve another avenue of communication with patients and gain considerable popularity, which is commendable. However, videos from clinicians did not demonstrate superiority in terms of reliability, suggesting that clinicians may not place sufficient emphasis clarifying reference sources, copyrights and qualifications etc.

Table 9 Multiple linear regression analysis of correlations between video characteristics and VPI, JAMA, mDISCERN, GQS, NPSS scores

The VPI, i.e. video popularity, comprises two components: views and likes. [30] Spearman correlation analysis indicated that there was no significant relationship between VPI and reliability or quality (Table 7). A higher VPI was associated with a greater number of comments (Table 9), showing better engagement with the video. The source of trainer and the corresponding content of exercise training may lead to a higher VPI (Fig. 4, Tables 6 and 9). Unfortunately, both were not able to predict higher reliability and quality, or even poorer than other sorts (Tables 6 and 9). Those videos may easily gain the favor of viewers, but lack in-depth elaboration of CS. On the contrary, videos about medical management, including both surgical or non-surgical categories, had relative advantages in reliability and quality (Table 6). However, neither topic showed a correlation with higher VPI. The results revealed that videos with higher dissemination value did not receive commensurate levels of attention. To address the conflict between video popularity and quality, it needs deeper consideration on what indeed influence the VPI.

By deconstructing the VPI into two sub-indexes, we further focused on the view ratio. The diagram demonstrated that those shorter videos exceled in engaging viewers (Fig. 5), however, being short in quality, as mentioned above. This precisely reflected the so called FFC. While videos from high-caliber uploaders, like academics and clinicians, had relatively longer duration, occupying 33% of the video sum and 52.34% (36,863/70425) of total duration, but only 29.55% (3,569,505/12078971) of total views. These authorities or official channels often had lower rank and fewer subscribers. Undeniably, long videos tend to exert greater viewing pressure, occupying more real-life time and imposing amount of information to the audience to absorb, which may lead to potential resistance from users. In the era chasing for efficiency, the FFC which is naturally facilitated by viewers has its own rational grounds.

Fig. 5
figure 5

Distribution of duration and corresponding view ratio of each video

The simply short videos yet to have incontrovertible dominance to users that shorter duration didn’t independently predict a higher VPI (Table 9). And as another component of VPI, the importance of the like ratio should be emphasized. Notably, higher like ratio was independently associated with higher JAMA and mDISCERN scores, i.e., reliability (Table 9). This suggested that reliable and high-quality videos could ultimately gain viewers’ positive feedback. However, the viewing propensity are mainly determined by superficial and brief information on the index page, such as title, cover, duration and uploader etc. The specific content quality couldn’t exert direct influence on viewers’ choices. Meanwhile, as the ability to access, integrate and absorb online information varies and is generally limited when dealing with knowledge from other fields among lay users, the cognitions from them towards high-quality professional content are insensitive and unprecise. As Staunton et al. noted, higher-quality information may not always be in a readable manner to engage users. [51] A common problem currently plagues the high-quality videos is the poor comprehensibility that were too profound and lengthy for users to efficiently absorb the information. These prompts us there is potential and need to direct viewers from purely short and coarse videos to the high-quality videos with more reliable, concise and comprehensible information.

Enhancing the quality of internet medical information and optimizing the online patient education requires the concerted efforts of multiple parties. One solution to improve the overall reliability and educational quality of CS-related videos on YouTube may be to motivate academics and clinicians to produce more contents. As professionals, academics and clinicians should devote more effort on popularizing medical knowledge for the public. The dominance of short videos should be reconsidered. There is a need for deeper consideration of how to extract important, systematic knowledge from expertise and present it comprehensibly within a limited duration. Lightweight but not necessarily short videos could benefit to both gaining preference and delivering accurate information. Simultaneously, the professional videos could be wrapped and presented in an appropriate way that caters to the viewers’ mentality to direct them to the high-quality content more effectively. In another aspect, it is essential to strengthen the cognition of information reliability of the uploaders. The importance of unbiased and evidence-based information, copyright awareness and disclosure of interests etc. should be emphasized to engage users and, more importantly, to reduce the possibility of users being exploited by commercial interests. For the YouTube platform, advanced artificial intelligence could be used to establish more stringent admission requirements and screening systems to resist poor and misleading contents. Additionally, combining with big-data analysis, the platform could selectively promote and push high-quality videos and relevant evidence-based information to the target population. Last but not least, the public should improve their ability to obtain and utilize the internet information dialectically, striving to understand the underlying mechanisms and scientific managements of the diseases.

Our study had some unavoidable limitations. Given the popularity and clout, we only included English videos on YouTube as subjects, which may induce selective bias and reduce external validity. The quality scales, GQS and CSSS, were subjective and lacked strict validation. Although we adopted a double-review process, confounders were inevitable. Additionally, due to the inherent limitations of the scoring systems, our assessment was limited to the breadth of the covered topics in a video and did not evaluate the comprehensibility and efficiency of specific information delivering for the common users. It should also be mentioned that our statistics were limited by timeliness. Most characteristics data of videos changed dynamically and may not be representative for all periods. Besides, users may obtain more complete CS-related information from multiple complementary videos on YouTube. The results evaluated from each single video may not accurately quantify the holistic information that users could perceive from YouTube. More comprehensive evaluation methods remain to be explored. Beyond the scope of this study are other factors that may affect video popularity, such as potential social and commercial influences. Therefore, our findings about video popularity should be considered with caution.

Conclusions

The internet has shown great potential in the medical field for patient education. However, the quality of online information is uneven and unregulated. This study indicated that the overall reliability and educational quality of current CS-related videos on YouTube are unsatisfactory. Users and patients searching for CS on YouTube are at high risk of encountering inadequate, inaccurate, or even misleading information. The videos with longer duration or from academic or clinician source could lead to higher reliability and quality. Optimizing the overall reliability and quality of online information requires a collaborative effort from multiple parties. We suggested motivating academics and clinicians to produce more concise and accurate contents. Meanwhile, the platform needs to establish stringent admission requirements and screening systems. And the public should endeavor to obtain and utilize internet information critically.

Availability of data and materials

The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.

Change history

Abbreviations

CS:

Cervical spondylosis

CR:

Cervical radiculopathy

CM:

Cervical myelopathy

VPI:

Video Power Index

JAMA:

Journal of American Medical Association

mDISCERN:

Modified DISCERN

GQS:

Global Quality Scale

CSSS:

Cervical-Spondylosis-Specific Scale

FFC:

Fast-food culture

References

  1. Degenerative TN, Spondylosis C. Ropper AH, editor. N Engl J Med. 2020;383(2):159–68.

    Google Scholar 

  2. Woods BI, Hilibrand AS. Cervical Radiculopathy: Epidemiology, Etiology, Diagnosis, and Treatment. Clin Spine Surg. 2015;28(5):E251.

    Google Scholar 

  3. Badhiwala JH, Ahuja CS, Akbar MA, Witiw CD, Nassiri F, Furlan JC, et al. Degenerative cervical myelopathy — update and future directions. Nat Rev Neurol. 2020;16(2):108–24.

    Article  PubMed  Google Scholar 

  4. Davies BM, Mowforth OD, Smith EK, Kotter MR. Degenerative cervical myelopathy. The. BMJ. 2018;22(360):k186.

    Article  Google Scholar 

  5. Lin SY, Sung FC, Lin CL, Chou LW, Hsu CY, Kao CH. Association of Depression and Cervical Spondylosis: A Nationwide Retrospective Propensity Score-Matched Cohort Study. J Clin Med. 2018;7(11):387.

    Article  PubMed  PubMed Central  Google Scholar 

  6. Bisson EF, Mummaneni PV, Michalopoulos GD, El Sammak S, Chan AK, Agarwal N, et al. Sleep Disturbances in Cervical Spondylotic Myelopathy: Prevalence and Postoperative Outcomes-an Analysis From the Quality Outcomes Database. Clin Spine Surg. 2023;36(3):112–9.

    Article  PubMed  Google Scholar 

  7. Chu Y, Wang X, Dai H. Prevalence and risk factors for anxiety and depression among community dwelling patients with cervical spondylosis during the COVID-19 pandemic. Heliyon. 2023;9(2):e13497.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  8. Lv Y, Tian W, Chen D, Liu Y, Wang L, Duan F. The prevalence and associated factors of symptomatic cervical Spondylosis in Chinese adults: a community-based cross-sectional study. BMC Musculoskelet Disord. 2018;11(19):325.

    Article  Google Scholar 

  9. Hurwitz EL, Randhawa K, Yu H, Côté P, Haldeman S. The global spine care initiative: a summary of the global burden of low back and neck pain studies. Eur Spine J. 2018;27(Suppl 6):796–801.

    Article  PubMed  Google Scholar 

  10. Hoy DG, Protani M, De R, Buchbinder R. The epidemiology of neck pain. Best Pract Res Clin Rheumatol. 2010;24(6):783–92.

    Article  CAS  PubMed  Google Scholar 

  11. Koller U, Waldstein W, Schatz KD, Windhager R. YouTube provides irrelevant information for the diagnosis and treatment of hip arthritis. Int Orthop. 2016;40(10):1995–2002.

    Article  PubMed  Google Scholar 

  12. Kocyigit BF, Nacitarhan V, Koca TT, Berk E. YouTube as a source of patient information for ankylosing spondylitis exercises. Clin Rheumatol. 2019;38(6):1747–51.

    Article  PubMed  Google Scholar 

  13. Statistics for YouTube, 2023. Available: https://www.youtube.com/yt/about/press/ Accessed 1 Jan 2023.

  14. Badarudeen S, Sabharwal S. Readability of patient education materials from the American academy of orthopaedic surgeons and pediatric orthopaedic society of North America web sites. J Bone Jt Surg Am. 2008;90(1):199–204.

    Article  Google Scholar 

  15. Tartaglione JP, Rosenbaum AJ, Abousayed M, Hushmendy SF, DiPreta JA. Evaluating the Quality, Accuracy, and Readability of Online Resources Pertaining to Hallux Valgus. Foot Ankle Spec. 2016;9(1):17–23.

    Article  PubMed  Google Scholar 

  16. Madathil KC, Rivera-Rodriguez AJ, Greenstein JS, Gramopadhye AK. Healthcare information on YouTube: A systematic review. Health Informatics J. 2015;21(3):173–94.

    Article  PubMed  Google Scholar 

  17. Lewis SP, Heath NL, Sornberger MJ, Arbuthnott AE. Helpful or Harmful? An Examination of Viewers’ Responses to Nonsuicidal Self-Injury Videos on YouTube. J Adolesc Health. 2012;51(4):380–5.

    Article  PubMed  Google Scholar 

  18. Morahan-Martin JM. How Internet Users Find, Evaluate, and Use Online Health Information: A Cross-Cultural Review. Cyberpsychol Behav. 2004;7(5):497–510.

    Article  PubMed  Google Scholar 

  19. Kunze KN, Krivicich LM, Verma NN, Chahla J. Quality of Online Video Resources Concerning Patient Education for the Meniscus: A YouTube-Based Quality-Control Study. Arthrosc J Arthrosc Relat Surg. 2020;36(1):233–8.

    Article  Google Scholar 

  20. Onder ME, Zengin O. Quality of healthcare information on YouTube: psoriatic arthritis. Z Rheumatol. 2023;82 Suppl 1:30–7.

  21. Gokcen HB, Gumussuyu G. A Quality Analysis of Disc Herniation Videos on YouTube. World Neurosurg. 2019;S1878–8750(19):30246–53.

    Google Scholar 

  22. Güneri FD, Forestier FBE, Forestier RJ, Karaarslan F, Odabaşi E. YouTube as a source of information for water treatments. Int J Biometeorol. 2022;66(4):781–9.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Szmuda T, Ali S, Słoniewski P. Letter to the editor regarding “A Quality analysis of disk herniation videos on YouTube.” World Neurosurg. 2019;130:570–2.

    Article  PubMed  Google Scholar 

  24. Hornung AL, Rudisill SS, Suleiman RW, Siyaji ZK, Sood S, Siddiqui S, et al. Low back pain: What is the role of YouTube content in patient education? J Orthop Res. 2022;40(4):901–8.

    Article  PubMed  Google Scholar 

  25. Bai G, Pan X, Zhao T, Chen X, Liu G, Fu W. quality assessment of YouTube videos as an information source for testicular torsion. Front Public Health. 2022;18(10):905609.

    Article  Google Scholar 

  26. Silberg WM. Assessing, controlling, and assuring the quality of medical information on the internet: caveant lector et viewor—let the reader and viewer beware. JAMA. 1997;277(15):1244.

    Article  CAS  PubMed  Google Scholar 

  27. Charnock D, Shepperd S, Needham G, Gann R. DISCERN: an instrument for judging the quality of written consumer health information on treatment choices. J Epidemiol Community Health. 1999;53(2):105–11.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  28. Singh AG, Singh S, Singh PP. YouTube for information on rheumatoid arthritis - a wakeup call? J Rheumatol. 2012;39(5):899–903.

    Article  PubMed  Google Scholar 

  29. Bernard A, Langille M, Hughes S, Rose C, Leddin D, van Veldhuyzen-Zanten S. A Systematic review of patient inflammatory bowel disease information resources on the World Wide Web. Am J Gastroenterol. 2007;102(9):2070–7.

    Article  PubMed  Google Scholar 

  30. Zhang X, Yang Y, Shen YW, Zhang KR, Ma LT, Ding C, et al. Quality of online video resources concerning patient education for neck pain: A YouTube-based quality-control study. Front Public Health. 2022;21(10):972348.

    Article  Google Scholar 

  31. Kwak D, Park JW, Won Y, Kwon Y, Lee JI. Quality and reliability evaluation of online videos on carpal tunnel syndrome: a YouTube video-based study. BMJ Open. 2022;12(4):e059239.

    Article  PubMed  PubMed Central  Google Scholar 

  32. Ss R, Nz S, Al H, S Z, Rm A, Zk S, et al. YouTube as a source of information on pediatric scoliosis: a reliability and educational quality analysis. Spine Deform. 2023;11(1). Cited 12 Jan 2023. Available from: https://pubmed.ncbi.nlm.nih.gov/35986883/

  33. Peduzzi P, Concato J, Feinstein AR, Holford TR. Importance of events per independent variable in proportional hazards regression analysis. II. Accuracy and precision of regression estimates. J Clin Epidemiol. 1995;48(12):1503–10.

    Article  CAS  PubMed  Google Scholar 

  34. Koo TK, Li MY. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J Chiropr Med. 2016;15(2):155–63.

    Article  PubMed  PubMed Central  Google Scholar 

  35. Baker JF, Devitt BM, Kiely PD, Green J, Mulhall KJ, Synnott KA, et al. Prevalence of Internet use amongst an elective spinal surgery outpatient population. Eur Spine J. 2010;19(10):1776–9.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Finney Rutten LJ, Blake KD, Greenberg-Worisek AJ, Allen SV, Moser RP, Hesse BW. Online Health Information Seeking Among US Adults: Measuring Progress Toward a Healthy People 2020 Objective. Public Health Rep. 2019;134(6):617–25.

    Article  PubMed  PubMed Central  Google Scholar 

  37. Erdem MN, Karaca S. Evaluating the Accuracy and Quality of the Information in Kyphosis Videos Shared on YouTube. Spine. 2018;43(22):E1334–9.

    Article  PubMed  Google Scholar 

  38. Stogowski P, Antkowiak L, Trzciński R, Rogalska M, Dułak NA, Anuszkiewicz K, et al. Content Quality and Audience Engagement Analysis of Online Videos for Anterior Lumbar Interbody Fusion. World Neurosurg. 2022;160:e636–42.

    Article  PubMed  Google Scholar 

  39. C UU, F U. YouTube as a source of information on systemic sclerosis. Int J Rheum Dis. 2022;25(8). Cited 1 Apr 2023. Available from: https://pubmed.ncbi.nlm.nih.gov/35666007/

  40. Ng CH, Lim GRS, Fong W. Quality of English-language videos on YouTube as a source of information on systemic lupus erythematosus. Int J Rheum Dis. 2020;23(12):1636–44.

    Article  PubMed  Google Scholar 

  41. Ovenden CD, Brooks FM. Anterior cervical discectomy and fusion YouTube Videos as a source of patient education. Asian Spine J. 2018;12:987–91.

    Article  PubMed  PubMed Central  Google Scholar 

  42. Mohile NV, Jenkins NW, Markowitz MI, Lee D, Donnally CJ. YouTube as an Information Source for Lumbar Disc Herniations: A Systematic Review. World Neurosurg. 2023;172:e250–5.

    Article  PubMed  Google Scholar 

  43. Rudisill SS, Saleh NZ, Hornung AL, Zbeidi S, Ali RM, Siyaji ZK, et al. YouTube as a source of information on pediatric scoliosis: a reliability and educational quality analysis. Spine Deform. 2023;11:3–9.

    Article  PubMed  Google Scholar 

  44. Richardson MA, Park W, Bernstein DN, Mesfin A. Analysis of the Quality, Reliability, and Educational Content of YouTube Videos Concerning Spine Tumors. Int J Spine Surg. 2022;16:278–82.

    Article  PubMed  PubMed Central  Google Scholar 

  45. Muller AL, Baker JF. Analysis of Lumbar Fusion and Lumbar Arthroplasty Videos on YouTube. Int J Spine Surg. 2022;16:283–90.

    Article  PubMed  PubMed Central  Google Scholar 

  46. Donaldson SI, Dormanesh A, Perez C, Zaffer MO, Majmundar A, Unger JB, et al. Monitoring the official Youtube channels of e-cigarette companies: a thematic analysis. Health Educ Behav. 2023;20:10901981221148964.

    Google Scholar 

  47. Bi X, Tang C. Research on the motives affecting the behavior of short video’s creators. Ieee Access. 2020;8:188415–28.

    Article  Google Scholar 

  48. McMahon KM, Schwartz J, Nilles-Melchert T, Ray K, Eaton V, Chakkalakal D. YouTube and the Achilles Tendon: An analysis of internet information reliability and content quality. Cureus. 2022;14(4):e23984.

    PubMed  PubMed Central  Google Scholar 

  49. Ozsoy-Unubol T, Alanbay-Yagci E. YouTube as a source of information on fibromyalgia. Int J Rheum Dis. 2021;24(2):197–202.

    Article  PubMed  Google Scholar 

  50. Crutchfield CR, Frank JS, Anderson MJ, Trofa DP, Lynch TS. A systematic assessment of YouTube content on femoroacetabular impingement: an updated review. Orthop J Sports Med. 2021;9(6):23259671211016340.

    Article  PubMed  PubMed Central  Google Scholar 

  51. Staunton PF, Baker JF, Green J, Devitt A. Online curves: a quality analysis of scoliosis videos on YouTube. Spine. 2015;40(23):1857–61.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

Not applicable

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

H.L. and X.Z. provided the concept. H.W., C.Y. and Z.L. acquired and analyzed data. H.W. and C.Y. and was major contributors in writing the manuscript. X.Z., T.W. and J.H. reviewed and corrected the article. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Liu Hao.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

The original version of this article was revised: the original publication of this article contained the redundant column “CSSS” in table 9.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, H., Yan, C., Wu, T. et al. YouTube online videos as a source for patient education of cervical spondylosis—a reliability and quality analysis. BMC Public Health 23, 1831 (2023). https://doi.org/10.1186/s12889-023-16495-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12889-023-16495-w

Keywords