Published on in Vol 8 , No 1 (2022) :Jan-Mar

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/29581, first published .
Assessment of the Quality, Understandability, and Reliability of YouTube Videos as a Source of Information on Basal Cell Carcinoma: Web-Based Analysis

Assessment of the Quality, Understandability, and Reliability of YouTube Videos as a Source of Information on Basal Cell Carcinoma: Web-Based Analysis

Assessment of the Quality, Understandability, and Reliability of YouTube Videos as a Source of Information on Basal Cell Carcinoma: Web-Based Analysis

Original Paper

1Department of Dermatology, Universitätsklinikum Erlangen, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany

2Comprehensive Cancer Center Erlangen - Europäische Metropolregion Nürnberg, Erlangen, Germany

3Department of Dermatology, University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden, Germany

Corresponding Author:

Theresa Steeb, MPH

Department of Dermatology

Universitätsklinikum Erlangen

Friedrich-Alexander-Universität Erlangen-Nürnberg

Ulmenweg 18

Erlangen, 91054

Germany

Phone: 49 9131 8545852

Email: Theresa.Steeb@uk-erlangen.de


Background: Patients with skin cancer increasingly watch online videos to acquire disease-related information. Until now, no scientific evaluation of the quality of videos available for German-speaking patients with basal cell carcinoma (BCC) has been performed.

Objective: In this study, we aimed to identify and evaluate videos about BCC provided on YouTube.

Methods: A video search on YouTube was conducted in July 2020, using German BCC-related keywords (eg, “Basalzellkarzinom,” “Basaliom,” “weißer hautkrebs,” and “heller hautkrebs”). The first three pages (ie, 60 videos) were searched by two independent researchers for each keyword. Two authors evaluated videos that met the predefined eligibility criteria. The quality of the information of the videos was evaluated using the DISCERN tool and the Global Quality Scale (GQS). The understandability and actionability were assessed with the Patient Education Materials Assessment Tool for Audiovisual Materials (PEMAT-A/V). The reliability was assessed with the JAMA (Journal of the American Medical Association) criteria score. Subgroup differences were identified using the Kruskal-Wallis test.

Results: A total of 41 videos were included in the evaluation. The mean assessment scores were as follows: DISCERN, 3.3 (SD 0.80); GQS, 3.8 (SD 1.1); JAMA, 27.74% (SD 22.1%); understandability, 70.8% (SD 13.3%); and actionability, 45.9% (SD 43.7%). These values indicated that the videos were of medium to good quality and had good understandability, low actionability, and poor reliability. The quality of videos provided by health professionals was significantly higher than that of videos provided by laypersons.

Conclusions: Optimization of health-related videos about BCC is desirable. In particular, adaptation to reliability criteria is necessary to support patient education and increase transparency.

JMIR Cancer 2022;8(1):e29581

doi:10.2196/29581

Keywords



Cutaneous basal cell carcinoma (BCC) represents the most common malignant tumor type in Central Europe, accounting for more than 80% of all epithelial skin carcinomas [1,2]. These tumors typically occur among fair-skinned individuals and are located most commonly on the head and neck, followed by the trunk and extremities [3]. The incidence of BCC continues to increase each year, with a current annual incidence of approximately 200 cases per 100,000 persons in Germany. However, the actual number is estimated to be much higher because cancer registries only document the first occurrence of BCC, and multiple tumors are not recorded [2,4]. BCC is rarely fatal, and surgical interventions remain to be the gold standard of treatment [1,5,6].

Patients with cancer in Germany commonly prefer to attend physician consultations in order to acquire disease-related information [7]. However, the physician’s time for a consultation is usually limited, while patients receive a large amount of medical and treatment-related information. Thus, patients may struggle with understanding all of the information provided and may subsequently feel inadequately informed [8]. While medical consultations and written information remain to be the most important sources of health information for patients, a steadily increasing number of patients are seeking health information on the internet [7,9-11]. YouTube is an open-access video-sharing platform, ranking second among the most-accessed websites worldwide, as it counts 5 billion visits per day and 1 billion hours watched daily [12]. It is increasingly used to disseminate health-related information and has become an easily accessible source for patients to acquire information related to their diseases [13]. The distribution of medical information to such a huge audience offers invaluable opportunities but also challenges, as the quality of unfiltered information posted can be of low scientific quality [14]. Information may even be misleading or harmful, as the credibility of the providers cannot be verified, and quality control of these videos has not yet been established [15-17]. Until now, no scientific evaluation of the quality of videos available for German-speaking patients with BCC has been performed. Therefore, the aim of this study was to identify YouTube videos about BCC and to assess their quality, reliability, usability, and understandability. The results of this study may encourage shared decision-making and be beneficial for both patients and health care providers in order to recommend appropriate videos to their patients.


Search Strategy

A video search on YouTube was conducted in July 2020, using German BCC-related keywords (eg, “Basalzellkarzinom,” “Basaliom,” “weißer hautkrebs,” and “heller hautkrebs”). The standard search options provided by YouTube were maintained. The first three pages (ie, 60 videos) were searched by two independent researchers for each keyword using Internet Explorer 11 (Microsoft). It has been observed that a significant proportion of users watch videos from only the first three pages. Furthermore, a similar methodology has been used in previous studies related to YouTube videos [18,19].

Eligibility Criteria

To be eligible for evaluation, videos had to meet the following inclusion criteria: (1) contain information referring to BCC, (2) be accessible for free and for all users, and (3) provide information in the German language. Videos were excluded if they were commercials, they did not have sound, they presented only photos, or if the duration was less than one minute. All search results were screened for duplicates, and the predefined eligibility criteria were applied.

Grouping of Videos

Due to the variety of the video providers, the videos were grouped according to their original source into the following categories: layperson, health professional (ie, hospital or practice), educational provider, noncommercial provider or professional society, pharmaceutical company, health portal, and unclassified. For television or news reports, we distinguished whether they were uploaded by the official channel or reuploaded by private providers.

Data Management

The available baseline information (ie, URL, title, name of the provider, video length, and year of upload) of each selected video was documented. Additionally, the numbers of views, likes, and dislikes were extracted. With this information, we calculated the video power index (VPI) to assess the popularity of the videos. The VPI is calculated as follows:

VPI = number of likes / (number of likes + number of dislikes) × 100

The baseline information was extracted to an internally piloted data extraction sheet using Microsoft Excel 2010.

Two reviewers (TS and MH) independently assessed the videos’ quality of information, reliability, and understandability. Prior to the assessment, the use of the assessment tools was piloted by independently evaluating the first five videos to discuss potential difficulties and resolve questions.

Quality of Information

The DISCERN tool is commonly used to assess the quality of cancer information and was developed for laypersons [20]. A modified German version of this tool was used in this study, consisting of nine items that were used (1) to review a video’s transparency (items 1-6), (2) to review a video’s content (items 7 and 8), and (3) to give an intuitive assessment summary (item 9). Items were scored on a 5-point scale ranging from 1 (“criterion is not met at all”) to 5 (“criterion is fully met”; Multimedia Appendix 1). Thus, videos that were rated, on average, 4 or higher were considered to be of good quality, those rated from 2 to below 4 were considered medium quality, and those rated less than 2 were considered low quality. A maximum of 45 points could be achieved.

Additionally, the Global Quality Scale (GQS) was used. The GQS includes a 5-point scale ranging from 1 (“low quality”) to 5 (“high quality”) [21]. Videos scoring 4 or 5 points were rated as high quality, those scoring 3 points were rated as medium quality, and those scoring 1 or 2 points were rated as low quality.

Understandability and Actionability

The Patient Education Materials Assessment Tool for Audiovisual Materials (PEMAT-A/V) was chosen to assess the individual videos’ understandability and actionability. The understandability section comprises 13 items that covered content, word choice and style, organization, layout and design, and the use of visual aids [22]. The second section covers actionability by four items. Each item can be scored as 0 (“disagree”), 1 (“agree”), or N/A (“not applicable”). Then, percentage scores for both sections are calculated by dividing the number of achieved points by the number of items the video was evaluated on in each section. PEMAT-A/V scores range from 0% to 100%, with higher values generally indicating better understandability or actionability.

Accuracy, Utility, and Reliability

The accuracy, utility, and reliability of each video source were explored according to the JAMA (Journal of the American Medical Association) benchmark criteria [23]. These four criteria included authorship (ie, authors, contributors, affiliations, and credentials), attribution (ie, references and sources used for the content and copyright information), disclosures (ie, sponsorship, advertising, commercial funding, and potential conflicts of interest), and currency (ie, dates of posted and updated information). Each item can be scored as 0 (“disagree”) or 1 (“agree”). Next, we calculated percentages of fulfilled items. The higher the value, the more accuracy, utility, and reliability elements were fulfilled.

Harms and Benefits

In order to summarize their potential benefit or harm, the videos were rated on an adapted 3-point scale as to whether they were perceived to be useful, neutral, or harmful for potential audiences [24]. Useful videos were judged to contain correct information and to be of value to patients, whereas harmful videos contained misleading or false information.

Statistical Analysis

Statistical analyses were conducted using SPSS Statistics for Windows (version 24; IBM Corp). Descriptive analyses included mean (SD) or median (range). Subgroup differences were explored using the Kruskal-Wallis test. The relationship between the individual items of the tests was examined using Spearman correlation. Statistical significance was set at P≤.05. The interrater agreement of the two reviewers was determined using the intraclass correlation coefficient, as well as by determining the interitem correlation, r, between the individual reviewers.


Video Identification and Baseline Characteristics

Our search identified 659 videos. Following a multistep process, three review authors (TS, MH, and LR) screened the videos for duplicates and checked them for compliance with the predefined eligibility criteria. Finally, 41 individual videos were considered for assessment (Figure 1). Most videos were provided by health professionals (15/41, 37%), followed by laypersons (6/41, 15%) and health portals (6/41, 15%). Furthermore, 10% of the videos (4/41) were offered by educational providers, and 7% (3/41) of the videos were TV reports uploaded by official TV channels or reuploaded by private providers. Out of 41 videos, 2 (5%) providers remained unclear.

Figure 1. Flowchart showing the identification process of the videos.
View this figure

The videos were uploaded between 2011 and 2020, with the majority (30/41, 73%) uploaded after 2017 (Table 1 and Multimedia Appendix 2). The number of views ranged from 25 to 386,195, with a mean of 27,853 views. The video length (minutes: seconds) ranged from 1:04 to 91:36. In 78% (32/41) of the videos, the duration was less than 10 minutes. The number of likes ranged from 0 to 17,925, with a median of 22. Most likes were given on a video dealing with the personal BCC history of a German influencer (video #20). The number of dislikes ranged from 0 to 333. The VPI was evaluable for 33 videos and ranged from 40 to 100.

Overall, video #8 (“Hautkrebs - Ein Überblick über Typen und Therapien”; Multimedia Appendix 2), provided by health professionals, and video #35 (“Weißer Hautkrebs – ein Patienteninformationsfilm”), created by a professional society, were rated best among all videos. Both videos gave an overview on the disease course. In contrast, video #30 (“Verjüngung mit Uta Baranovskyy: Weißer Hautkrebs Teil 3”) provided by a layperson was rated the worst due to misleading information regarding the treatment of BCC.

Table 1. Overview of baseline characteristics, quality, understandability, actionability, and reliability of the videos according to the respective categorization of the provider.
Characteristic Provider

AllLaypersonTV reupload by private accountHealth prof.aOfficial TV reportEduc.b providerProf. society or NCc providerHealth portalUnclear
Videos, n (%)41 (100)6 (15)3 (7)15 (37)3 (7)4 (10)2 (5)6 (15)2 (4.9)
Views, mean (SD)27,853 (70,693)101,427 (1,591,378)46,049 (77,209)4221 (3105)68,499 (57,295)2177 (2751)37,602 (25,232)6980 (6836)326 (350)
Video length (min:s), mean (range)11:12 (1:04- 91:36)15:07 (1:15- 37:44)6:31 (4:20- 9:12)8:25 (1:16- 65:50)12:41 (4:04- 29:13)5:30 (1:04- 16:52)25:51 (8:42- 43.01)2:48 (1:49- 3:57)46:55 (2:15- 91:36)
Year of upload, range2011- 20202017- 20192011- 20182012- 20202012- 20202016- 201920162013- 20202018- 2020
Likes, mean (range)672 (0-17,925)4383 (20-17,925)54 (0-139)19 (0-59)172 (53-344)14 (0-42)24.50 (0-49)26 (0-68)3 (2-3)
Dislikes, mean (range)15 (0-333)74 (0-333)12 (0-34)1 (0-3)28 (12-51)1 (0-3)5 (0-9)2 (0-5)0 (0)
DISCERN scored, mean (SD)3.30 (0.80)1.95 (0.41)3.50 (0.44)3.64 (0.53)3.08 (0.94)3.33 (0.41)3.72 (0.93)3.46 (0.70)4.03 (0.04)
GQS scoree, mean (SD)3.76 (1.13)1.83 (0.41)4.12 (0.29)4.47 (0.64)3.33 (1.26)3.50 (0.41)4.00 (0.71)3.83 (1.21)4.25 (1.06)
PEMAT-A/V scoref (% U), mean (SD)70.84 (13.32)60.28 (11.50)76.77 (10.16)75.32 (13.26)78.59 (11.33)71.44 (11.49)74.66 (18.16)61.55 (10.92)71.14 (19.61)
PEMAT-A/V scoref (% A), mean (SD)45.94 (43.74)55.56 (50.18)27.78 (25.46)56.67 (40.21)33.33 (33.33)25.00 (50.00)50.00 (70.71)36.11 (37.14)50.00 (70.71)
JAMA scoreg (%), mean (SD)27.74 (22.10)12.50 (15.81)20.83 (7.22)30.00 (16.23)37.50 (21.65)43.75 (38.86)62.50 (17.67)12.50 (15.81)31.25 (26.52)

aprof: professional.

beduc: educational.

cNC: noncommercial.

dDISCERN items were scored on a 5-point scale ranging from 1 (“criterion is not met at all”) to 5 (“criterion is fully met”); videos were considered good quality (≥4), medium quality (≥2 to <4), or low quality (<2).

eGQS: Global Quality Scale; the GSQ was scored a 5-point scale ranging from 1 (“low quality”) to 5 (“high quality”); videos were considered high quality (4 or 5), medium quality (3), or low quality (1 or 2).

fPEMAT-A/V: Patient Education Materials Assessment Tool for Audiovisual Materials; scores range from 0% to 100%, with higher values indicating better understandability (U) or actionability (A).

gJAMA: Journal of the American Medical Association; each of four criteria were scored as 0 (“disagree”) or 1 (“agree”); scores range from 0% to 100%, with higher values indicating higher reliability.

Quality: DISCERN and GQS Results

Out of 45 points in total, the 41 individual videos ranged between 10.5 and 35.0 points according to the DISCERN tool. The mean DISCERN scores per video ranged from 1.31 to 4.38 points, with an average mean score of 3.31 (SD 0.80) points, indicating medium quality (Table 1). Most score deductions were due to lacking information about the sources used to create the respective video or missing complementary information. The mean GQS score was 3.8 (SD 1.1) points, indicating medium quality as well.

Understandability and Actionability: PEMAT-A/V Results

The average PEMAT-A/V score was 70.84% (SD 13.32%, range 43.18%-100%) for understandability and 45.94% (SD 43.74%, range 0%-100%) for actionability. Most score deductions for the understandability domain were due to a lack of a summary and because no visual aids were deployed. For the actionability domain, information was often missing regarding the interpretation of certain figures in order to take action.

Accuracy, Utility, and Reliability: JAMA Results

In total, a mean of 27.74% (SD 22.1%, range 0%-87.5%) of the JAMA benchmark criteria were fulfilled, indicating rather poor reliability. The main reasons for score deductions were missing information regarding the currency of videos (ie, the upload date) and missing disclosure of the provider.

Harms and Benefits

A total of 49% (20/41) of the videos were evaluated as useful, 7% (3/41) were evaluated as harmful, and the remaining videos were evaluated as neither beneficial nor harmful. All videos estimated to be harmful were provided by laypersons.

Interrater Agreement

We calculated intraclass correlation coefficients ranging from 0.940 to 0.955 with a Cronbach α of .973, indicating high overall interrater agreement concerning the assessments by the DISCERN tool, the GQS, the JAMA criteria, and the PEMAT-A/V. The interitem correlation, r, was 0.949, indicating high individual agreement among the two reviewers when assessing the individual items.

Subgroup Analyses

Significant differences in video quality, according to the DISCERN tool and the GQS, were identified between videos provided by laypersons and health professionals (P=.01; ie, videos by health professionals were judged as having higher quality than those provided by laypersons).

Regarding the assessment of whether videos were beneficial or not, differences were found in terms of the quality of the videos. Videos rated as beneficial showed significantly better quality in comparison to those rated as harmful (DISCERN: P=.004; GQS: P=.002) and neutral (DISCERN: P=.006; GQS: P<.001), according to the DISCERN tool and the GQS. No further subgroup differences were identified.

Correlation Analysis

A significant positive correlation was found between DISCERN and GQS values (r=0.836) as well as between DISCERN values and reliability and understandability criteria (r=0.488 and r=0.460, respectively; Table 2). In addition, the quality according to the GQS also significantly correlated with the reliability (r=0.426) and understandability (r=0.482) of the videos. Furthermore, the longer the duration of a video, the more understandability (r=0.454) and actionability (r=0.314) items had been deployed. No further significant correlations between the baseline characteristics and the quality, reliability, understandability, or actionability of the videos were identified.

Table 2. Correlation analysis (Spearman r and two-tailed P value) among the research variables.
VariableBaseline characteristicsQualityReliabilityPEMAT-A/Va
LikesDislikesViewsDurationUploadsDISCERNGQSbJAMAcUnderstandabilityActionability
Baseline characteristics

Likes


r10.784d0.604d0.375e–0.060–0.085–0.082–0.0840.0310.215


P valuef<.001<.001.02.71.60.61.60.85.18

Dislikes


r0.784d10.619d0.228–0.323e–0.186–0.1510.037–0.1190.066


P value<.001<.001.15.04.25.35.82.46.68

Views


r0.604d0.619d10.200–0.380e0.0860.0030.0680.0290.055


P value<.001<.001.21.01.59.99.67.86.74

Duration


r0.375e0.2280.20010.226–0.013–0.0250.1150.454d0.314


P value.02.15.21.16.93.87.47<.001.06

Uploads


r–0.060–0.323e–0.380e0.22610.0590.106–0.1520.3000.157


P value.71.04.01.16.72.51.34.06.33
Quality measures

DISCERN


r–0.085–0.1860.086–0.0130.05910.836d0.488d0.460d0.135


P value.60.25.59.93.72<.001<.001<.001.40

GQS


r–0.082–0.1510.003–0.0250.1060.836d10.426d0.482d0.186


P value.61.35.99.87.51<.001<.001<.001.24
Reliability measures

JAMA


r–0.0840.0370.0680.115–0.1520.488d0.426d10.469d–0.052


P value.60.82.67.47.34<.001<.001<.001.75
PEMAT-A/V measures

Understandability


r0.031–0.1190.0290.454d0.3000.460d0.482d0.469d10.220


P value.85.46.86<.001.06<.001<.001<.001.17

Actionability


r0.2150.0660.0550.314e0.1570.1350.186–0.0520.2201


P value.18.68.74.05.33.40.24.75.17

aPEMAT-A/V: Patient Education Materials Assessment Tool for Audiovisual Materials.

bGQS: Global Quality Scale.

cJAMA: Journal of the American Medical Association.

dThe correlation is significant at a significance level of <.001 (two-tailed).

eThe correlation is significant at a significance level of .05 (two-tailed).

fNot applicable.


Principal Findings

In this study, 41 YouTube videos about BCC have been systematically identified and evaluated by two independent reviewers. For the first time, we present an in-depth and objective assessment of the quality, understandability, and reliability of the information about BCC provided by YouTube videos on this subject. There were more than 1 million views among the 41 videos identified in our search, highlighting the importance of the internet and platforms like YouTube as sources of health information. Half of the assessed videos were estimated to be beneficial for patients, showing that YouTube may be an important tool for information broadcasting. The percentage of beneficial videos was similar compared to the results of previous studies evaluating video contents about other diseases [25-27].

Our results complement the currently available evidence on informational material available for other types of skin cancer, such as videos, brochures, or websites [14,27,28]. Our evaluation shows that currently available BCC videos were, overall, of medium to good quality and understandability but had low actionability and poor reliability. In addition, we have shown that videos of longer duration applied more understandability and actionability items and that the quality of videos provided by health professionals was significantly higher than that of videos provided by laypersons.

Interestingly, none of the videos identified in our search were provided by pharmaceutical companies, which sharply contrasts with our previous search and evaluation of videos on melanoma [27]. In that study, 16% of the videos had been created by pharmaceutical companies and nearly one-third by laypersons, while most videos on BCC had been supplied by health professionals. A potential explanation might be that pharmaceutical companies offer more videos on melanoma, as the interest in disease-specific knowledge is judged to be more important due to the complexity and abundance of different therapy regimens. Nevertheless, our evaluation revealed that videos about BCC provided by health professionals scored the best ratings in terms of quality, understandability, and reliability. This may be explained by the fact that these providers have better resources and scientific backgrounds to produce such high-quality videos.

In summary, the quality, understandability, and reliability of the BCC videos were comparable to those about melanoma [27]. However, BCC videos were judged to score more points on actionability items and fewer points on reliability items. Notably, the most likes were awarded for the two videos uploaded by a female influencer describing her own personal history with BCC as well as her therapy and follow-up. While these videos were mostly inferior in comparison to other videos, they highlight that the involvement of testimonials or influencers might be a feasible approach to maximize the awareness of skin cancer, in general, and to promote preventive measures. However, on the other hand, they may also use their coverage to distribute incorrect or harmful information.

YouTube is a growing online video platform providing easy access [12] with steadily increasing popularity among patients and medical professionals [29]. Distribution of medical information to such a huge audience offers invaluable opportunities but also risks of misinformation and biased presentation. Since the accuracy of online information is variable and since there is no peer review of such videos, the credibility and trustworthiness of the providers cannot be verified [15-17]. Moreover, quality certificates, like HONcode (Health on the Net Foundation Code of Conduct), which are awarded for reliable health-related webpages, are missing for YouTube videos [30]. Additionally, YouTube can be used as an advertising tool. As users can share their personal opinions without sufficient information and experience, videos may mislead patients and affect the physician-patient relationship [31]. Obtaining correct information from reliable sources is crucial, as it increases patients’ satisfaction and empowerment and may improve treatment results [32,33]. Efforts should be undertaken to introduce regular quality control of videos with medical content on YouTube.

We are aware that this study has some limitations. YouTube search results are highly dynamic and will change when new videos are uploaded and when old videos are removed. Additionally, we did not include videos with restricted access (eg, asking for log-in information).

Conclusions

Overall, our study demonstrates that online videos on BCC are currently of medium to good quality and are predominantly uploaded by health professionals. However, the reliability of the videos was poor. As more and more patients use online material, including YouTube videos, for acquiring disease-specific knowledge, it is crucial to ensure good quality, understandability, and reliability prior to publication. Thus, optimization of the videos is desirable. In particular, adaptation to reliability criteria is necessary to support patient education and increase transparency. Patients should be advised to check the sources of the videos and whether their content is up to date.

Acknowledgments

This study was funded by the German Skin Cancer Council.

We acknowledge financial support by Deutsche Forschungsgemeinschaft and Friedrich-Alexander-Universität Erlangen-Nürnberg within the funding programme “Open Access Publication Funding”.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Overview of the different assessment tools used for the evaluation of the videos.

DOCX File , 18 KB

Multimedia Appendix 2

Summary of information about 41 videos on melanoma in German analyzed in this study.

DOCX File , 31 KB

  1. Lomas A, Leonardi-Bee J, Bath-Hextall F. A systematic review of worldwide incidence of nonmelanoma skin cancer. Br J Dermatol 2012 May;166(5):1069-1080. [CrossRef] [Medline]
  2. Asgari MM, Moffet HH, Ray GT, Quesenberry CP. Trends in basal cell carcinoma incidence and identification of high-risk subgroups, 1998-2012. JAMA Dermatol 2015 Sep;151(9):976-981. [CrossRef] [Medline]
  3. Venables ZC, Nijsten T, Wong KF, Autier P, Broggio J, Deas A, et al. Epidemiology of basal and cutaneous squamous cell carcinoma in the UK 2013-15: A cohort study. Br J Dermatol 2019 Sep;181(3):474-482 [FREE Full text] [CrossRef] [Medline]
  4. Erdmann F, Spix C, Katalinic A, Christ M, Folkerts J, Hansmann J, et al. Krebs in Deutschland für 2017/2018. Berlin, Germany: Robert Koch-Institut; 2021.   URL: https://edoc.rki.de/handle/176904/9042 [accessed 2022-02-27]
  5. Lang BM, Balermpas P, Bauer A, Blum A, Brölsch GF, Dirschka T, et al. S2k guidelines for cutaneous basal cell carcinoma - Part 1: Epidemiology, genetics and diagnosis. J Dtsch Dermatol Ges 2019 Jan;17(1):94-103. [CrossRef] [Medline]
  6. Lang BM, Balermpas P, Bauer A, Blum A, Brölsch GF, Dirschka T, et al. S2k guidelines for cutaneous basal cell carcinoma - Part 2: Treatment, prevention and follow-up. J Dtsch Dermatol Ges 2019 Feb;17(2):214-230. [CrossRef] [Medline]
  7. Heimer A, Henkel M. Bedarf an Krebsinformation in der Bevölkerung: Analyse des Informationsverhaltens von Ratsuchenden. Heidelberg, Germany: Akademische Verlagsgesellschaft AKA GmbH; 2012.   URL: https://www.krebsinformationsdienst.de/info/krebsinformationsbedarf-2012-bevoelkerung.pdf [accessed 2022-02-27]
  8. Brütting J, Bergmann M, Garzarolli M, Rauschenberg R, Weber C, Berking C, NVKH supporting group. Unmet information needs of patients with melanoma in Germany. Melanoma Res 2019 Apr;29(2):196-204. [CrossRef] [Medline]
  9. Hamilton SN, Scali EP, Yu I, Gusnowski E, Ingledew P. Sifting through it all: Characterizing melanoma patients' utilization of the internet as an information source. J Cancer Educ 2015 Sep;30(3):580-584. [CrossRef] [Medline]
  10. van de Poll-Franse LV, van Eenbergen MCHJ. Internet use by cancer survivors: Current use and future wishes. Support Care Cancer 2008 Oct;16(10):1189-1195. [CrossRef] [Medline]
  11. Ebel M, Stellamanns J, Keinki C, Rudolph I, Huebner J. Cancer patients and the internet: A survey among German cancer patients. J Cancer Educ 2015 Nov 9:503-508. [CrossRef] [Medline]
  12. Ceci L. Statista. 2022 Feb 07.   URL: https://www.statista.com/topics/2019/youtube/#topicHeader__wrapper [accessed 2022-02-28]
  13. Madathil KC, Rivera-Rodriguez AJ, Greenstein JS, Gramopadhye AK. Healthcare information on YouTube: A systematic review. Health Informatics J 2015 Sep;21(3):173-194. [CrossRef] [Medline]
  14. Brütting J, Steeb T, Reinhardt L, Berking C, Meier F. Exploring the most visible German websites on melanoma immunotherapy: A web-based analysis. JMIR Cancer 2018 Dec 13;4(2):e10676 [FREE Full text] [CrossRef] [Medline]
  15. Stamelou M, Edwards MJ, Espay AJ, Fung VSC, Hallett M, Lang AE, et al. Movement disorders on YouTube--Caveat spectator. N Engl J Med 2011 Sep 22;365(12):1160-1161 [FREE Full text] [CrossRef] [Medline]
  16. Steinberg PL, Wason S, Stern JM, Deters L, Kowal B, Seigne J. YouTube as source of prostate cancer information. Urology 2010 Mar;75(3):619-622. [CrossRef] [Medline]
  17. Cline RJ, Haynes KM. Consumer health information seeking on the internet: The state of the art. Health Educ Res 2001 Dec;16(6):671-692. [CrossRef] [Medline]
  18. Jansen BJ, Spink A. An analysis of web documents retrieved and viewed. In: Proceedings of the 4th International Conference on Internet Computing. 2003 Presented at: The 4th International Conference on Internet Computing; June 23-26, 2003; Las Vegas, NV p. 65-69   URL: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.108.7248&rep=rep1&type=pdf
  19. Rittberg R, Dissanayake T, Katz SJ. A qualitative analysis of methotrexate self-injection education videos on YouTube. Clin Rheumatol 2016 May;35(5):1329-1333. [CrossRef] [Medline]
  20. Discern.: Abteilung Epidemiologie, Sozialmedizin und Gesundheitssystemforschung, Medizinische Hochschule Hannover / Ärztliche Zentralstelle Qualitätssicherung; 2005.   URL: http://www.discern.de/ [accessed 2022-02-18]
  21. Bernard A, Langille M, Hughes S, Rose C, Leddin D, Veldhuyzen van Zanten S. A systematic review of patient inflammatory bowel disease information resources on the World Wide Web. Am J Gastroenterol 2007 Sep;102(9):2070-2077. [CrossRef] [Medline]
  22. Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): A new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns 2014 Sep;96(3):395-403. [CrossRef] [Medline]
  23. Silberg WM, Lundberg GD, Musacchio RA. Assessing, controlling, and assuring the quality of medical information on the internet: Caveant lector et viewor--Let the reader and viewer beware. JAMA 1997 Apr 16;277(15):1244-1245. [Medline]
  24. Li M, Yan S, Yang D, Li B, Cui W. YouTube™ as a source of information on food poisoning. BMC Public Health 2019 Jul 16;19(1):952 [FREE Full text] [CrossRef] [Medline]
  25. Delli K, Livas C, Vissink A, Spijkervet FKL. Is YouTube useful as a source of information for Sjögren's syndrome? Oral Dis 2016 Apr;22(3):196-201. [CrossRef] [Medline]
  26. Garg N, Venkatraman A, Pandey A, Kumar N. YouTube as a source of information on dialysis: A content analysis. Nephrology (Carlton) 2015 May;20(5):315-320. [CrossRef] [Medline]
  27. Steeb T, Reinhardt L, Görgmayr C, Weingarten H, Doppler A, Brütting J, German Skin Cancer Council. German YouTube™ videos as a source of information on cutaneous melanoma: A critical appraisal. J Eur Acad Dermatol Venereol 2020 Oct;34(10):e642-e644. [CrossRef] [Medline]
  28. Brütting J, Reinhardt L, Bergmann M, Schadendorf D, Weber C, Tilgen W, on behalf of the NVKH. Quality, readability, and understandability of German booklets addressing melanoma patients. J Cancer Educ 2019 Aug;34(4):760-767. [CrossRef] [Medline]
  29. Desai T, Shariff A, Dhingra V, Minhas D, Eure M, Kats M. Is content really king? An objective analysis of the public's response to medical videos on YouTube. PLoS One 2013;8(12):e82469 [FREE Full text] [CrossRef] [Medline]
  30. HONcode. Health On the Net.   URL: https://www.hon.ch/HONcode/Patients/Visitor/visitor.html [accessed 2022-02-18]
  31. Nason GJ, Baker JF, Byrne DP, Noel J, Moore D, Kiely PJ. Scoliosis-specific information on the internet: Has the "information highway" led to better information provision? Spine (Phila Pa 1976) 2012 Oct 01;37(21):E1364-E1369. [CrossRef] [Medline]
  32. Sechrest RC. The internet and the physician-patient relationship. Clin Orthop Relat Res 2010 Oct;468(10):2566-2571 [FREE Full text] [CrossRef] [Medline]
  33. Hungerford DS. Internet access produces misinformed patients: Managing the confusion. Orthopedics 2009 Sep;32(9):658-660 [FREE Full text] [CrossRef] [Medline]


BCC: basal cell carcinoma
GQS: Global Quality Scale
HONcode: Health on the Net Foundation Code of Conduct
JAMA: Journal of the American Medical Association
N/A: not applicable
PEMAT-A/V: Patient Education Materials Assessment Tool for Audiovisual Materials
VPI: video power index


Edited by A Mavragani; submitted 13.04.21; peer-reviewed by A Wessely, A Hidki; comments to author 28.06.21; revised version received 07.07.21; accepted 20.12.21; published 11.03.22

Copyright

©Theresa Steeb, Lydia Reinhardt, Matthias Harlaß, Markus Vincent Heppt, Friedegund Meier, Carola Berking. Originally published in JMIR Cancer (https://cancer.jmir.org), 11.03.2022.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Cancer, is properly cited. The complete bibliographic information, a link to the original publication on https://cancer.jmir.org/, as well as this copyright and license information must be included.