Accessibility settings

Published on in Vol 12 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/76044, first published .
Extracting Quality of Life Information of Patients Diagnosed With Breast Cancer From Health Care Online Forum Posts: Data Feasibility Study

Extracting Quality of Life Information of Patients Diagnosed With Breast Cancer From Health Care Online Forum Posts: Data Feasibility Study

Extracting Quality of Life Information of Patients Diagnosed With Breast Cancer From Health Care Online Forum Posts: Data Feasibility Study

1Center for Cognitive Interaction Technology, Faculty of Technology, Bielefeld University, Inspiration 1, Bielefeld, Germany

2Inspire, Arlington, VA, United States

3Department of Medical Psychology and Medical Sociology, University of Leipzig, Leipzig, Germany

4Semalytix GmbH, Bielefeld, Germany

Corresponding Author:

David Maria Schmidt, MSc


Background: Quality of life (QoL) questionnaires are used in many disease areas to measure the burden that a disease causes for patients, which help provide insights into disease impact, identify unmet medical needs, and inform patient-centered drug development and value assessment for treatments. The collection of data imposes both a significant burden on patients as well as effort on health care personnel, thus incurring high costs for the health care system. Given that patients share detailed information about their condition and treatment experiences on social media and patient forums, an important research question is to what extent information about QoL can be obtained from patients’ online forum posts to potentially complement information obtained from questionnaires.

Objective: This study aimed to assess how much QoL information can be gained from the analysis of posts by patients in online health care communities and whether this information is rich enough to estimate individual patient’s QoL based on their posts. We conducted this feasibility study in the context of breast cancer as it is the most prevalent cancer in the female population.

Methods: We recruited 134 female patients diagnosed with breast cancer on the Inspire patient online forum, who voluntarily participated in our feasibility study. They filled in the EORTC (European Organisation for Research and Treatment of Cancer) QLQ-C30 and QLQ-BR23 questionnaires consisting of 30 general questions and 23 additional breast cancer–specific questions and provided consent to analyze their posts and comments on the online forum (756 posts and 19,478 comments). Posts were coded manually to identify parts of the text providing answers to 1 of the aforementioned 53 questions.

Results: The data annotation yielded a substantial agreement (mean Fleiss κ of 0.5, SD 0.28). Overall, we found answers in the coded data for 50 out of 53 EORTC QLQ-C30 and QLQ-BR23 questions. The information coded in the posts reliably predicted the answers given in the questionnaires (F1-score=0.7), with even better results when grouping similar questions (F1-score=0.8 for fine-grained and 0.9 for coarse-grained grouping). The 5 questions that were most frequently answered on the basis of the coded posts were “Did you feel ill or unwell?” (304 of 2683 annotated posts and comments), “Did you worry?” (105 posts and comments), “Have you had pain?” (104 posts and comments), “Did you feel tense?” (85 posts and comments), and “Were you limited in doing either your work or other daily activities?” (77 posts and comments).

Conclusions: Our feasibility study shows that there is valuable QoL-related information in posts of online patient communities, which can potentially serve as an innovative low-burden QoL monitoring approach. Future research should consider how these insights can be used to complement existing QoL instruments and whether the process of extracting QoL-related information can be automated.

JMIR Cancer 2026;12:e76044

doi:10.2196/76044

Keywords



Quality of life (QoL), as defined by the World Health Organization, encompasses physical, mental, and social well-being beyond the mere absence of disease [1]. Assessing QoL is essential in health care, providing insights into disease impact, identifying unmet medical needs, and informing patient-centered drug development and value assessment for treatments. Traditionally, QoL is measured using patient-reported outcome questionnaires, such as the widely used EORTC (European Organisation for Research and Treatment of Cancer) QLQ-C30 for patients with cancer [2]. These standardized tools facilitate comparisons across patient groups and with the general population.

However, QoL assessments are often time-consuming and administratively burdensome for both patients and health care providers [3-7]. Efforts to streamline data collection have explored digital and computer-supported approaches [6-8]. Given the increasing use of social media and online patient communities for sharing health experiences, a key question arises: Can patient-generated online content serve as a proxy or complementary source for QoL data collection? While not a replacement for structured surveys, social media data may provide real-time, cost-effective insights into patient well-being, preferences, and treatment experiences.

Previous research has leveraged social media to assess disease burden and treatment experiences across conditions, such as Parkinson disease [9,10], Sjögren disease [11], chronic obstructive pulmonary disease [12], and mental disorders [13,14]. Some related work also analyzed patient posts compared to a specific QoL ground truth, but on an aggregated level. For example, Zivanovic et al [15] compared Twitter data with official QoL data of the city of Bristol. Similarly, Cotté et al [16] also investigated cancer-specific forums and QoL data, but with a focus on immune checkpoint inhibitors and not the QoL impact for individual users. Therefore, studies have yet to systematically evaluate whether individual QoL data can be reliably extracted from online patient discussions.

To explore this, we conducted a study with patients diagnosed with breast cancer from the Inspire patient online forum, assessing whether online forum posts contain relevant QoL information. Breast cancer remains the most commonly diagnosed cancer in women, with over 2.3 million new cases in 2020 [17]. QoL is a critical concern for patients diagnosed with breast cancer and survivors, who frequently experience significant impairments [18-21].

In our study, patients diagnosed with breast cancer completed standard QoL questionnaires (EORTC QLQ-C30 and QLQ-BR23), while their online posts and comments were manually coded to determine the presence of QoL-related content. We then assessed the prevalence of QoL discussions and their correlation with questionnaire responses. Our research focuses on the following questions:

  • RQ1: Do patient online forum posts contain QoL-related information?
  • RQ2: Which QoL questions from EORTC QLQ-C30 and QLQ-BR23 are most frequently addressed in online posts?
  • RQ3: How well do these online discussions align with individual patients’ QoL questionnaire responses?

This study aims to determine whether patient-generated content from online health communities can complement traditional QoL assessments, potentially offering a scalable, low-burden alternative for understanding patient well-being.


Overall Structure

This feasibility study aimed to assess the presence of QoL information in online discussions by patients diagnosed with breast cancer. Breast cancer was selected due to its high prevalence, and data were collected from patient-generated content on the Inspire patient online forum. Patient communities on this forum are organized by condition, including “breast cancer” and “advanced breast cancer,” where each thread of natural peer-to-peer conversation consists of an initial post and subsequent comments. An overview of the workflow followed in this study is given in Figure 1.

To compare patient-reported QoL in online posts with structured assessments, we (the authors of this study) administered the QLQ-C30 questionnaire by the EORTC [2] and its breast cancer–specific module, QLQ-BR23. These validated instruments provided a reference (ground truth) against which the content of online discussions could be analyzed.

Participants were recruited from the “Breast Cancer” and “Advanced Breast Cancer” communities on Inspire. Eligibility criteria included female patients diagnosed with breast cancer, age 18 years or older, residency in the United States, and at least 1 post or comment in the respective Inspire communities.

Residency in the United States was included due to Inspire being based in the United States.

The EORTC QLQ-C30 questionnaire was used to assess general QoL in patients with cancer. It comprises 30 items grouped into 5 functioning scales, a global health status/QoL scale, 3 symptom scales, and 6 single-item measures. Scores range from 0 to 100, with higher functioning scores and lower symptom scores indicating better QoL. Normative values for this instrument have been previously established [22,23]. The breast cancer–specific module, EORTC QLQ-BR23, consists of 23 items categorized into 8 scales: systemic therapy side effects, hair loss, arm symptoms, breast symptoms, body image, future perspective, sexual functioning, and sexual enjoyment. Response options align with those of the EORTC QLQ-C30.

The final survey also included demographic questions, such as year of birth, US state of residence, gender, sex at birth, insurance type, and residential setting (urban, suburban, rural, or other). The survey required approximately 5 to 10 minutes to complete and was conducted between May and June 2024.

Figure 1. Overview of the study workflow from planning survey conduction to data analysis. The online survey using the EORTC QLQ-C30 and QLQ-BR23 questionnaires was conducted between May and June 2024 with patients diagnosed with breast cancer posting in the Inspire patient online forums. EORTC: European Organisation for Research and Treatment of Cancer.

Ethical Considerations

Our study has been approved by the Ethics Committee of Bielefeld University under application number 2023‐216-W1. A filled CHERRIES (Checklist for Reporting Results of Internet E-Surveys) [24] checklist can be found in Checklist 1. Informed consent was obtained from all participants, granting permission to analyze their posts and comments for this study. The collected data, especially the survey responses, are not publicly shared and only available to the authors of the study. Additionally, the data are stored in encrypted form only. All eligible community members were invited to participate (ie, convenience sampling), with the first 100 respondents receiving a US $20 gift card as compensation.

Annotation and Coding Process

The objective of the coding process was to identify text spans within collected posts and comments that could correspond to responses in the EORTC QoL questionnaires. To achieve this, we implemented an annotation schema similar to named-entity recognition, where each of the 53 QoL questions served as a label. Unlike traditional named-entity recognition, which identifies entities, such as persons or organizations [25,26], our approach used QoL-related labels to tag relevant text segments.

In addition to assigning labels corresponding to each QoL question, we introduced supplementary labels to capture both reported presence and absence of symptoms: “positive” or “negative.” Coders applied these labels alongside question-specific tags to indicate whether a response reflected a favorable or unfavorable QoL experience. Furthermore, an “Exact match” label was used to denote text spans that provided precise or highly relevant answers, while standard labels without this addition were applied to responses that were partial or vague.

The annotation schema and guidelines were developed in a 2-phase process. In the first phase, 3 annotators labeled 30 posts using a preliminary schema. After completing this pilot phase, the results were reviewed, and minor refinements were made. For instance, the label “Did you feel ill or unwell?” was designated for general QoL descriptions that did not fit into other, more specific categories. The annotators were explicitly trained to only use the fallback label when they could not find any other label that fits better, and this was also added to the annotation guidelines (Multimedia Appendix 1). Compliance with these rules was ensured by closely checking fallback annotations during the beginning of the annotation process of the full data. These pilot annotations were subsequently excluded from the final dataset.

In the second phase, the full annotation process commenced, with no further modifications needed. Additionally, posts and comments from the last 6 months (relative to each user’s participation date) were labeled to indicate whether they referred to past events. This distinction was critical for evaluating the correlation between forum data and survey responses, as older event descriptions may not accurately reflect a patient’s current QoL. Filtering out such posts allowed for a more precise alignment between forum-derived insights and survey-based QoL assessments.

Measuring Information Consistency Between Online Posts and Survey Responses

To assess the alignment between information shared in online posts and responses in the structured questionnaires, we applied classification metrics commonly used in machine learning. The coding scheme and questionnaire responses were simplified into a binary format, mapping each piece of information from posts and survey answers to either a “positive” or “negative” response. The intended meaning of “positive” here is that the corresponding question would be answered with “yes,” thus indicating the presence of symptoms or other issues. Analogously, “negative” corresponds to answering the question with “no,” indicating the absence of symptoms or other issues. Specifically, the answer categories “a little,” “quite a bit,” and “very much” were classified as “positive,” whereas “not at all” was categorized as “negative.” This enabled the computation of key classification metrics as follows:

  • True positives (TP): Instances where both the online post and the corresponding questionnaire response indicate a “positive” answer (eg, a user stating, “I had so much pain last week!” and selecting “yes” for the survey question, “Have you had pain?”).
  • False positives (FP): Cases where an online post is labeled as “positive,” but the corresponding survey response is “negative” (eg, a user stating, “I am so worried!” but responding “no” to the question, “Did you worry?”).
  • True negatives (TN): Instances where both the online post and the survey response indicate a “negative” answer (eg, a user stating, “I do not have any pain!” and selecting “no” for the survey question, “Have you had pain?”).
  • False negatives (FN): Cases where an online post is labeled as “negative,” but the corresponding survey response is “positive” (eg, a user stating, “I did not worry at all last week!” but selecting a “positive” response in the survey question, “Did you worry?”).

If no labeled data existed for a specific user and QoL question, that user did not contribute to the TP, FP, TN, or FN calculations for that question. This approach allowed us to evaluate how well the available data could predict survey responses.

Using the TP, FP, TN, and FN values, we computed standard machine-learning classification metrics, including precision (TP/[TP + FP]), recall (TP/[TP + FN]), and F1-score as harmonic mean between precision and recall, that is, 2× precision × recall/(precision + recall) [27]. Given that data sparsity increases when analyzing results at the individual user level across all 53 survey questions, we explored grouping similar survey questions based on the EORTC scoring manual. This resulted in 2 levels of categorization: (1) fine-grained grouping: based on detailed QoL question clusters (Table 1) and (2) coarse-grained grouping: functional (questions 1-7, 20-27, 39-46), symptoms (questions 8-19, 28, 31-38, 47-53), and global health (questions 29 and 30).

These groupings enabled more robust analyses by reducing sparsity and improving interpretability of the correspondence between online forum discussions and structured survey responses.

Table 1. Fine-grained grouping based on the EORTCa QLQ-C30 and QLQ-BR23 scoring manualb.
GroupQuestion IDs
Global health29, 30
Physical functioning1, 2, 3, 4, 5
Role functioning6, 7
Emotional functioning21, 22, 23, 24
Cognitive functioning20, 25
Social functioning26, 27
Fatigue10, 12, 18
Nausea and vomiting14, 15
Pain9, 19
Dyspnea8
Insomnia11
Appetite loss13
Constipation16
Diarrhea17
Financial difficulties28
Systemic therapy side effects31, 32, 33, 34, 36, 37, 38
Upset by hair loss35
Arm symptoms47, 48, 49
Breast symptoms50, 51, 52, 53
Body image39, 40, 41, 42
Future perspective43
Sexual functioning44, 45
Sexual enjoyment46

aEORTC: European Organisation for Research and Treatment of Cancer.

bQuestion IDs refer to the corresponding EORTC questionnaire question numbers. The online survey using the EORTC QLQ-C30 and QLQ-BR23 questionnaires was conducted between May and June 2024 with patients diagnosed with breast cancer posting in the Inspire patient online forums.


General Statistics

The collected dataset comprises 20,204 posts and comments from 2006 to 2024, with 580 posts and comments originating within the 6 months preceding the end of the study. The distribution of posts per user follows a long-tail pattern, as illustrated in Figure 2, with absolute values presented in Table 2). As shown in Figure 2, the majority of contributions are comments rather than original posts. Table 3 shows that our data cover a broad spectrum of patients diagnosed with breast cancer, both in terms of age and distance between survey and breast cancer diagnosis.

Figure 2. Posts per user, with 123 users on the x-axis and the (log-scaled) number of posts on the y-axis, differentiating between combined posts and comments in light blue and comments only in dark blue. Users without posts or comments were disregarded. The online survey using the EORTC QLQ-C30 and QLQ-BR23 questionnaires was conducted between May and June 2024 with patients diagnosed with breast cancer posting in the Inspire patient online forums. EORTC: European Organisation for Research and Treatment of Cancer.
Table 2. Overview of the number of posts per usera.
Number of posts and commentsNumber of users
16
2-1028
11-10063
101-100022
>10004

aThe online survey using the EORTC QLQ-C30 and QLQ-BR23 questionnaires was conducted between May and June 2024 with breast cancer patients posting in the Inspire patient online forums.

The average post length is 408.4 (SD 366.8) characters. A total of 134 participants were enrolled in the study, of whom 11 did not contribute at least 1 post or comment, resulting in 123 participants meeting the inclusion criteria.

For annotation, the last 6 months of data (580 posts and comments) were labeled by 3 annotators, with an average annotation time of 15.9 (SD 10.3) hours per annotator. The last 24 months of data (an additional 2103 posts and comments) were labeled by 2 annotators, requiring an average of 76.5 (SD 12.7) hours per annotator. In total, 2683 posts were annotated. Fourteen posts contained QoL-related information referring to someone other than the posting user. These posts were annotated but excluded from statistical evaluation.

Of the 2669 annotated posts and comments included in the final dataset, 613 (23%) contained at least 1 label corresponding to a QoL question. When considering only the last 6 months, the proportion was higher at 37.4% (217 out of 580 posts), indicating that approximately one-third to one-fourth of all posts contained QoL-relevant content.

Table 3. Statistics of self-reported demographic data of survey participantsa.
Data categoryMean (SD)MinimumMaximum
Age68.2 (11.5)27116
Age at diagnosis55.7 (12.6)26103
Years since diagnosis13.1 (9.4)045

aThe online survey using the EORTC QLQ-C30 and QLQ-BR23 questionnaires was conducted between May and June 2024 with patients diagnosed with breast cancer patients posting in the Inspire patient online forums.

Among the 53 QoL questions, 50 were labeled at least once, with the most frequently occurring questions summarized in Table 4. The most commonly assigned label was “Did you feel ill or unwell?,” with 301 annotations, representing 11% (n=301) of all posts and 57.3% (n=47) of users who actively posted during the study period. It has to be noted that this is used as a fallback label for the case where posts or comments contain information which is clearly QoL-relevant, but no other better-fitting label could be found. An example for this is when patients describe some very specific or very vague symptoms that do not correspond well to any other question (eg, “the symptoms increased a lot”). The next most frequently annotated questions were “Did you worry?” (105 annotations) and “Have you had pain?” (102 annotations), both capturing broad categories of patient discomfort. The overall label distribution is visualized in Multimedia Appendix 2.

Table 4. Top 10 most frequent labels with respect to number of posts they are annotated in for all annotated posts (most recent 24 months)a.
QuestionNumber of postsPortion of posts, %Number of usersPortion of users, %
Did you feel ill or unwell?301114757.3
Did you worry?1053.92631.7
Have you had pain?1023.73239
Did you feel tense?843.12125.6
Were you limited in doing either your work or other daily activities?772.81822
Were you tired?5521822
Have you felt nauseated?511.91822
Were you worried about your health in the future?411.51619.5
Have you lacked appetite?321.256.1
Did you feel depressed?321.21417.1

a“Portion of posts” refers to the percentage of posts with the corresponding label relative to all posts of that time period. Similarly, “Portion of users” refers to the number of users who have posts annotated with that label relative to all users who posted in that time period. The online survey using the European Organisation for Research and Treatment of Cancer (EORTC) quality of life questionnaires (QLQ) QLQ-C30 and QLQ-BR23 was conducted between May and June 2024 with patients diagnosed with breast cancer posting in the Inspire patient online forums.

Annotator Agreement

As every post and comment within the most recent 2 years was annotated by 2 or 3 annotators, the interannotator agreement can be calculated via Fleiss κ values [28]. This measure aids in determining to which degree annotators agreed on which text segments or posts answered that question at least partially. We only consider the basic labels corresponding to the questions for the calculation of interannotator agreement, ignoring the additional labels “negative” and “exact match,” and handle each question separately, as this fits best with the setting assumed by the Fleiss κ metric for interannotator agreement.

We calculated the Fleiss κ for 2 different levels of annotation granularity, that is, on word and on post or comment level, respectively. The word level reflects the actual annotations, that is, the text segments created during annotation, whereas the post level is a simplification aggregating all annotations of a post such that the annotations only have to agree on whether the whole text answers a specific question or not, but not where exactly.

The top and bottom 10 Fleiss κ values for the most recent 6 months of data are provided in Table 5, together with the absolute numbers of posts in which the corresponding question has been answered.

It can be observed that many frequently occurring labels exhibit solid to high annotator agreement, for example, “Have you lost any hair?” has a Fleiss κ value of 0.7 for word units and even 1.0 for post units while being present in 26 posts in total. Even the fallback label “Did you feel ill or unwell?” with 285 labeled posts has Fleiss κ values of 0.6 and 0.7 for word and post granularity, respectively.

Most questions in the lower range of Fleiss κ values have a very low number of labeled posts. The 6 questions with the lowest observed agreement of 0.0 together only account for 14 posts in total, so these appear to be cases where the annotators do not agree whether the label is applicable. Further investigating the data, the most disagreement is caused by vaguely formulated posts for which a label could but does not necessarily have to apply. For example, for labels, such as “Did you have any pain in your arm or shoulder?,” issues can range from just vaguely described body areas to mentions of certain diseases or syndromes, which are closely connected to pain in that area, without pain being explicitly mentioned. Although we cannot give real posts as examples for data privacy reasons, we summarized prototypical posts to illustrate the reasons for disagreement. For “Did you have any pain in your arm or shoulder?,” annotators disagreed for posts, such as “I take that medication for 5 years now and started to have problems with hand-foot syndrome some time ago,” as hand-foot syndrome itself is not an explicit mention of pain and can occur with or without pain [29]. Similarly, annotators disagreed for posts, such as “In the last months I had cramps in different parts of my hand from time to time.” Another label with low annotator agreement was “Have you had skin problems on or in the area of your affected breast (e.g., itchy, dry, flaky)?,” for example, for posts like “I try to avoid foods with high folic acid as much as possible as they can cause pain in the arms for some of us. For the dryness, I use some ointment all the time,” as it is not very clear what part of the body “dryness” refers to.

Table 5. Annotator agreement for the 10 questions with the highest agreement and 10 questions with the lowest agreement (Fleiss κ, the most recent 6 months of data before the survey) using word units and post unitsa.
QuestionFleiss κ (word units)Fleiss κ (post units)Labeled posts
Top 10
Did you have headaches?1.01.03
Have you had difficulty remembering things?0.91.03
Have you lacked appetite?0.90.915
Did you have hot flushes?0.80.89
Have you lost any hair?0.71.026
To what extent were you sexually active? (with or without intercourse)0.71.06
Did you feel ill or unwell?0.60.7285
Have you had diarrhea?0.60.910
Do you have any trouble taking a short walk outside of the house?0.60.719
Were you tired?0.60.835
Bottom 10
Have you had skin problems on or in the area of your affected breast (eg, itchy, dry, and flaky)?0.20.311
Have you vomited?0.20.52
Do you have any trouble taking a long walk?0.20.37
Did you have any pain in your arm or shoulder?0.20.411
How would you rate your overall health during the past week?0.00.01
Did you need to rest?0.00.01
Was it difficult to raise your arm or to move it sideways?0.00.01
Did pain interfere with your daily activities?0.00.03
Did food and drink taste different than usual?0.00.06
How would you rate your overall quality of life during the past week?0.00.02

aThe online survey using the European Organisation for Research and Treatment of Cancer (EORTC) quality of life questionnaires (QLQ) QLQ-C30 and QLQ-BR23 was conducted between May and June 2024 with patients diagnosed with breast cancer posting in the Inspire patient online forums.

Correlation Results Between Annotated Data and Survey Data

We estimate the correlation between the information provided by the authors to QoL aspects and the data provided in the questionnaires by computing F1-score as described in the “Measuring Information Consistency Between Online Posts and Survey Responses” section. Table 6 shows the F1-scores and total annotations for the 10 most frequently annotated questions based on data originating from the most recent 6-month period and compares these numbers to the corresponding values for 24 months.

As observed from Table 6, all of the 10 most frequently answered questions can be predicted reliably with an F1-score of 0.6 or higher for the last 6 months of data. For 7 of the 10 questions, the F1-score is even 0.8 or higher. The lowest score of 0.6 is achieved for the most frequent question, that is, “Did you feel ill or unwell?” This specific question is also used as a fallback label for utterances, which do not fit in any other category, thus increasing prediction difficulty compared to other labels. Even when considering a much larger time period of 24 months, the F1-scores for these questions do not drop below 0.5, with some even increasing.

Although the observed F1-scores are high for many questions, for example, “Were you tired?” or “Were you worried about your health in the future?” with each of an F1-score of 1.0, for some questions, the available data get sparse and the data sparsity increases even further when considering single users or shorter time spans, leading to lower F1-scores in some cases, for example, for “Have you vomited?” (F1-score of 0.2 with 2 labeled posts in 6 months of data). One way of reducing data sparsity is to consider a larger time span. However, this comes at the cost of posts and comments with a larger temporal distance from the survey date being potentially less relevant with respect to the QoL at that time. However, in Table 6, despite the inclusion of older posts, we can observe solid to high F1-scores (0.6-1.0, mean 0.8, SD 0.1) for all 10 most frequently answered questions. This shows that including posts from dates outside the time period asked for in the QoL surveys can still be useful.

Table 6. F1-scores for the 10 most frequently annotated questions, describing how well annotations predict the actual answers given in the survey, considering the last 6 and 24 months of data, which are not filtered by time deviationa.
QuestionF1-score (6 mo)Total annotations (6 mo)F1-score (24 mo)Total annotations (24 mo)
Did you feel ill or unwell?0.61270.7222
Did you worry?0.8360.947
Did you feel tense?0.9330.840
Have you had pain?0.8280.874
Were you tired?1.0140.950
Have you felt nauseated?0.7140.534
Has your physical condition or medical treatment interfered with your social activities?0.7130.713
Has your physical condition or medical treatment caused you financial difficulties?1.0121.012
Were you worried about your health in the future?1.0111.015
Have you lost any hair?0.8100.829

aThe online survey using the European Organisation for Research and Treatment of Cancer (EORTC) quality of life questionnaires (QLQ) QLQ-C30 and QLQ-BR23 was conducted between May and June 2024 with patients diagnosed with breast cancer posting in the Inspire patient online forums.

As some questions occur sparsely, we group the answers from the posts and the questionnaires into the groups “Symptomatic” and “Functional” used in the EORTC scoring manual. The results for the grouped questions are shown in Table 7.

The only group in Table 7 having annotations available and exhibiting a lower F1-score than 0.7 for the last 6 months is “cognitive functioning,” which comprises only 2 questions, that is “Have you had difficulty in concentrating on things, like reading a newspaper or watching television?” and “Have you had difficulty remembering things?” All other groups show very promising results, with over half of the groups having scores of 0.9 or higher. Groups of questions for which there were no annotations in the data, that is “Appetite loss,” “Global health,” “Insomnia,” “Sexual enjoyment,” “Sexual functioning,” and “Upset by hair loss,” have been omitted in the table. Even when considering much older data from the last 24 months as well, F1-scores do not drop below 0.5, with many scores even increasing, underlining the relevance of older data.

For 6 months of data, a total of 31 users have shared posts containing QoL information with on average 10.8% (SD 9.8%) of the 53 QoL questions answered. For the full 24 months of annotated data, there are 46 users with QoL-relevant posts, with 10.8% (SD 10.1%) of the questionnaire questions answered on average.

Table 7. F1-scores describing how well annotations predict the actual answers given in the survey, considering the last 6 and 24 months of data, using fine-grained and coarse-grained groups, which are filtered by time deviationa.
QuestionF1-score (6 mo)Total annotations (6 mo)F1-score (24 mo)Total annotations (24 mo)
Fine grouping
Fatigue1.050.944
Future perspective1.051.09
Arm symptoms1.041.06
Diarrhea1.030.616
Dyspnea1.020.914
Nausea and vomiting1.020.625
Body image1.011.01
Financial difficulties0.990.99
Emotional functioning0.9380.970
Social functioning0.9110.911
Physical functioning0.990.812
Breast symptoms0.970.911
Role functioning0.8100.721
Systemic therapy side effects0.7590.9189
Pain0.7150.861
Constipation0.720.810
Cognitive functioning0.010.53
Insomniab01.01
Appetite loss00.716
Coarse grouping
Functional0.9680.9120
Symptomatic0.81010.9385

aThe online survey using the EORTC QLQ-C30 and QLQ-BR23 questionnaires was conducted between May and June 2024 with patients diagnosed with breast cancer posting in the Inspire patient online forums.

bNot applicable.


The goal of this feasibility study is to investigate whether we can reliably gain QoL information from posts of patients in online communities. To this end, we have introduced a new methodology that examines the content of posts in relation to data from QoL surveys completed by the same cohort of patients whose posts we have access to. This has allowed us to not only assess how frequently users post information that is relevant to QoL, but also measure the degree to which this information corresponds to the answers to the corresponding questions in the questionnaires. This methodology is novel, as it has not previously been applied to analyzing whether information shared by patients online can reliably estimate individual QoL data. We have provided evidence that this methodology is sound, showing that the information can be coded at good levels of agreement and that there is some objective “ground truth” in the information shared by patients online that can be consistently interpreted as providing answers to the survey questions. With regard to RQ1, we can say that patient online forum posts indeed contain QoL-related information. The questions most frequently addressed in those posts (RQ2) are “Did you worry?” “Did you feel tense?” and “Have you had pain?” Moreover, the online posts align well with the actual QoL questionnaire responses (RQ3), achieving an F1-score of 0.7, with even better results when grouping similar questions (F1-score =0.8 for fine-grained and 0.9 for coarse-grained grouping).

In particular, Fleiss κ values were nonnegative with “Did you have headaches?” achieving the highest score of 1.0. Only 6 out of 53 questions showed agreement values of 0, with the mean Fleiss κ value being 0.5 (SD 0.28), indicating moderate to solid agreement among the annotators, especially considering the complexity of the task.

We have further quantified the degree to which patients share information that can be mapped to questions from the EORTC QLQ-C30 and QLQ-BR23 questionnaires.

From the total of 2669 annotated posts and comments, 613 (23%) answered at least 1 question from the used EORTC questionnaire. For the last 6 months (since the end of the survey), even 217 out of 580 posts (37.4%) contain QoL-relevant information. As these numbers might be affected by selection bias (indicated by older posts containing less QoL-relevant information), they should be seen more as upper bounds compared to scenarios where the data of an arbitrary patient are investigated.

We have further shown that there is a substantial correlation between the information that participants have provided in their online posts and the survey responses. We have proposed to rely on standard evaluation metrics as used in machine-learning research to estimate the correctness of the information provided by participants in their online posts compared to their answers to corresponding questions in the questionnaires.

The average score for the most recent 6 months of data across all 33 questionnaire questions that had answers in the respective data was an F1-score of 0.7, indicating that posts and comments containing QoL-relevant information can reliably predict questionnaire answers. Similarly, from the patient level, an average of 63.4% (SD 33.5%) of the estimated answers to the questions is correct. This indicates a high predictive power of forum posts which contain QoL information.

When grouping similar questions together, we even get a mean F1-score of 0.8 (SD 0.154) for fine-grained grouping and a mean F1-score of 0.9 (SD 0.01) for coarse-grained grouping when applying the time deviation filtering, excluding posts and comments talking about events and experiences from an unspecified or distant past. Without filtering, we get mean F1-scores of 0.7 (SD 0.353) without grouping, 0.8 (0.158) with fine-grained grouping, and 0.9 (SD 0.051) with coarse-grained grouping.

The most frequently answered questions (RQ2) include “Did you worry?,” “Did you feel tense?,” “Have you had pain?,” “Were you limited in doing either your work or other daily activities?,” “Were you tired?,” “Have you felt nauseated?,” “Were you worried about your health in the future?,” “Have you lacked appetite?,” and “Did you feel depressed?” These questions are not particularly specific to breast cancer, thus indicating patients typically discuss their QoL online more on a general level or in such a specialized manner that none of the EORTC questions fit well as shown by the high number of “Did you feel ill or unwell?” occurrences.

One inherent problem of the collection and analysis of the questionnaire data is related to missing data, as often patients do not fill out all information. In our case, the rate of completeness for the survey data was 100%, as the completion of all questionnaire items was required. Two questions of the form “If you were/had …” allowed choosing “no answer,” which, however, we do not consider missing data but instead a different way to answer the question.

However, not all questions were answered in posts of all patients. For patients who have shared QoL-relevant information in their posts, only about 10.8% (SD 9.8%) of the questions could be answered by information given in the posts. This means that the available data are typically not sufficient to answer all questions of the EORTC QoL questionnaire, but at least some parts can be reliably extracted for each patient. We also investigated ways to deal with data sparsity by grouping questions and considering longer time periods, which help to get more insights into the QoL without losing too much reliability.

Overall, our results show that users of online communities, such as Inspire, indeed post information about their QoL, providing important information that can be consistently and objectively mapped to the categories of standard questionnaires.

Our results have important implications. First, the results suggest that we could potentially seek answers to questions that are typically not included in standardized QoL questionnaires. Furthermore, there are nuances in the posts of patients that are not directly captured by the questionnaires. On one hand, the topics that patients do or do not bring up in online discussions may indicate which topics are currently the most important for that patient, potentially referring to specific aspects of the relatively broad areas covered by the QoL questionnaires. On the other hand, this can provide insights into areas, which are not covered by the QLQ because they are unique to that patient (eg, due to certain living circumstances). This hints at the potential of our methodology to complement data obtained from standard QoL questionnaires.

Comparison With Prior Work

In recent years, there have been numerous studies investigating whether QoL-related information can be gained from unstructured data sources. However, our work differs from related approaches in substantial ways.

For one, our approach leverages different types of data considered. We considered social media posts from dedicated health care online communities or more precisely the breast cancer forum of Inspire. Other work has focused on clinical reports [30,31] or health records [31,32], which, in spite of also having unstructured elements, considerably differ in terms of content and style from social media data. Among similar approaches considering social media, many approaches have sourced Twitter data [13,15,33-37], in contrast to the specific health care online community data that we used. An exception is the work by Leung et al [38,39] who have used data from cancer support groups. However, their data originate from chats moderated by health care professionals instead of posts from organic, forum-like online communities. Their study had a narrower focus by studying signs of emotional distress in patients participating in the online discussions instead of the broader focus of capturing QoL-related information in our study. An overview of how social media contents can be analyzed for health care is given by Fu et al [40].

Our study has aimed to gain insights into the QoL of individual patients based on their posts in health care communities. In contrast, the goal of many other studies is to discover aggregate trends for a certain geographic segment of the target population [13,15,36]. Other studies are interested in population-wide trends and effects related to how a disease, intervention, or treatment affects patients in general, instead of estimating values for specific users [9,11,12,37,41-52]. Such population-wide analyses can support the creation of conceptual models of a disease [9,42], whereas other studies have focused on general and individual health [33], recognition of discussions about health services [53], or emotional distress in participants of online chats [38,39].

Some studies have been conducted in relation to existing QoL questionnaires. For example, Tapi Nzali et al [54] have also used the EORTC QLQ-C30 and QLQ-BR23 questionnaires, aiming at identifying general topics that patients diagnosed with breast cancer discuss in relation to EORTC questionnaires. Gries and Fastenau [52] have relied on items in the EORTC QLQ-C30 and QLQ-MY20 questionnaires to investigate concepts of interest, symptoms, and impacts discussed by patients with multiple myeloma on a specific health care social media platform. These studies analyze general trends rather than predicting QoL for individual patients.

Renner et al [55] annotated data from 19 health-related forums using the Short Form Health Survey SF-36 [56] and Euro Quality of Life 5 Dimensions [57], but they did not compare and contrast survey responses with the information in patient posts as we do.

There are few studies that have analyzed the information posted by patients in relation to ground truth QoL data as available in the existing QoL questionnaire data. For example, Zivanovic et al [15] have compared Twitter data originating from the city of Bristol with official QoL data for this city. They analyzed the data on an aggregate level for the whole city rather than for individual patients as we do. Twitter has also been used by Sarma et al [35] who compared information obtained from Tweets with ground truth data of Centers for Disease Control Healthy Days Questionnaire [58]. The questionnaire used is much shorter than the one used in our study, providing a less detailed view on patients’ QoL. Cotté et al [16] have compared data from cancer-specific forums with data obtained from the Functional Assessment of Cancer Therapy-General [59] and EORTC QLQ-C30 questionnaires. Their main focus has been on investigating the impact of immune checkpoint inhibitors, but not investigating the QoL impact for individual users.

Although this work focuses on breast cancer, related work covers a wide range of diseases. These diseases include Parkinson disease [9,10], amblyopia [41], presbyopia [42,49], cancer [16,30,48,51], dry eye disease [43], chronic ocular pain [44], Sjögren disease [11], chronic obstructive pulmonary disease [12], systemic lupus erythematosus [45,60], immunoglobulin A nephropathy [46], complement 3 glomerulopathy [47], atopic dermatitis [50,61], obesity [36], depression [62], and (general) mental health [13,14].

In summary, our work advances the field by paving the way for a new option to monitor patients’ QoL. As elaborated in the previous section, analyzing forum posts with respect to QoL questions can be used to complement existing instruments for assessing QoL, helping to focus on the most relevant aspects of patients’ QoL as well as to discover areas potentially not covered by existing instruments in detail. Additionally, possible ways to automate the process of identifying QoL-relevant information in online posts using, for example, artificial intelligence methods, pose an interesting path of future research.

Limitations

Despite carefully annotating the data and collecting ground truth data through a survey, our work has multiple limitations, which are briefly discussed in this section.

In our study, the investigated population has been limited to those active on the Inspire platform. However, many other online patient communities exist, and both the frequency of QoL-relevant utterances as well as their predictive power might be different.

Similarly, the fact that older posts in our dataset discuss QoL-relevant topics less often indicates that selection bias might be present in our data, as patients currently interested in QoL might be more likely to have participated in our study. Thus, our results have to be considered more like an upper bound to the possible insights about patients’ QoL from posts in online forums.

Moreover, we also found signs of recall bias in the data. For example, we found that some patients talked about how much they worried about different topics shortly before participating in the study, but then answered that they did not worry at all. The occurrence of such phenomena has not been systematically investigated in this work and might thus have an impact on the results.

Certain highly intimate topics, such as aspects related to sexuality or body image, were rarely discussed in the forum posts. Given the public nature of the data, which is to be expected, however, it represents a limitation in terms of depth and scope of insights that can be obtained from such a dataset. Compared to one-on-one interviews, forum data may provide a less comprehensive picture of patients’ concerns in these areas.

Finally, data sparsity and unequal distribution of posts and comments among forum users are typical problems when considering social media. As illustrated by the long-tail distribution in Figure 2, there are a few users with many posts and comments in our dataset, but the majority of users have much fewer posts available. The degree to which we can obtain QoL information is thus dependent on the posting behavior of a particular user. Statistics and numbers presented in this work thus have to be interpreted as upper bounds for the potential to extract QoL information from a randomly chosen user of the Inspire communities. Similarly, due to the general data sparsity, labels that are scarcely covered by our data sample might be present more prominently for other users, which have not been part of our study, and analogously, labels with many occurrences in our data might be more rare to find for other users.

Conclusions

This study has investigated the feasibility of gaining insights about QoL from posts of patients in online forums. Toward this goal, we have recruited patients diagnosed with breast cancer from Inspire. These patients have filled the EORTC Quality of Life Questionnaire and allowed us to analyze their posts and comments in relation to their answers given in the questionnaire. By this, we have been able to compare the information provided in their posts with the answers provided in the questionnaires. Our key findings are that between 23% and 37.4% of patients’ posts contain information that can be mapped to at least one of the questions of the questionnaire. Furthermore, we could show that the QoL information in posts is reliable, corresponding to a large extent to the answers given in the questionnaire. Our findings thus show that the information shared by patients in online posts can contribute to a better understanding of how the disease affects their QoL.

While the information provided by patients in posts shared in online forums cannot replace the collection of data via questionnaires, it represents a low-cost, low-burden alternative to collect insights about QoL effectively. Most importantly, the information in posts can help us to discover aspects related to disease burden and QoL that are not captured by existing questionnaires. This could even provide guidance on how survey instruments could be extended to cover those aspects that are most burdensome to patients.

Future work should investigate how exactly the information provided by patients online can complement existing QoL assessment methodologies. A further line of research relates to how the information can be extracted automatically, in order to avoid having to scan and read thousands of posts. Given the current state of research on information extraction, this seems to be a promising avenue to investigate. Additionally, first experiments conducted in this direction yielded promising results.

All in all, our work contributes to the foundation for a novel low-burden alternative for monitoring patients’ QoL and complementing existing QoL instruments. Moreover, this opens up interesting new avenues of research on how to automate the process including using artificial intelligence methods and thus further lowering the burden for both patients and clinical staff.

Acknowledgments

The authors thank the European Organisation for Research and Treatment of Cancer for giving them the permission to use their questionnaires QLQ-C30 and BR-23 (request IDs 96009 and 102328). They also thank Antonia Frank for her help with annotating the data and developing the annotation guidelines. No generative artificial intelligence tools were used to write this paper.

Funding

This work is partially funded by the Ministry of Culture and Science of the State of North Rhine-Westphalia under grant number NW21-059A (SAIL). The funder had no involvement in the study design, data collection, analysis, interpretation, or the writing of the manuscript.

Data Availability

The datasets generated or analyzed during this study are not publicly available due to data privacy reasons.

Authors' Contributions

DMS is the main author of all sections of this paper. He is the main author of the annotation guidelines and annotated the data, supervised the annotation process, and conducted all experiments and results presented in this paper. RS contributed to the Methods and Limitations sections as well as the annotation guideline development and annotated the data. BPHC, DK, and VK implemented the survey as well as contributed to all sections of the paper. AH helped with the survey construction as well as contributed to the Introduction and Methods sections. PC supervised the work and contributed to all sections. All authors reviewed and approved the manuscript.

Conflicts of Interest

BR-HC, DK, and VK are employees of Inspire, the online community where we conducted the survey and collected the forum data. PC is a shareholder of Semalytix, a company providing social media listening services to the life science industry. All other authors declare no conflicts of interest.

Multimedia Appendix 1

Annotation guidelines and insights from annotation.

DOCX File, 22 KB

Multimedia Appendix 2

Distribution of annotations between labeled questions.

DOCX File, 54 KB

Checklist 1

CHERRIES checklist.

PDF File, 73 KB

  1. Constitution of the World Health Organization. World Health Organization; 1948. URL: https:/​/www.​globalhealthrights.org/​wp-content/​uploads/​2013/​10/​Constitution-of-the-World-Health-Organization-WHO.​pdf [Accessed 2026-04-06]
  2. Aaronson NK, Ahmedzai S, Bergman B, et al. The European Organization for Research and Treatment of Cancer QLQ-C30: a quality-of-life instrument for use in international clinical trials in oncology. J Natl Cancer Inst. Mar 3, 1993;85(5):365-376. [CrossRef] [Medline]
  3. Jacobsen PB, Davis K, Cella D. Assessing quality of life in research and clinical practice. Oncology (Williston Park). Sep 2002;16(9 Suppl 10):133-139. [Medline]
  4. Detmar SB, Aaronson NK. Quality of life assessment in daily clinical oncology practice: a feasibility study. Eur J Cancer. Jul 1998;34(8):1181-1186. [CrossRef] [Medline]
  5. Bezjak A, Ng P, Skeel R, Depetrillo AD, Comis R, Taylor KM. Oncologists’ use of quality of life information: results of a survey of Eastern Cooperative Oncology Group physicians. Qual Life Res. 2001;10(1):1-13. [CrossRef] [Medline]
  6. Buxton J, White M, Osoba D. Patients’ experiences using a computerized program with a touch-sensitive video monitor for the assessment of health-related quality of life. Qual Life Res. Aug 1998;7(6):513-519. [CrossRef] [Medline]
  7. Velikova G, Brown JM, Smith AB, Selby PJ. Computer-based quality of life questionnaires may contribute to doctor-patient interactions in oncology. Br J Cancer. Jan 7, 2002;86(1):51-59. [CrossRef] [Medline]
  8. Giesinger J, Kemmler G, Meraner V, et al. Towards the implementation of quality of life monitoring in daily clinical routine: methodological issues and clinical implication. Breast Care (Basel). 2009;4(3):148-154. [CrossRef] [Medline]
  9. Staunton H, Kelly K, Newton L, et al. A patient-centered conceptual model of symptoms and their impact in early Parkinson’s disease: a qualitative study. J Parkinsons Dis. 2022;12(1):137-151. [CrossRef] [Medline]
  10. Javidnia M, Arbatti L, Hosamath A, Eberly SW, Oakes D, Shoulson I. Predictive value of verbatim Parkinson’s disease patient-reported symptoms of postural instability and falling. J Parkinsons Dis. 2021;11(4):1957-1964. [CrossRef] [Medline]
  11. Perella C, Steenackers M, Robbins B, et al. Patient experience of Sjögren’s disease and its multifaceted impact on patients’ lives. Rheumatol Ther. Jun 2023;10(3):601-614. [CrossRef] [Medline]
  12. Cook NS, Kostikas K, Gruenberger JB, et al. Patients’ perspectives on COPD: findings from a social media listening study. ERJ Open Res. Feb 2019;5(1):00128-2018. [CrossRef] [Medline]
  13. McClellan C, Ali MM, Mutter R, Kroutil L, Landwehr J. Using social media to monitor mental health discussions - evidence from Twitter. J Am Med Inform Assoc. May 1, 2017;24(3):496-502. [CrossRef] [Medline]
  14. Karmen C, Hsiung RC, Wetter T. Screening Internet forum participants for depression symptoms by assembling and enhancing multiple NLP methods. Comput Methods Programs Biomed. Jun 2015;120(1):27-36. [CrossRef] [Medline]
  15. Zivanovic S, Martinez J, Verplanke J. Capturing and mapping quality of life using Twitter data. GeoJournal. Feb 2020;85(1):237-255. [CrossRef]
  16. Cotté FE, Voillot P, Bennett B, et al. Exploring the health-related quality of life of patients treated with immune checkpoint inhibitors: social media study. J Med Internet Res. Sep 11, 2020;22(9):e19694. [CrossRef] [Medline]
  17. Sung H, Ferlay J, Siegel RL, et al. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. May 2021;71(3):209-249. [CrossRef] [Medline]
  18. Arraras JI, Manterola A, Asin G, et al. Quality of life in elderly patients with localized breast cancer treated with radiotherapy. A prospective study. Breast. Apr 2016;26:46-53. [CrossRef] [Medline]
  19. Carreira H, Williams R, Dempsey H, Stanway S, Smeeth L, Bhaskaran K. Quality of life and mental health in breast cancer survivors compared with non-cancer controls: a study of patient-reported outcomes in the United Kingdom. J Cancer Surviv. Aug 2021;15(4):564-575. [CrossRef] [Medline]
  20. Mols F, Vingerhoets AJJM, Coebergh JW, van de Poll-Franse LV. Quality of life among long-term breast cancer survivors: a systematic review. Eur J Cancer. Nov 2005;41(17):2613-2619. [CrossRef] [Medline]
  21. Montazeri A. Health-related quality of life in breast cancer patients: a bibliographic review of the literature from 1974 to 2007. J Exp Clin Cancer Res. Aug 29, 2008;27(1):32. [CrossRef] [Medline]
  22. Nolte S, Liegl G, Petersen MA, et al. General population normative data for the EORTC QLQ-C30 health-related quality of life questionnaire based on 15,386 persons across 13 European countries, Canada and the Unites States. Eur J Cancer. Jan 2019;107:153-163. [CrossRef] [Medline]
  23. Hinz A, Singer S, Brähler E. European reference values for the quality of life questionnaire EORTC QLQ-C30: results of a German investigation and a summarizing analysis of six European general population normative studies. Acta Oncol. Jul 2014;53(7):958-965. [CrossRef] [Medline]
  24. Eysenbach G. Improving the quality of web surveys: the checklist for reporting results of internet e-surveys (CHERRIES). J Med Internet Res. Sep 29, 2004;6(3):e34. [CrossRef] [Medline]
  25. Grishman R, Sundheim B. Message understanding conference - 6: a brief history. Presented at: Proceedings of the 16th International Conference on Computational Linguistics (COLING 1996); Aug 5-9, 1996. [CrossRef]
  26. Li J, Sun A, Han J, Li C. A survey on deep learning for named entity recognition. IEEE Trans Knowl Data Eng. 2022;34(1):50-70. [CrossRef]
  27. Advanced Data Mining Techniques. Springer; 2008. [CrossRef]
  28. Fleiss JL. Measuring nominal scale agreement among many raters. Psychol Bull. Nov 1971;76(5):378-382. [CrossRef]
  29. Hsu YH, Shen WC, Wang CH, Lin YF, Chen SC. Hand-foot syndrome and its impact on daily activities in breast cancer patients receiving docetaxel-based chemotherapy. Eur J Oncol Nurs. Dec 2019;43:101670. [CrossRef] [Medline]
  30. Naseri H, Kafi K, Skamene S, et al. Development of a generalizable natural language processing pipeline to extract physician-reported pain from clinical reports: generated using publicly-available datasets and tested on institutional clinical reports for cancer patients with bone metastases. J Biomed Inform. Aug 2021;120:103864. [CrossRef] [Medline]
  31. Forsyth AW, Barzilay R, Hughes KS, et al. Machine learning methods to extract documentation of breast cancer symptoms from electronic health records. J Pain Symptom Manage. Jun 2018;55(6):1492-1499. [CrossRef] [Medline]
  32. Pakhomov SVS, Shah ND, Van Houten HK, Hanson PL, Smith SA. The role of the electronic medical record in the assessment of health related quality of life. AMIA Annu Symp Proc. 2011;2011:1080-1088. [Medline]
  33. Yin Z, Fabbri D, Rosenbloom ST, Malin B. A scalable framework to detect personal health mentions on Twitter. J Med Internet Res. Jun 5, 2015;17(6):e138. [CrossRef] [Medline]
  34. Riga M, Stocker M. Atmospheric environment and quality of life information extraction from Twitter with the use of self-organizing maps. J Environ Inform. 2015;26. [CrossRef]
  35. Sarma KV, Spiegel BMR, Reid MW, et al. Estimating the health-related quality of life of Twitter users using semantic processing. Stud Health Technol Inform. Aug 21, 2019;264:1065-1069. [CrossRef] [Medline]
  36. Zhang X, Athanasiadou R, Razavian N. Tracing state-level obesity prevalence from sentence embeddings of tweets: a feasibility study. Presented at: Heterogeneous Data Management, Polystores, and Analytics for Healthcare (DMAH 2020, Poly 2020); Aug 31 to Sep 4, 2020. [CrossRef]
  37. Shankar R, Xu Q, Bundele A. Patient voices in dialysis care: sentiment analysis and topic modeling study of social media discourse. J Med Internet Res. May 15, 2025;27:e70128. [CrossRef] [Medline]
  38. Leung YW, Wouterloot E, Adikari A, et al. Natural language processing-based virtual cofacilitator for online cancer support groups: protocol for an algorithm development and validation study. JMIR Res Protoc. Jan 7, 2021;10(1):e21453. [CrossRef] [Medline]
  39. Leung YW, Park B, Heo R, et al. Providing care beyond therapy sessions with a natural language processing-based recommender system that identifies cancer patients who experience psychosocial challenges and provides self-care support: pilot study. JMIR Cancer. Jul 29, 2022;8(3):e35893. [CrossRef] [Medline]
  40. Fu J, Li C, Zhou C, et al. Methods for analyzing the contents of social media for health care: scoping review. J Med Internet Res. Jun 26, 2023;25:e43349. [CrossRef] [Medline]
  41. Syntosi A, Felizzi F, Bouchet C. A social media listening study to understand the unmet needs and quality of life in adult and pediatric amblyopia patients. Ophthalmol Ther. Dec 2022;11(6):2183-2196. [CrossRef] [Medline]
  42. Findley A, Sharma G, Bentley S, et al. Comparison of literature review, social media listening, and qualitative interview research methods in generating patient-reported symptom and functional impact concepts of presbyopia. Ophthalmol Ther. Feb 2023;12(1):501-516. [CrossRef] [Medline]
  43. Cook N, Mullins A, Gautam R, et al. Evaluating patient experiences in dry eye disease through social media listening research. Ophthalmol Ther. Sep 2019;8(3):407-420. [CrossRef] [Medline]
  44. Goel K, Parashar N, Aasaithambi S, Verma H, O’Brien P, Sloesen B. P33 patient experiences and insights on chronic ocular pain from a social media listening study. Value Health. Jan 2022;25(1):S8. [CrossRef]
  45. Spies E, Andreu T, Koelling J, Hartung M, Kamudoni P, Park J. PCR251 retrospective social listening study of patients living with systemic lupus erythematosus (SLE): understanding the patient experience. Value Health. Dec 2022;25(12):S438. [CrossRef]
  46. Tyagi N, Aasaithambi S, Chauhan J, George A, Zaour N. PUK32 patient insights for immunoglobulin A nephropathy (IgAN) using social media listening. Value Health. Nov 2019;22:S919. [CrossRef]
  47. Tyagi N, Aasaithambi S, Chauhan J, George A, Zaour N. PUK33 patient insights for complement 3 glomerulopathy (C3G) using social media listening. Value Health. Nov 2019;22:S919. [CrossRef]
  48. Tadmouri A, Alivon M, Andreu T, et al. RWD141 exploration of melanoma patient-generated real-world data using an AI-based social listening approach. Value Health. Dec 2022;25(12):S476-S477. [CrossRef]
  49. Wolffsohn JS, Leteneux-Pantais C, Chiva-Razavi S, et al. Social media listening to understand the lived experience of presbyopia: systematic search and content analysis study. J Med Internet Res. Sep 21, 2020;22(9):e18306. [CrossRef] [Medline]
  50. Voillot P, Riche B, Portafax M, et al. Social media platforms listening study on atopic dermatitis: quantitative and qualitative findings. J Med Internet Res. Jan 28, 2022;24(1):e31140. [CrossRef] [Medline]
  51. Rodrigues A, Chauhan J, Sagkriotis A, Aasaithambi S, Montrone M. Understanding the lived experience of lung cancer: a European social media listening study. BMC Cancer. Apr 30, 2022;22(1):475. [CrossRef] [Medline]
  52. Gries KS, Fastenau J. Using a digital patient powered research network to identify outcomes of importance to patients with multiple myeloma. J Patient Rep Outcomes. Sep 1, 2020;4(1):74. [CrossRef] [Medline]
  53. Yang YC, Al-Garadi MA, Bremer W, Zhu JM, Grande D, Sarker A. Developing an automatic system for classifying chatter about health services on Twitter: case study for Medicaid. J Med Internet Res. May 3, 2021;23(5):e26616. [CrossRef] [Medline]
  54. Tapi Nzali MD, Bringay S, Lavergne C, Mollevi C, Opitz T. What patients can tell us: topic analysis for social media on breast cancer. JMIR Med Inform. Jul 31, 2017;5(3):e23. [CrossRef] [Medline]
  55. Renner S, Marty T, Khadhar M, et al. A new method to extract health-related quality of life data from social media testimonies: algorithm development and validation. J Med Internet Res. Jan 28, 2022;24(1):e31528. [CrossRef] [Medline]
  56. Ware Jr JE. SF-36 health survey. In: The Use of Psychological Testing for Treatment Planning and Outcomes Assessment. 2nd ed. Lawrence Erlbaum Associates; 1999:1227-1246. ISBN: 0585176663
  57. EuroQol Group. EuroQol--a new facility for the measurement of health-related quality of life. Health Policy. Dec 1990;16(3):199-208. [CrossRef] [Medline]
  58. Taylor VR. Measuring healthy days: population assessment of health-related quality of life. Centers for Disease Control and Prevention (CDC); 2000. URL: https://stacks.cdc.gov/view/cdc/6406
  59. Cella DF, Tulsky DS, Gray G, et al. The Functional Assessment of Cancer Therapy scale: development and validation of the general measure. J Clin Oncol. Mar 1993;11(3):570-579. [CrossRef] [Medline]
  60. Matsuda S, Ohtomo T, Tomizawa S, et al. Incorporating unstructured patient narratives and health insurance claims data in pharmacovigilance: natural language processing analysis of patient-generated texts about systemic lupus erythematosus. JMIR Public Health Surveill. Jun 29, 2021;7(6):e29238. [CrossRef] [Medline]
  61. Falissard B, Simpson EL, Guttman-Yassky E, et al. Qualitative assessment of adult patients’ perception of atopic dermatitis using natural language processing analysis in a cross-sectional study. Dermatol Ther (Heidelb). Apr 2020;10(2):297-305. [CrossRef] [Medline]
  62. Tejaswini V, Sathya Babu K, Sahoo B. Depression detection from social media text analysis using natural language processing techniques and hybrid deep learning model. ACM Trans Asian Low Resour Lang Inf Process. Jan 31, 2024;23(1):1-20. [CrossRef]


CHERRIES: Checklist for Reporting Results of Internet E-Surveys
EORTC: European Organisation for Research and Treatment of Cancer
FN: false negatives
FP: false positives
QoL: quality of life
TN: true negatives
TP: true positives


Edited by Matthew Balcarras; submitted 15.Apr.2025; peer-reviewed by Ali AL-Asadi, Álvaro García-Barragán, Thierry Escudier; final revised version received 27.Jan.2026; accepted 23.Feb.2026; published 30.Apr.2026.

Copyright

© David Maria Schmidt, Raoul Schubert, Brian Po-Han Chen, Deborah Kuk, Valmeek Kudesia, Andreas Hinz, Philipp Cimiano. Originally published in JMIR Cancer (https://cancer.jmir.org), 30.Apr.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Cancer, is properly cited. The complete bibliographic information, a link to the original publication on https://cancer.jmir.org/, as well as this copyright and license information must be included.