Skip to main content

Buffering against exposure to mental health misinformation in online communities on Facebook: the interplay of depression literacy and expert moderation



The proliferation of health misinformation on social media is a growing public health concern. Online communities for mental health (OCMHs) are also considered an outlet for exposure to misinformation. This study explored the impact of the self-reported volume of exposure to mental health misinformation in misinformation agreement and the moderating effects of depression literacy and type of OCMHs participation (expert vs. peer-led).


Participants (n = 403) were recruited in Italian-speaking OCMHs on Facebook. We conducted regression analyses using PROCESS macro (moderated moderation, Model 3). Measures included: the Depression Literacy Questionnaire (Griffiths et al., 2004), the self-reported misinformation exposure in the OCMHs (3 items), and misinformation agreement with the exposure items (3 items). Whether participants were members of expert or peer-led OCMHs was also investigated.


The final model explained the 12% variance in the agreement. There was a positive and significant relationship between misinformation exposure and misinformation agreement (β = 0.3221, p < .001), a significant two-way interaction between misinformation exposure and depression literacy (β = − 0.2179, p = .0014 ), and between self-reported misinformation exposure and type of OCMH (β = − 0.2322, p = .0254), such that at higher levels of depression literacy and in case of participation to expert-led OCMHs, the relationship misinformation exposure-misinformation agreement was weaker. Finally, a three-way interaction was found (β = 0.2497, p = .0144) that showed that depression literacy moderated the positive relationship between misinformation exposure and misinformation agreement such that the more misinformation participants were exposed to, the more they agreed with it unless they had higher levels of depression literacy; this, however, occurred only if they participated in peer-led groups.


Results provide evidence that the more members reported being exposed to mental health misinformation, the more they tended to agree with it, however this was only visible when participants had lower depression literacy and were participating in peer-led OCMHs. Results of this study suggest that both internal factors (i.e., high depression literacy) and external factors (the type of online community individuals were participating in) can buffer the negative effects of misinformation exposure. It also suggests that increasing depression literacy and expert community moderation could curb the negative consequences of misinformation exposure related to mental health. Results will guide interventions to mitigate the effects of misinformation in OCMHs, including encouraging health professionals in their administration and implementing health education programs.

Peer Review reports


Health misinformation and the illusory truth effect

Due to the increasing popularity of the internet [1, 2], more specifically of social media [3] as a venue for seeking and sharing health information, there is a growing concern about the spread of health misinformation [4, 5]. Recently, these concerns have intensified due to the COVID-19 pandemic [6]. A recent systematic review of reviews found that the prevalence of health-related misinformation on social media ranged from 0.2 to 28.8% [7].

Swire-Thompson and colleagues [8] define misinformation as “information that is contrary to the epistemic consensus of the scientific community regarding a phenomenon”. Health misinformation is a specific type of misinformation that refers to a “health-related claim of fact that is currently false due to a lack of scientific evidence” [9]. As users generate information on social media, it can be subjective or inaccurate, therefore a worrisome source of health misinformation [6, 10] as it can also be archived and persist over time until it is corrected or deleted, becoming a dangerous resource for future health information seekers [11].

Public health researchers and practitioners are increasingly preoccupied with the potential for health misinformation to misinform and mislead the public as it not only creates erroneous health beliefs confusion and reduces trust in health professionals but can also “delay or prevent effective care, in some cases threatening the lives of individuals” [5, 10]. Thus, combating its effects has become crucial for public health [9, 12] and can be accomplished only by understanding its psychological drivers [13] and complementary buffers. One particularly prominent finding that helps explain why people are susceptible to misinformation is the ‘illusory truth effect’, according to which repeated information is perceived as more truthful than new information [14,15,16,17,18].

In the context of the illusory truth effect, it has been found that the effects of repeated exposure to misinformation on perceptions of accuracy disappeared when the receiver knew the actual truth and that people were more likely to believe misinformation when they were unfamiliar with the issue at hand [19]. Other studies have shown that knowledge is key to buffering against misinformation exposure [20,21,22].

Mental health misinformation: underestimated and understudied

According to a recent systematic review, previous studies on health misinformation on social media have concentrated more on topics of physical-related illnesses such as vaccines (32%), drugs or smoking (22%), non-communicable diseases (19%) and pandemics (10%) [23]. Few studies have examined misinformation regarding mental health specifically, although this might be crucial for two reasons. Firstly, mental health is a growing public health concern [24], which has been underestimated even though about 14% of the global disease burden has been attributed to neuropsychiatric disorders such as depression [25]. Secondly, mental health conditions are frequently stigmatized and misunderstood, resulting in a greater prevalence of misinformation online and offline [26,27,28]. Therefore, it is crucial to understand the conditions that can mitigate the outcomes of mental health misinformation exposure.

Mental health misinformation in online communities

Online health communities are virtual platforms where individuals (or caregivers) with similar health conditions or concerns gather to share information, seek support, and engage in discussions related to health. These communities are typically specific to a specific illness and are highly prevalent for chronic and marginalized diseases [29], such as mental health disorders. Although these communities can exist in different forms such as forums [30] or as social media groups such as on Facebook [31, 32] or Reddit [33] they all share similar affordances [34] such as the question-and-answer format. Online communities generally rely on the work of volunteers to police themselves [35], some of them being health professionals, others peers with no expert credentials [31, 36, 37]. Although healthcare professionals play a critical role in ensuring information quality in online health communities [38, 39], the literature on this topic needs to be more extensive, especially regarding the differences that these two types of groups might perform, particularly about misinformation.

Online communities for mental health symptoms (OCMHs) are communities specific for mental health topics that serve as virtual spaces where individuals suffering from mental health conditions (or their caregivers) can connect with others experiencing similar challenges. These communities cater to a wide range of mental health topics, including general discussions about mental health [40] as well as more specific topics such as depression and anxiety [33, 41].

OCMHs are becoming increasingly present on social networking sites, especially among younger generations [42] and they can also be considered an outlet for misinformation. A recent content analysis [43] has found extremely high levels of misinformation in OCMHs, and even communities moderated by health professionals (expert-led) were not exempt from this issue. This is in line with other studies showing that healthcare professionals can also spread misinformation in various ways [10].

Research gaps and current study

The present study aims to investigate the relationship between exposure to mental health misinformation in Italian online communities for mental health on Facebook and related agreement, focusing on two aspects that might impact this relationship and the interplay between them: depression literacy and type of OCMHs moderation.

Health literacy is at the heart of any discussion of health-related misinformation, which can be defined as “the degree to which people have the capacity to obtain, process, and understand basic health information and services needed to make appropriate health decisions” [44]. In a systematic review, results indicated that low health literacy was negatively related to the ability to evaluate online health information [45]. Previous studies have shown that knowledge moderated the relationship between exposure and beliefs but focused on other aspects of literacy, such as news literacy [46] or media literacy [47]. One study found that a higher level of cancer literacy helped participants identify misinformation and prevented them from being persuaded by it [48].

As in the health context, different types of literacy exist in different contexts; in the present study, we focused on declarative knowledge about depression literacy, depression being one of the most common mental illnesses in Italy [49, 50] and worldwide [51]. Depression literacy [52] is a facet of the mental health literacy concept, the latter defined as an individual’s knowledge regarding mental health [53], but with a specific focus on depression intended as a major depressive episode. Mental health literacy, or lack thereof, has been used as a possible factor to explain uncertainties or lack of knowledge about mental health and the ensuing effects on effective treatment and care [54]. Determining whether depression literacy levels can buffer the effects of misinformation exposure is critical, also as identifying which segments of the population are especially vulnerable to health misinformation and developing interventions for individuals at risk.

However, there may be more effective approaches than focusing on individual differences in susceptibility to misinformation while ignoring other potential external factors. As we mentioned earlier, OCMH typologies can vary depending on their moderators’ expertise (peers or mental health experts) and this factor might influence the relationship between exposure and agreement with misinformation. Content moderation scholars posit that content moderation “has exploded as a public, advocacy, and policy concern” [55]. Typically platforms such as Facebook use a combination of algorithmic tools, user reporting and human review [56]. Previous research on misinformation has primarily focused on automated content moderation implemented by platforms through machine learning classifiers [57, 58] and has demonstrated the effectiveness of platforms’ content moderation practices in mitigating the spread of conspiracy theories and other forms of misinformation [59]. However, the effects of human moderation have not been investigated extensively, especially in the context of mental health misinformation, a type of misinformation that poses unique challenges as the complexity of mental health issues requires nuanced approaches to content moderation. However, until now, the literature has predominantly focused on examining the social effects of moderation within health communities (e.g [60, 61]), rather than its significance in countering misinformation. Previous research has in fact demonstrated that the knowledge and guidance provided by peer patients differed significantly from that offered by professional healthcare providers [62].

Therefore, it is critical to investigate potential differences between these groups in misinformation exposure outcomes. Given that the quality and accuracy of online health information provided by OCMHs can vary significantly [63], it is crucial to consider the potential implications of participating in OCMHs with different content moderation types, as this may affect the degree of exposure to health-related misinformation and its associated consequences.

Furthermore, the interplay between internal (depression literacy) and external (type of OCMHs moderation) factors sheds light on the most vulnerable individuals within online communities. This approach will expand upon previous research on individual differences in misinformation susceptibility [16, 22] and provide a more comprehensive understanding of the complex dynamics that may influence agreement with misinformation exposure within OCMHs.

To the best of our knowledge, no study has examined whether and how individual differences such as depression literacy might interact with the external factors embedded in online communities, including the type of moderators’ expertise. However, it is crucial to explore their potential interaction as previous research highlighted the heterogeneous nature of users within online communities, with varying levels of health literacy [64]. Additionally, it is plausible that different types of guidance provided by moderators, depending on their expertise levels, might be less or more able to mitigate the effects of poor literacy, particularly when users are exposed to high volumes of misleading content.

Understanding how depression literacy and moderator expertise interact in online communities can provide valuable insights into optimizing mental health support and interventions. By exploring this interaction, we can identify effective strategies for addressing the challenges posed by varying levels of health literacy and the diverse expertise of moderators. This research can inform the development of targeted interventions and improve outcomes for individuals seeking support for depression within online communities.

We have chosen to focus our research on Italian-speaking OCMHs on Facebook as OCMHs are transitioning from to forums social networking sites [42]. Furthermore, while Facebook’s overall usage worldwide is declining, particularly among younger generations [65], it remains one of the most widely used social networking platforms in Italy as in 2022, 77.5% of surveyed Italian internet users reported using Facebook [66]. Moreover, differently from other Italian-speaking forums or mental health subreddit communities, we have observed a vast amount of active OCMHs communities on the Facebook social media platform, making it a suitable case study. Furthermore, as claimed by Bayer and colleagues [67], as social media will continue to iterate, emerging research can extrapolate the findings from one platform to other platforms such as that “even when a platform is decommissioned, findings linked to its elements can be compared to future channels that share characteristics within the same element”. In other words, we believe that our findings could have the potential to extend beyond the specific OCMHs under study and be used to build research also across other social media platforms.

Hypotheses and research question

We tested the following hypotheses using a moderated moderation model (see Fig. 1 for the hypothesized model). Model 3, as proposed by Hayes [68], is a statistical model used to analyze moderated moderation effects that extends the traditional moderation analysis by examining the interactive effects of two moderating variables on the relationship between an independent variable and a dependent variable. The model allowed us to examine the interaction effects between the two moderating variables (1) depression literacy (individual difference) and (2) type of OCMHs (external factor) on the relationship between the independent (misinformation exposure) and dependent (agreement with misinformation) variables.

First, based on the literature on the Illusory truth effect [69], we expect that:


Misinformation exposure will be positively and significantly associated with misinformation agreement.

However, based on the above literature on the influence of the protective role of knowledge [46,47,48] we hypothesize that:


The positive association between misinformation exposure and misinformation agreement will be moderated by depression literacy.

Furthermore, we expect that:


The positive association between misinformation exposure and misinformation agreement will be moderated by type of OCMHs participation.

Then, as a research question, we will test whether depression literacy and type of OCMHs participation also interact with each other in the following way:


The moderating effect of depression literacy on the relationship between misinformation exposure and misinformation agreement will be further moderated by the type of OCMHs participation.

Fig. 1
figure 1

Hypothesized model


Participants and procedure

The data for this study were collected through an online survey on Qualtrics in spring 2022. To recruit participants, the principal investigator reached out to administrators of 72 Italian-speaking OCMHs, all identified by entering mental health-related keywords into the Facebook search bar. OCMHs included in the study could have been moderated by mental health professionals (expert-led) or peer-led. In total, 65% of OCMHs (n = 51) agreed to participate, while 12% (n = 9) declined, and 15% (n = 12) did not respond. Participants were recruited through a video presentation shared on the collaborating OCMHs, including a questionnaire link on Qualtrics. They were rewarded with a financial incentive and access to artistic/poetry videos.

At the beginning of the Qualtrics survey, after providing the informed consent, participants were presented with items related to sociodemographic variables. Subsequently, they were asked to select the specific online mental health group in which they had participated in the previous month from a provided list. Following this, participants were asked to indicate the frequency of their participation in the selected group. Next, they were presented with three items pertaining to misinformation exposure, followed by corresponding questions regarding their agreement with the presented misinformation. Finally, participants completed the depression literacy measure.

The Ethics Committee of the university approved the study design. In total, we collected 493 responses with all variables of the present study filled out. 74 responses were eliminated from the dataset as participants signed “I do not remember” in the OCMHs they were participating in. Thus, they did not specify the OCMH, making it impossible to code an important variable for the study. In the case of double completion of the wave (n = 16 cases), the earliest wave was included in the study. The Ethics Committee of Università della Svizzera italiana approved the study design (CE_2021_10). All methods were performed in accordance with the Declaration of Helsinki. Informed consent was obtained from all participants.


The survey’s items are reported in Additional file 1.

Misinformation exposure

The Italian scale was created ad hoc based on prior studies measuring susceptibility to misinformation. Participants were asked to rate how often they encountered three misinforming claims in the preceding 30 days. The three items were based on the results of a previous content analysis on the groups [31]. The response options were from 1 = “Never” to 5 = “Very often”. Other studies measured exposure similarly by asking participants how frequently they had seen or heard misinforming statements in the past 30 days e.g [70, 71]. We averaged scores on the three misinformation items into an overall index as in [72].

Misinformation agreement

Misinformation agreement was measured by asking the participants their degree of agreement with the same items presented in the Misinformation exposure variable (Likert scale: 1 = completely disagree, 7 = completely agree). In line with other researchers, an index score was created by taking the average of all item ratings e.g [73].

Depression literacy questionnaire

The Depression Literacy Questionnaire [52] assesses mental health literacy specific to depression. The questionnaire consists of 22 items which are true or false. Respondents can answer each item with one of three options – true, false, or don’t know. Each correct response received one point, and a sum score represented the extent of depression literacy (α = 0.74). The Italian version used in the study was validated (unpublished). Translation and back translation were conducted to confirm the scale’s accuracy and appropriateness. A bilingual expert panel composed of an expert in the topic and the two translators was convened to identify and resolve the inadequate expressions/concepts of the translation, as well as any discrepancies between the forward translation and the existing previous versions of the questionnaire. The questionnaire was pilot tested with a small sample of participants (n = 10) to ensure the clarity of the items. We then administered the questionnaire to a larger sample of participants (n = 286) and conducted a series of analyses to assess its reliability and validity. Internal consistency was assessed using Cronbach’s alpha, and we found good internal consistency for the scale (α = 0.74). We also examined the convergent validity of the scale by comparing it to well-established health literacy measured: the Newest Vital Sign (r = .135, p = .022) and the HLS-EU-Q16 (r = .231, p < .001).

Type of OCMHs participation

Participants marked from a list the number of OCMHs they participated in the last month preceding the survey. The first author checked the presence of expert moderators in the OCMHs mentioned during the survey. This led to creating a dummy coded variable in which 0 = peer-led only OCMHs participation, 1 = mixed participation, and 2 = expert-led only participation. Although mixed participation was not highly informative on the type of content moderation participants were exposed to, we decided to retain this level in the variable due to its popularity in the sample (15,9% see section Preliminary results) and excluding this category would have compromised the external validity of the study and limited our ability to capture the diverse experiences of individuals engaged in multiple types of online mental health communities.


Covariates included were frequency of participation, gender and age. The frequency of participation in the groups was assessed using a 4-point scale, ranging from “rarely” (less than once a month) to “very often” (almost every day). The frequency of participation was based on previous studies used to measure the frequency of participation [42]. We controlled for the frequency of participation as misinformation agreement might be impacted by the amount of time they spend on OCMHs platforms [74].

Participants were also asked to indicate the gender they identified to (male, female, other, or prefer not to answer). Due to the scarcity of values in the other categories, the latest was the further dummy coded in 0 = other genders, 1 = male. Furthermore, previous research has shown that identifying with the male gender is associated with greater susceptibility to health misinformation [75] and men have higher reticence to seek professional mental health care services due to mental health stigma [76, 77]. As only two participants selected “other”, and one participant preferred not to respond, we grouped them together with the female responses, to distinguish them from the male category, which we expected to be more distinct. This approach aligns with previous studies [78, 79]. Furthermore, we included age as previous studies showed that older age was associated with less susceptibility to health misinformation [22].

Data analyses

Descriptive statistics were calculated for all measures to assess the normal distribution and detect outliers. Scale reliability was assessed with Cronbach’s alpha for Depression Literacy. However, as its values are quite sensitive to the number of items in the scale, and having two short scales with three items (misinformation exposure and misinformation agreement), according to [80], we report the inter-item correlation range for the scales [81]. recommends an optimal range for the inter-item correlation of.15 to 0.50. A collinearity analysis was carried out to check the prerequisites for a regression analysis, which showed no signs of collinearity (tolerance factor > 0.10 and variance inflation factor < 10) [82]. No multivariate outliers were found according to Mahalanobis distance. SPSS 28.0 was used in our study for statistical analyses. The significance level was set at 0.05. We used PROCESS Model 3 [83] with 95% bias-corrected bootstrap confidence intervals (CIs, with 10,000 bootstrap samples) to test the model of moderated moderation. The effects of gender, age, and frequency were controlled by including them as covariates in the moderation analysis in the PROCESS macro. Following suggestions outlined in Hayes [68], mean centering was conducted on variables that defined products.

Continuous variables were standardized before conducting the model. As the variable Type of OCMHs participation had three levels we used the multi-categorical moderator function in PROCESS. To represent a multi-categorical moderator with k groups, PROCESS creates k − 1 variables and adds them to the model, in addition to k − 1 products representing the interaction [84]. Thus, two dummy variables were created to represent the Type of OCMHs participation and two interaction terms to represent the interaction between the type of OCMHs participation, depression literacy, and misinformation exposure.


Preliminary results

Survey data were obtained from 403 participants. The sample consisted of 333 (82.6%) females, 67 males (16.6%), and 3 participants (0.7%) who responded with others. Participants were members of either a peer-led (n = 126, 31.3%), expert-led (n = 213, 52.9%), or both peer-led and expert-led OCMHs (‘mixed’ category, n = 64; 15,9%). The descriptive statistics, reliability coefficients, and Pearson correlations for the study variables are presented in Table 1.

Table 1 Descriptive statistics and correlations among variables in the whole sample (N = 403)

Participants were members of 64 groups, of which only five were moderated by health professionals (three psychotherapists, one psychologist, one nurse).

Main results

We assessed a moderated-moderation model in which the moderation by depression literacy of the misinformation exposure effect is moderated by the type of OCMHs participation (see Table 2).

Table 2 Coefficients and 95% confidence intervals of moderated-moderation for predicting Misinformation Agreement

Hypothesis 1

Misinformation exposure predicted a larger Misinformation agreement (β = 0.3221, t = 3.8808, p < .001).

Hypothesis 2

Depression literacy predicted lower misinformation agreement (β = -0.2701, t =-3.2933, p = .0011). The positive association between misinformation exposure and misinformation agreement was moderated by depression literacy (β = -0.2179, t = -3.2217, p = .0014).

Hypothesis 3

Participating in expert-led OCMHs predicted lower misinformation agreement (contrast: peer-expert; β = -0.2351, t = -2.0741, p = .0387). The positive association between misinformation exposure and misinformation agreement was moderated by type of OCMHs participation as the contrast interaction peer-expert OCMHs had a significant effect (β = − 0.2322, t =-2.2436, p = .0254).

Research question

The three-way interaction between misinformation exposure, depression literacy, and type of OCMHs participation in misinformation agreement was also significant (ΔR2 = 0.02; F(2,387) = 3.8037, p = .0231). The bootstrap CIs indicated significant effects (p < .001) for 2(of 3) levels of depression literacy: low (β = 0.5338) and medium (β = 0.3182), with an effect significant only in peer-led communities. See Figs. 2 and 3.

Fig. 2
figure 2

Plot representing the three-way interaction

Fig. 3
figure 3

Model with coefficients (contrast: peer-expert; *p < .05 **p < .01)


The present study provides compelling evidence that individuals, when repeatedly exposed to misinformation, are more likely to agree with it (H1), which aligns with the theoretical framework of the illusory truth effect [69]. This first result is concerning as members look for information on online communities to make important health decisions [85]. However, the study also found that individuals with higher levels of depression literacy were better equipped to resist mental health-related misinformation (H2), consistent with prior research [46,47,48]. Therefore, we can conclude that depression literacy is an individual difference that impacts the probability of agreeing (or not) with the misinformation people are exposed to, highlighting the importance of health literacy education to mitigate the effects of misinformation exposure.

The second buffer: the type of OCMHs participation was not only a moderator of the association Misinformation exposure-misinformation agreement (H3), with participants in expert-led OCMHs having lower chances of agreeing to misinformation, but also a moderator of the moderation between depression literacy and misinformation exposure on misinformation agreement (RQ). Specifically, members of peer-led OCMH groups with higher levels of depression literacy were almost as susceptible to misinformation as those in expert-led OCMH groups. Conversely, those with lower levels of depression literacy and lacking the external help of experts in peer-led OCMH groups had the worst outcomes. Future studies should primarily address these vulnerable members of OCMHs, which we identified by combining the analysis of internal (literacy) and external resources (type of participation).

The findings highlight the importance of expert sources to correct health misinformation in the social media [86]. Healthcare professionals, in particular, play a critical role in countering misinformation, as suggested by Schulz & Nakamoto [87], and in line with previous studies, are integral to the success of online communities as they are equipped with the required professional health knowledge to offer trustworthy advice on the causes, prevention, and treatment of disease [88].

One possible strategy to diminish misinformation exposure is to encourage health professionals to administer OCMHs to ensure accurate and trustworthy information and explore possible incentives to promote their involvement or tools that might contribute to their role in the communities, such as automated moderation. Another avenue for future research is to examine the factors that influence individuals to participate in either peer-led or expert-led communities and the perceived differences between these types of communities. Furthermore, gaining insight into the mechanisms underlying the potential positive outcomes associated with participation in expert-led OCMHs can contribute to enhancing the quality of peer-led communities and fostering improved regulation within these virtual spaces.

Furthermore, as suggested by [64], through their participation in OCMHs, health professionals might also inform their professional practice about the misinformation and ambiguous discourses regarding symptoms or treatments. To conclude, while they represented a minority compared to other participants, future studies should better explore members’ participation patterns in both peer-led and expert-led communities (‘mixed’ category). This will provide a more thorough understanding of their experiences and the potential risks they may face.

Limitations and future directions

While our study offers valuable insights into the relationship between exposure to misinformation and mental health misbeliefs, several limitations need to be considered. Firstly, given the cross-sectional nature of the employed data, causative associations between exposure to misinformation and mental health misbeliefs could only be speculated here. Future longitudinal studies could clarify the temporal ordering of variables investigated by examining the cross-lagged paths. Specifically, such studies can help determine whether participants’ beliefs (i.e. agreement with misinformation) precede their volume of exposure to health misinformation or the OCMHs (peer- or expert-led) they decide to participate to and shed light on how the dynamics of misinformation exposure and agreement unfold over time.

Secondly, our study used a self-selection sampling method, limiting our results’ generalizability to larger populations. Future studies should use probability sampling methods to address this limitation.

Thirdly, with respect to limitations regarding the measurement of the variables investigated, our study only measured the volume of misinformation exposure and did not account for the amount of misinformation correction. This is an important variable that should be included in future studies to fully understand the impact of misinformation exposure on mental health misbeliefs.

Furthermore, our use of a depression literacy scale only measured declarative knowledge and may not have captured the full complexity of the literacy construct. Future studies should complement this by investigating the role of critical skills. Another limitation is the potential for recall bias as our measures of misinformation exposure were based on self-reported past experiences. While the specific nature of the questions minimized this bias, future studies should consider using more objective measures of exposure to misinformation. Previous studies have shown that confirmation bias might influence memory and recall, such as that participants in our study could have had a better memory for instances consistent with their prior beliefs [89].

About the frequency measure included as a covariate, it is important to acknowledge that it provides a general indication of participation but lacks specificity, particularly in mixed groups where it does not distinguish whether members attended more peer- or expert-led OCMHs. To obtain a more accurate measurement, future studies should consider inquiring about the frequency of participation for each OCMH or request participants to rank the OCMHs in terms of their level of involvement.

We also acknowledge that, as mental health misinformation exposure and agreement were measured with ad hoc scales, although the items were created from a content analysis, some items were too vague or exhibited a double-barreled issue as there might be differences in exposure/agreement to the different kinds of pharmacological and psychological therapies. However, with the exclusion of item 1, as suggested by reviewers, results were similar to the ones found with all the 3 items. Future studies should address separately the different kinds of treatments. In general, there is a need to improve the operationalization of these variables considering that online people might be exposed to varying types of mental health misinformation.

Lastly, we recognize that there might be heterogeneity in the moderation of OCMHs, even among experts. Future studies should explore the quality (debunking, scientific dissemination, censuring) and quantity of moderation in OCMHs as a continuum rather than a dichotomy.


The study examined responses from 403 members of Facebook OCMHs and found that the vaster the quantity of misinformation regarding mental health the people were exposed to, the more they were likely to believe in it. Therefore, efforts are necessary to reduce exposure to health misinformation and interventions to reduce its impact. Depression literacy was found to have a mitigating effect, suggesting that improving depression literacy could serve as a public health goal to counter the negative impact of misinformation on mental health outcomes. Moreover, the study highlighted the potential role of the type of participation in OCMHs (expert- vs. peer-led) in reducing misbeliefs. It emphasized the importance of health professionals in OCMHs and a resource to incentivize. Findings have significant implications for understanding the complex interplay between the variables investigated and informing interventions and policies to reduce misinformation’s negative impact on mental health, especially for the most vulnerable members identified.

Data Availability

The datasets used and analysed during the current study available from the corresponding author on reasonable request.



Online community for mental health


  1. Hunsaker A, Hargittai E. A review of internet use among older adults. New Media Soc. 2018;20(10):3937–54.

    Google Scholar 

  2. Fox S. The social life of health information [Internet]. Pew Research Center. [cited 2023 Mar 27]. Available from:

  3. Southwell BG, Thorson EA. The prevalence, consequence, and Remedy of Misinformation in Mass Media Systems. J Commun. 2015;65(4):589–95.

    Google Scholar 

  4. Confronting. Health Misinformation.

  5. Wang Y, McKee M, Torbica A, Stuckler D. Systematic literature review on the spread of Health-related misinformation on Social Media. Soc Sci Med. 2019;240:112552.

    PubMed  PubMed Central  Google Scholar 

  6. Bridgman A, Merkley E, Loewen PJ, Owen T, Ruths D, Teichmann L et al. The causes and consequences of COVID-19 misperceptions: Understanding the role of news and social media. Harv Kennedy Sch Misinformation Rev [Internet]. 2020 Jun 18 [cited 2023 Apr 3];1(3). Available from:

  7. Borges do Nascimento IJ, Beatriz Pizarro A, Almeida J, Azzopardi-Muscat N, André Gonçalves M, Björklund M, et al. Infodemics and health misinformation: a systematic review of reviews. Bull World Health Organ. 2022;100(9):544–61.

    PubMed  PubMed Central  Google Scholar 

  8. Swire-Thompson B, DeGutis J, Lazer D. Searching for the Backfire Effect: measurement and design considerations. J Appl Res Mem Cogn. 2020;9(3):286–99.

    PubMed  PubMed Central  Google Scholar 

  9. Chou WYS, Oh A, Klein WMP. Addressing Health-Related misinformation on Social Media. JAMA. 2018;320(23):2417–8.

    PubMed  Google Scholar 

  10. Nan X, Wang Y, Thier K. Health Misinformation. The Routledge Handbook of Health Communication. 3rd ed. Routledge; 2021.

  11. Introne J, Goggins S. Advice reification, learning, and emergent collective intelligence in online health support communities. Comput Hum Behav. 2019;99:205–18.

    Google Scholar 

  12. Southwell BG, Niederdeppe J, Cappella JN, Gaysynsky A, Kelley DE, Oh A, et al. Misinformation as a Misunderstood Challenge to Public Health. Am J Prev Med. 2019;57(2):282–5.

    PubMed  Google Scholar 

  13. Sylvia Chou WY, Gaysynsky A, Cappella JN. Where we go from Here: Health Misinformation on Social Media. Am J Public Health. 2020;110(S3):273–5.

    Google Scholar 

  14. Brashier NM, Marsh EJ. Judging Truth. Annu Rev Psychol. 2020;71(1):499–515.

    PubMed  Google Scholar 

  15. Unkelbach C, Koch A, Silva RR, Garcia-Marques T. Truth by Repetition: explanations and implications. Curr Dir Psychol Sci. 2019;28(3):247–53.

    Google Scholar 

  16. van der Linden S. Misinformation: susceptibility, spread, and interventions to immunize the public. Nat Med. 2022;28(3):460–7.

    PubMed  Google Scholar 

  17. Morgan JC, Cappella JN. The Effect of Repetition on the Perceived Truth of Tobacco-Related Health Misinformation among U.S. adults. J Health Commun. 2023;0(0):1–8.

    Google Scholar 

  18. Ecker UKH, Lewandowsky S, Cook J, Schmid P, Fazio LK, Brashier N, et al. The psychological drivers of misinformation belief and its resistance to correction. Nat Rev Psychol. 2022;1(1):13–29.

    Google Scholar 

  19. Dechêne A, Stahl C, Hansen J, Wänke M. The Truth about the truth: a Meta-Analytic Review of the Truth Effect. Personal Soc Psychol Rev. 2010;14(2):238–57.

    Google Scholar 

  20. Miller JM, Saunders KL, Farhart CE. Conspiracy endorsement as motivated reasoning: the moderating roles of political knowledge and trust. Am J Polit Sci. 2016;60(4):824–44.

    Google Scholar 

  21. Spence A, Spence K. Knowledge mitigates misinformation. Nat Energy. 2021;6(4):329–30.

    Google Scholar 

  22. Nan X, Wang Y, Thier K. Why do people believe health misinformation and who is at risk? A systematic review of individual differences in susceptibility to health misinformation. Soc Sci Med [Internet]. 2022 Dec 1 [cited 2023 Jan 26];314:115398. Available from:

  23. Suarez-Lledo V, Alvarez-Galvez J. Prevalence of Health Misinformation on Social Media: Systematic Review. J Med Internet Res [Internet]. 2021 Jan 20 [cited 2022 May 11];23(1):e17187. Available from:

  24. Campion J. Public mental health: key challenges and opportunities. BJPsych Int. 2018;15(3):51–4.

    PubMed  PubMed Central  Google Scholar 

  25. Prince M, Patel V, Saxena S, Maj M, Maselko J, Phillips MR, et al. No health without mental health. The Lancet. 2007;370(9590):859–77.

    Google Scholar 

  26. Corrigan P. How stigma interferes with mental health care. Am Psychol. 2004;59:614–25.

    PubMed  Google Scholar 

  27. Henderson C, Thornicroft G. Stigma and discrimination in mental illness: time to change. The Lancet. 2009;373(9679):1928–30.

    Google Scholar 

  28. Gaiha SM, Taylor Salisbury T, Koschorke M, Raman U, Petticrew M. Stigma associated with mental health problems among young people in India: a systematic review of magnitude, manifestations and recommendations. BMC Psychiatry. 2020;20(1):538.

    PubMed  PubMed Central  Google Scholar 

  29. Willis E, Royne MB. Online Health Communities and Chronic Disease Self-Management. Health Commun. 2017;32(3):269–78.

    PubMed  Google Scholar 

  30. Smith-Merry J, Goggin G, Campbell A, McKenzie K, Ridout B, Baylosis C. Social connection and Online Engagement: insights from interviews with users of a Mental Health Online Forum. JMIR Ment Health. 2019;6(3):e11084.

    PubMed  PubMed Central  Google Scholar 

  31. Bizzotto N, Morlino S, Schulz PJ. Misinformation in Italian Online Mental Health Communities During the COVID-19 Pandemic: Protocol for a Content Analysis Study. JMIR Res Protoc. 2022; 11(5):e35347.

  32. Pak J, Kim HS, Rhee ES. Characterising social structural and linguistic behaviours of subgroup interactions: a case of online health communities for postpartum depression on Facebook. Int J Web Based Communities. 2020;16(3):225–48.

    Google Scholar 

  33. Park A, Conway M, Chen AT. Examining thematic similarity, difference, and membership in three online mental health communities from reddit: a text mining and visualization approach. Comput Hum Behav. 2018;78:98–112.

    Google Scholar 

  34. Leonardi PM, Nardi BA, Kallinikos J. Materiality and Organizing: Social Interaction in a Technological World. OUP Oxford; 2012.

  35. Myers West S. Censored, suspended, shadowbanned: user interpretations of content moderation on social media platforms. New Media Soc. 2018;20(11):4366–83.

    Google Scholar 

  36. Coulson NS, Shaw RL. Nurturing health-related online support groups: exploring the experiences of patient moderators. Comput Hum Behav. 2013;29(4):1695–701.

    Google Scholar 

  37. Kanthawala S, Peng W. Credibility in Online Health Communities: Effects of Moderator Credentials and endorsement cues. J Media. 2021;2(3):379–96.

    Google Scholar 

  38. Bautista JR, Zhang Y, Gwizdka J. Healthcare professionals’ acts of correcting health misinformation on social media. Int J Med Inf. 2021;148:104375.

    Google Scholar 

  39. Van Oerle S, Lievens A, Mahr D. Value co-creation in online healthcare communities: the impact of patients’ reference frames on cure and care. Psychol Mark. 2018;35(9):629–39.

    Google Scholar 

  40. Deng D, Rogers T, Naslund JA. The role of moderators in facilitating and encouraging peer-to-peer support in an online Mental Health Community: a qualitative exploratory study. J Technol Behav Sci. 2023;8(2):128–39.

    PubMed  PubMed Central  Google Scholar 

  41. Griffiths KM, Calear AL, Banfield M. Systematic review on internet support groups (ISGs) and depression (1): do ISGs reduce depressive symptoms? J Med Internet Res. 2009;11(3):e1270.

    Google Scholar 

  42. Bizzotto N, Marciano L, de Bruijn GJ, Schulz PJ. The empowering role of web-based help seeking on depressive symptoms: systematic review and Meta-analysis. J Med Internet Res. 2023;25(1):e36964.

    PubMed  PubMed Central  Google Scholar 

  43. Bizzotto N, Schulz PJ, De Bruijn GJ. The “Loci” of Misinformation and Its Correction on Peer- and Expert-led Online Communities for Mental Health. Content Analysis. J Med Internet Res. in press

  44. Ratzan SC, Parker RM. Health literacy—identification and response. J Health Commun. 2006;11(8):713–5.

    PubMed  Google Scholar 

  45. Diviani N, Putte B, van den, Giani S, van Weert JC. Low Health Literacy and Evaluation of Online Health Information: A Systematic Review of the Literature. J Med Internet Res [Internet]. 2015 May 7 [cited 2023 Feb 8];17(5):e4018. Available from:

  46. Ashley S, Craft S, Maksl A, Tully M, Vraga EK. Can News Literacy Help Reduce Belief in COVID Misinformation? Mass Commun Soc [Internet]. 2022 Nov 1 [cited 2023 Jan 25];0(0):1–25. Available from:

  47. Xiao X, Su Y. Stumble on information or misinformation? Examining the interplay of incidental news exposure, narcissism, and new media literacy in misinformation engagement. Internet Res [Internet]. 2022 Jan 1 [cited 2023 Feb 15];ahead-of-print(ahead-of-print). Available from:

  48. Wang W, Jacobson S. Effects of health misinformation on misbeliefs: understanding the moderating roles of different types of knowledge. J Inf Commun Ethics Soc. 2022;21(1):76–93.

    CAS  Google Scholar 

  49. cycles T text provides. general information S assumes no liability for the information given being complete or correct D to varying update, Text SCDM up to DDTR in the. Topic: Depression in Italy [Internet]. Statista. [cited 2023 Mar 27]. Available from:

  50. Silvestri C, Carpita B, Cassioli E, Lazzeretti M, Rossi E, Messina V, et al. Prevalence study of mental disorders in an italian region. Preliminary report. BMC Psychiatry. 2023;23(1):12.

    PubMed  PubMed Central  Google Scholar 

  51. World Health Organization. Depression and other common mental disorders: global health estimates [Internet]. World Health Organization; 2017 [cited 2023 Mar 19]. Report No.: WHO/MSD/MER/2017.2. Available from:

  52. Griffiths KM, Christensen H, Jorm AF, Evans K, Groves C. Effect of web-based depression literacy and cognitive-behavioural therapy interventions on stigmatising attitudes to depression: randomised controlled trial. Br J Psychiatry J Ment Sci. 2004;185:342–9.

    Google Scholar 

  53. Ibrahim N, Amit N, Shahar S, Wee LH, Ismail R, Khairuddin R et al. Do depression literacy, mental illness beliefs and stigma influence mental health help-seeking attitude? A cross-sectional study of secondary school and university students from B40 households in Malaysia. BMC Public Health [Internet]. 2019 Jun 13 [cited 2023 Jan 27];19(4):544. Available from:

  54. Rudd RE, Anderson JE, Oppenheimer S, Nath C. Health literacy: an update of medical and public health literature. Rev Adult Learn Lit. 2007;7:175–203.

    Google Scholar 

  55. Gillespie T, Aufderheide P, Carmi E, Gerrard Y, Gorwa R, Matamoros Fernandez A et al. Expanding the Debate about Content Moderation: Scholarly Research Agendas for the Coming Policy Debates [Internet]. Rochester, NY; 2023 [cited 2023 Jun 12]. Available from:

  56. Grygiel J, Brown N. Are social media companies motivated to be good corporate citizens? Examination of the connection between corporate social responsibility and social media safety. Telecommun Policy. 2019;43(5):445–60.

    Google Scholar 

  57. Hartmann IA. A new framework for online content moderation. Comput Law Secur Rev. 2020;36:105376.

    Google Scholar 

  58. Gillespie T. Do not recommend? Reduction as a form of Content Moderation. Soc Media Soc. 2022;8(3):20563051221117550.

    Google Scholar 

  59. Papakyriakopoulos O, Medina Serrano JC, Hegelich S. The spread of COVID-19 conspiracy theories on social me-dia and the effect of content moderation. Harv Kennedy Sch Misinformation Rev [Internet]. 2020 Aug 17 [cited 2023 Jun 12]; Available from:

  60. Saha K, Ernala SK, Dutta S, Sharma E, De Choudhury M. Understanding moderation in Online Mental Health Communities. In: Meiselwitz G, editor. Social Computing and Social Media participation, user experience, consumer experience, and applications of Social Computing. Cham: Springer International Publishing; 2020. pp. 87–107. (Lecture Notes in Computer Science).

    Google Scholar 

  61. Wadden D, August T, Li Q, Althoff T. The effect of moderation on online mental health conversations. ICWSM. 2021.

  62. Hartzler A, Pratt W. Managing the personal side of health: how patient expertise differs from the expertise of clinicians. J Med Internet Res. 2011;13(3):e62.

    PubMed  PubMed Central  Google Scholar 

  63. Eysenbach G, Powell J, Kuss O, Sa ER. Empirical studies assessing the quality of Health Information for Consumers on the world wide WebA systematic review. JAMA. 2002;287(20):2691–700.

    PubMed  Google Scholar 

  64. Petrič G, Atanasova S, Kamin T. Ill Literates or Illiterates? Investigating the eHealth literacy of users of Online Health Communities. J Med Internet Res. 2017;19(10):e7372.

    Google Scholar 

  65. NW, St 1615L, Washington S. 800, Inquiries D 20036 U 419 4300 | M 857 8562 | F 419 4372 | M. Since 2014-15, TikTok has arisen; Facebook usage has dropped; Instagram, Snapchat have grown [Internet]. Pew Research Center: Internet, Science & Tech. [cited 2023 Jun 12]. Available from:

  66. Italy. : leading social media platforms 2022 [Internet]. Statista. [cited 2023 Jun 12]. Available from:

  67. Bayer JB, Triệu P, Ellison NB. Social media elements, Ecologies, and Effects. Annu Rev Psychol. 2020;71(1):471–97.

    PubMed  Google Scholar 

  68. Hayes AF. Introduction to Mediation, Moderation, and conditional process analysis, Second Edition: a regression-based Approach. Guilford Publications; 2017. p. 714.

  69. Henderson EL, Westwood SJ, Simons DJ. A reproducible systematic map of research on the illusory truth effect. Psychon Bull Rev. 2022;29(3):1065–88.

    PubMed  Google Scholar 

  70. Hwang Y, Jeong SH. Misinformation exposure and Acceptance: the role of information seeking and Processing. Health Commun. 2023;38(3):585–93.

    PubMed  Google Scholar 

  71. Yoo W, Oh SH, Choi DH. COVID-19, Digital Media, and Health| exposure to COVID-19 misinformation across instant messaging apps: moderating roles of News Media and Interpersonal Communication. Int J Commun. 2023;17(0):23.

    Google Scholar 

  72. Roozenbeek J, Schneider CR, Dryhurst S, Kerr J, Freeman ALJ, Recchia G, et al. Susceptibility to misinformation about COVID-19 around the world. R Soc Open Sci. 2020;7(10):201199.

    CAS  PubMed  PubMed Central  Google Scholar 

  73. Maertens R, Roozenbeek J, Basol M, van der Linden S. Long-term effectiveness of inoculation against misinformation: three longitudinal experiments. J Exp Psychol Appl. 2021;27:1–16.

    PubMed  Google Scholar 

  74. Enders AM, Uscinski JE, Seelig MI, Klofstad CA, Wuchty S, Funchion JR et al. The Relationship Between Social Media Use and Beliefs in Conspiracy Theories and Misinformation. Polit Behav [Internet]. 2021 Jul 7 [cited 2023 Mar 16]; Available from:

  75. Filkuková P, Ayton P, Rand K, Langguth J. What Should I Trust? Individual Differences in Attitudes to Conflicting Information and Misinformation on COVID-19. Front Psychol [Internet]. 2021 [cited 2023 Jun 12];12. Available from:

  76. Oliffe JL, Phillips MJ. Men, depression and masculinities: a review and recommendations. J Mens Health. 2008;5(3):194–202.

    Google Scholar 

  77. Chatmon BN. Males and Mental Health Stigma. Am J Mens Health. 2020;14(4):1557988320949322.

    PubMed  PubMed Central  Google Scholar 

  78. Lougheed JP, Keskin G, Morgan S. The hazards of Daily Stressors: comparing the Experiences of sexual and gender minority young adults to cisgender heterosexual young adults during the COVID-19 pandemic. Collabra Psychol. 2023;9(1):73649.

    Google Scholar 

  79. Wang Y, Chen X, Yang Y, Cui Y, Xu R. Risk perception and resource scarcity in food procurement during the early outbreak of COVID-19. Public Health. 2021;195:152–7.

    CAS  PubMed  Google Scholar 

  80. Pallant J. SPSS Survival Manual: a step by step guide to data analysis using IBM SPSS. 7th ed. London: Routledge; 2020. p. 378.

    Google Scholar 

  81. Clark LA, Watson D. Constructing validity: basic issues in objective scale development. Psychol Assess. 1995;7:309–19.

    Google Scholar 

  82. Hair JF, Black WC, Babin BJ, Anderson RE. Multivariate Data Analysis. Pearson Education Limited; 2013. p. 734.

  83. PROCESS macro for SPSS and SAS [Internet]. The PROCESS macro for SPSS, SAS, and R. [cited 2023 Feb 17]. Available from:

  84. Hayes AF, Montoya AK. A tutorial on testing, visualizing, and probing an Interaction Involving a Multicategorical Variable in Linear Regression Analysis. Commun Methods Meas. 2017;11(1):1–30.

    Google Scholar 

  85. Yan Z, Wang T, Chen Y, Zhang H. Knowledge sharing in online health communities: a social exchange theory perspective. Inf Manage. 2016;53(5):643–53.

    Google Scholar 

  86. Vraga EK, Bode L. Using Expert sources to correct Health Misinformation in Social Media. Sci Commun. 2017;39(5):621–45.

    Google Scholar 

  87. Schulz PJ, Nakamoto K. The perils of misinformation: when health literacy goes awry. Nat Rev Nephrol. 2022;18(3):135–136.

  88. Zhang X, Liu S, Deng Z, Chen X. Knowledge sharing motivations in online health communities: a comparative study of health professionals and normal users. Comput Hum Behav. 2017;75:797–810.

    Google Scholar 

  89. Frost P, Casey B, Griffin K, Raymundo L, Farrell C, Carrigan R. The influence of confirmation Bias on memory and source monitoring. J Gen Psychol. 2015;142(4):238–52.

    PubMed  Google Scholar 

Download references


The authors wish to thank the admins of the OCMHs for their invaluable support for this study.


The Swiss National Science Foundation supported this work under grant 200396.

Author information

Authors and Affiliations



N.B. wrote the main manuscript text. All authors reviewed the manuscript.

Corresponding author

Correspondence to Peter Johannes Schulz.

Ethics declarations

Ethics approval and consent to participate

The Ethics Committee of Università della Svizzera italiana approved the study design (CE_2021_10). All methods were performed in accordance with the Declaration of Helsinki. Informed consent was obtained from all subjects.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary Material 1

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Bizzotto, N., de Bruijn, GJ. & Schulz, P.J. Buffering against exposure to mental health misinformation in online communities on Facebook: the interplay of depression literacy and expert moderation. BMC Public Health 23, 1577 (2023).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: