Abstract

The ubiquity of social media use and the digital data traces it produces has triggered a potential methodological shift in the psychological sciences away from traditional, laboratory-based experimentation. The hope is that, by using computational social science methods to analyse large-scale observational data from social media, human behaviour can be studied with greater statistical power and ecological validity. However, current standards of null hypothesis significance testing and correlational statistics seem ill-suited to markedly noisy, high-dimensional social media datasets. We explore this point by probing the moral contagion phenomenon, whereby the use of moral-emotional language increases the probability of message spread. Through out-of-sample prediction, model comparisons and specification curve analyses, we find that the moral contagion model performs no better than an implausible XYZ contagion model. This highlights the risks of using purely correlational evidence from large observational datasets and sounds a cautionary note for psychology’s merge with big data.

Main

The digitalization of society raises many substantive questions (see, for example, refs. 1,2,3). At the same time, however, it provides unmistakable methodological opportunities for social science research. For all of the interactions that take place online, such as communications between social media users, digital data traces are left behind. Not only do these data traces capture naturalistic behaviours, but the sheer scale and variety of such data means theories of, for example, collective reasoning and opinion dynamics can now be readily observed quantitatively ‘in the wild’4,5,6. With more social science domains recognizing this trend and utilizing large-scale social media data to test theories of human behaviour7, there is a pressing need for researchers to better understand the strength of inferences made from such data, and for methodological standards to be queried.

One high-profile study that combines social media data and psychological theory recently presented findings of a ‘moral contagion’ effect8. In the study, Brady et al.8 apply a dictionary-based text analysis procedure to quantify moral-emotional language in hundreds of thousands of tweets capturing the naturally occurring communications of Twitter users. By then fitting a regression model and performing a series of robustness checks, they show that the mere presence of moral-emotional words increases messages’ retweet counts by a factor of 20%, regardless of the messages’ informational quality8. The implications of this moral contagion phenomenon, where exposure to moral emotions shapes the diffusion of information, are undoubtedly significant. Invoking morality in reasoning has previously been shown to harden existing belief structures, delegitimize authority and, in extreme cases, dehumanize opposing perspectives9,10. While injections of moral reasoning into discourse can be beneficial—providing shared identities and guiding ethical behaviour—the introduction of unnecessary moralization and its emotional underpinnings may jeopardize rational debate. It is for this reason that moral justifications carry weight in some domains but not others. For example, loading an argument with moral-emotional language might be an effective strategy in a debate over social policy and human rights, yet that same strategy is likely to be penalized in an argument over mathematics. However, if moral contagion is as widespread and domain-general as Brady et al.8 suggest, then it seems plausible that sentiments about where moralization is appropriate are changing. This also suggests that we are susceptible to new forms of political persuasion online. As Brady et al.8 conclude, “it seems likely that politicians, community leaders, and organizers of social movements express moral emotions…in an effort to increase message exposure and to influence perceived norms within social networks” (p. 7316). Beyond this substantive contribution, the authors also recognize the methodological implications of their study8, because “in comparison with laboratory-based studies, the social network approach offers much greater ecological validity” (p. 7317).

The work of Brady et al.8 is one example of what is an ongoing methodological shift in the psychological sciences (also see, for example, refs. 11,12,13), whereby statistical analyses of large-scale digital data traces, namely social media data, are used as the basis for inferences about human emotions, behaviours and motivations. But digital data traces produced by social media users are inherently noisy and high-dimensional. In contrast to the ‘custom-made’ data generated via controlled experimentation, material harvested from online platforms is usually not created with research in mind14. Social media data can be ambiguous, confounded by proprietary algorithms and restricted access, and unrepresentative of wider populations, which may limit the generalizability of findings between platforms and between online and offline populations14,15,16. These documented observations may be less problematic if one’s research objective concerns itself only with behaviour on a given platform; however, in the absence of agreed-upon methodological standards for handling social media data, the space for ‘researcher degrees of freedom’17 is particularly vast. This means that conclusions from analyses of observational social media data alone may face deeper issues, insofar as they are intended to teach us something about real human behaviour.

Here, we probe the finding of moral contagion, illustrating possible methodological pitfalls that might be encountered when standard practices of null hypothesis significance testing are applied to large-scale social media datasets. How robust is correlational evidence from large-scale observational data? What inferences and generalizations can be made from such evidence? Answering these questions seems crucial for psychology’s merge with big data.

Results

Out-of-sample prediction

The diffusion of information via social networks has been likened to a biological pathogen, spreading from person to person through direct contact. For a behaviour, psychological state or other condition to qualify as a simple social contagion, the probability of the condition being adopted by an individual should increase monotonically with the number of times that individual is exposed to said condition18. In the case of moral contagion, moral-emotional words (for example, ‘kill’, ‘protest’ and ‘compassion’) are considered to be the ‘contagious’ cue because their presence is presumed to be a central factor in an individual’s decision to retweet (or diffuse) the message in which it is included. Based on this logic, moral contagion should be present in other corpora of tweets pertaining to contentious, politicized topics. To test this proposal, we recreated Brady et al.’s 8 methodology and applied it to other Twitter corpora spanning a variety of socio-political issues and events.

Using the dictionary-based text analysis of Brady et al.8 to quantify distinctly moral, distinctly emotional and moral-emotional language (see Supplementary Sect. 1.2 for details), we tested the influence of language use on message diffusion across six corpora of tweets that capture the naturally occurring communications among users (four of these corpora were pre-existing). Each corpus pertained to a specific issue or event: the COVID-19 pandemic (n = 172,697), the #MeToo movement19 (n = 151,035), the #MuellerReport investigation (n = 39,068), the #WomensMarch protest20 (n = 3,778), the announcement of the 2016 EU Referendum result in the United Kingdom (Brexit)21 (n = 5,660) and the 2016 US Presidential Election (n = 8,233; this corpus contained only ‘viral’ tweets that received more than 1,000 retweets)22 (see Supplementary Sect. 1.1 for details on each corpus). Diffusion was measured as the sum of a message’s retweet count as captured in the metadata and the number of times that message’s text appeared in a corpus. Identical messages were then collapsed into a single observation with other relevant metadata from the earliest posting (for example, the number of followers a message poster has; whether the post included uniform resource locators (URLs), an image or video media). This approach avoids penalizing retweet chains, which are important indicators of diffusion on Twitter, while also accounting for unconventional retweets where a user copies and pastes someone’s message rather than clicking the retweet button. With diffusion as our dependent variable and the three language measures as predictors, we then followed Brady et al.8 in fitting a negative binomial regression model to each dataset (referred to as the ‘main moral contagion model’). The presence of contagion was determined by exponentiating the regression coefficients of each predictor (distinctly emotional, distinctly moral and moral-emotional language) to generate incidence rate ratios (IRRs), the most central measure being moral-emotional language’s IRR. Note that, as a ratio measure, IRRs greater than 1.00 signify a positive contagion effect (for example, an IRR of 1.10 suggests a 10% increase in diffusion), and vice versa.

Prior to analysing our corpora, we checked our model specifications by reanalysing Brady et al.’s8 cleaned data, which they have made available online. Across the three corpora comprising 313,002 analysable tweets spanning three topics (same-sex marriage, n = 29,060; gun control, n = 48,394; climate change, n = 235,548), our analysis reproduced their findings. Moral-emotional language was significantly associated with an increase in retweets in each corpus when covariates were controlled for (same-sex marriage, IRR 1.17, P < 0.001, 95% CI 1.09–1.27; gun control, IRR 1.19, P < 0.001, 95% CI 1.14–1.23; climate change, IRR 1.24, P < 0.001, 95% CI 1.22–1.27), and in two out of three corpora when covariates were not controlled for (same-sex marriage, IRR 1.08, P = 0.059, 95% CI 0.99–1.18; gun control, IRR 1.36, P < 0.001, 95% CI 1.30–1.42; climate change, IRR 1.15, P < 0.001, 95% CI 1.12–1.17). However, these results did not consistently generalize across the six corpora we analysed.

Taking Brady et al.’s main moral contagion model, as well as the nested single-variable model in which only moral-emotional language is used as a predictor, we found moral contagion to be present in only two of six corpora before controlling for covariates: COVID-19 tweets (IRR 1.15, P < 0.001, 95% CI 1.11–1.18) and #MuellerReport tweets (IRR 1.28, P < 0.001, 95% CI 1.16–1.42). In the four pre-existing corpora, moral-emotional language either had no significant relationship with message diffusion or had a negative effect where moral-emotional language predicted a decrease in diffusion (Table 1). While we could not control for the same covariates as Brady et al.8 and were therefore unable to provide direct replications in the four pre-existing corpora due to missing metadata, we did so in the COVID-19 and #MuellerReport corpora (we do this to aid comparison with Brady et al.’s8 original results; however, we strongly caution against basing one’s interpretation of these results on covariates; see Covariates, outliers and the analytical multiverse section). Once Brady et al.’s8 chosen covariates were controlled for in the regression model to provide a direct replication of the original analysis, the significant association between moral-emotional words and message diffusion remained in the #MuellerReport tweets (IRR 1.27, P < 0.001, 95% CI 1.16–1.40), but no statistically significant relationship was observed in the COVID-19 tweets (IRR 1.01, P = 0.320, 95% CI 0.99–1.04).

The limits of correlational data

The inconsistent results of out-of-sample prediction tests using procedures similar to those applied by Brady et al.8 point to the limitations of purely correlational data. The inherent difficulty of distinguishing true causal contagion from confounding network homophily has been noted in detail elsewhere (see, for example, refs. 23,24). However, large observational datasets carry even more fundamental risks of spurious correlation and endogeneity. To demonstrate this, we conducted a follow-up analysis in the spirit of Hilbig25, who cautioned against correlational data as a sole source of evidence for heuristic use in judgement and decision-making tasks. Specifically, we created an absurd factor for illustrative purposes, what we call XYZ contagion, and tested whether the number of X’s, Y’s and Z’s included in messages’ text predicted diffusion (note that we were unable to test for XYZ contagion in Brady et al.’s8 original data because their raw data did not include metadata retweet counts, which meant that our analysis scripts could not be properly applied).

Our analysis found XYZ contagion to be present in four of our six corpora such that the presence of the letters X, Y and Z predicted an increase in message diffusion: COVID-19 tweets (IRR 1.08, P < 0.001, 95% CI 1.07–1.08), #MeToo tweets (IRR 1.13, P < 0.001, 95% CI 1.12–1.15), #MuellerReport tweets (IRR 1.12, P < 0.001, 95% CI 1.10–1.14) and the 2016 US Election tweets (IRR 1.01, P = 0.030, 95% CI 1.00–1.03). While there was no positive relationship between the presence of X, Y and Z and message diffusion in the #WomensMarch and Post-Brexit tweets, the finding that XYZ contagion passes a key test of robustness, viz. out-of-sample prediction, demonstrates the potential of large-scale social media datasets to contain spurious correlations (Table 1; also see Supplementary Sect. 2.3 for a bootstrap resampling analysis).

In addition, we calculated Akaike information criteria (AIC) as measures of model adequacy and found that our model of XYZ contagion actually outperforms the main, multi-variable moral contagion model in two of the six corpora (Table 1). We further tested the XYZ contagion model against the single-variable moral contagion model such that the predictive value of the count of letters X, Y and Z was compared with the count of moral-emotional words in isolation. This analysis revealed that the count of letters X, Y, and Z was in fact a better predictor of message diffusion than moral-emotional words in five out of six corpora, despite being nonsensical (Table 1).

Covariates, outliers and the analytical multiverse

Out-of-sample prediction tests and model comparisons demonstrate how social media datasets may be susceptible to unfounded correlations. However, we need to consider the influence of outliers and covariates in more detail, which are indeed sensible and widely recognized checks that can and have been put in place to guard against spurious results. But as we show next, in the context of social media data, neither of these are sufficient to solve the problems identified here, facing both methodological and conceptual limitations.

Regarding outliers, the problem is that social media data are a typical case of fat-tailed distribution, and it is unclear how ‘outlier’ should be defined. The prevalence of extreme values (for example, a tweet garnering 100,000 retweets when the median is 0) is likely a constitutive feature of the dataset, rather than a bug or error to be neglected. Consequently, decisions on outliers are seemingly arbitrary. For example, consider a traditional psychology experiment measuring reaction times in the laboratory. Outliers in this case are readily identifiable: A reaction time that is ten times the mean indicates that a participant was not paying attention or had not read the instructions, or that the data were entered incorrectly. Yet, in the domain of social media, there is no such judgement that can be made. That a message may be retweeted 0, 1 or 100,000 times is in fact an intrinsic part of the paradigm. What does it mean if, in a study of message sharing, the top 10 or 100 most shared messages determine what statistical results are retrieved from a corpus of hundreds of thousands of messages? Are these observations to be excluded, or are they indicative of a recipe for going viral?

Covariates might be considered even more important. Indeed, there is a wide range of potential covariates that plague social media data, relating to both the content of messages and the accounts of message posters. Specifically relating to Twitter, it has previously been shown that the presence of hashtags and URLs in a message, the number of followers and followees a message poster has and the age of the message poster’s account all influence retweet rates26. There are also questions around the potential need to account for the influence of automated and semi-automated bots16,27,28. Despite existing literature highlighting these covariates, the controls that researchers put in place are often inconsistent, even when the hypotheses in question are relatively similar. For example, consider three studies investigating the role of emotion in message sharing on Twitter: Stieglitz and Dang-Xuan29 control for the number of hashtags a tweet contains, the presence of URLs, the number of followers a message poster has and the number of tweets a user has posted during the sampling period; Ferrara and Yang30 excluded tweets containing URLs or media (that is, a photo or video); and Brady et al.8 control for the number of followers the message poster has, whether media or URLs are present in a tweet and whether the message poster is ‘verified’ (a status indicating that the user is a celebrity or public figure). Not only do these studies identify different covariates, but they also control for them in different ways. For instance, where Ferrara and Yang30 excluded tweets containing URLs and media, Brady et al.8 input these covariates as binary variables in a regression. While each study’s controls are certainly defensible, this points to another problem: any given set of controls will not be exhaustive and there is no agreed-upon standard for what controls must be applied to separate a publishable finding from a coincidental statistic; and even more fundamentally, against what ground truth could these methodological practices be evaluated?

Taken together, the ambiguity surrounding outliers and covariates highlights the increased ‘researcher degrees of freedom’17 in analyses of social media data. That is, researchers must make many arbitrary analytical decisions when collecting, processing and analysing the data. While this is not unique to social media data or any type of digital trace data, it may be especially consequential in this context. To investigate how decisions on covariates and outliers influence the moral contagion and XYZ contagion results, we conducted specification curve analyses (SCA)31 on our three largest corpora (COVID-19, #MeToo and #MuellerReport). In short, SCA is a way to make analytic flexibility transparent by running all justifiable model specifications (for example, what covariates to control for, what data subsets to analyse, what independent variable to assess, etc.), and then making joint inferences across the results of all these specifications31. SCA is closely related to the concept of a ‘garden of forking paths’32 and ‘multiverse analysis’33, and serves to clarify the fragility or robustness of statistical findings by identifying which analytical choices they hinge on.

For our SCA, we consider the results of negative binomial regression specifications with either the number of X’s, Y’s and Z’s or the number of moral-emotional words in a tweet predicting diffusion, with or without controlling for covariates, and with or without the removal of (arbitrary) increments of outliers (the tweets with the top 10, 100 or 1,000 diffusion counts). The covariates we consider are the number of distinctly moral words, the number of distinctly emotional words, the number of characters in a tweet, the number of followers a message poster has, whether the message poster’s account is verified and whether media, URLs and hashtags are present (binary). Because the #MeToo corpus is a pre-existing dataset that was not collected by the authors of the present study, not all of the relevant metadata are included and only some of the covariates could be considered. Figure 1 displays the outcome (unstandardized regression coefficient) of each model specification (x axis) when fitted to each corpus as three, vertically aligned points corresponding to the independent variable, covariates and outliers accounted for (y axis). We then plot these outcomes as specification curves in Fig. 2, visualizing how negative, positive and non-significant moral contagion effects can be retrieved, depending on the chosen corpus and model specification (also see Supplementary Figs. 25 for SCA applied to Brady et al.’s original corpora). The specification curves also allow for comparative evaluations between moral contagion and XYZ contagion. Namely, we observe that, while the median regression coefficient across model specifications with moral-emotional words as the independent variable is positive in the COVID-19 (n = 40, median B = 0.18, s.d. 0.08) and #MuellerReport corpora (n = 39, median B = 0.10, s.d. 0.13), it is negative in the #MeToo corpus (n = 28, median B = −0.02, s.d. 0.08). Meanwhile, the median regression coefficient across model specifications with the number of X’s, Y’s and Z’s as the independent variable is positive in all three corpora (COVID-19, n = 39, median B = 0.07, s.d. 0.05; #MeToo, n = 28, median B = 0.04, s.d. 0.06; #MuellerReport, n = 39, median B = 0.05, s.d. 0.05). This could be taken to suggest that the XYZ contagion effect is, if anything, more stable than the moral contagion effect across theoretically justifiable model specifications in the three corpora addressed here. Of course, we strongly doubt that the letters X, Y and Z play a central role in shaping the diffusion of information on Twitter. What our analyses show, however, is that the evidence of moral contagion provided by Brady et al.8 seems to be virtually indistinguishable from our atheoretical XYZ contagion effect, regardless of whether it is framed as a causal or correlational effect.

Fig. 1: Qualitative results of SCA.
figure1

a, COVID-19 corpus. b, #MeToo corpus. c, #MuellerReport corpus. Each possible model specification (x axis) is represented by three vertically aligned points corresponding to the outliers removed and covariates and independent variable included in the negative binomial regression equation (y axis). Red indicates a significant (P < 0.05) negative regression coefficient, grey indicates a non-significant coefficient and blue indicates a significant positive coefficient. There are fewer specifications (56) in the #MeToo SCA (b) because metadata on some covariates were absent. Of the 80 possible specifications for the COVID-19 and #MuellerReport data, one specification was excluded from the COVID-19 SCA and two from the #MuellerReport SCA because these algorithms did not converge.

Full size image
Fig. 2: Specification curves for moral contagion and XYZ contagion effects.
figure2

a, COVID-19 corpus (moral contagion, n = 40, median B = 0.18, s.d. 0.08; XYZ contagion, n = 39, median B = 0.07, s.d. 0.05). b, #MeToo corpus (moral contagion, n = 28, median B = −0.02, s.d. 0.08; XYZ contagion, n = 28, median B = 0.04, s.d. 0.06). c, #MuellerReport corpus (moral contagion, n = 39, median B = 0.10, s.d. 0.13; XYZ contagion, n = 39, median B = 0.05, s.d. 0.05). Each model specification (x axis) is represented by a single point indicating the resulting unstandardized regression coefficient and vertical bars indicating 95% confidence intervals (y axis). Red indicates a significant (P < 0.05) negative regression coefficient, grey indicates a non-significant coefficient and blue indicates a significant positive coefficient. There are fewer specifications in the #MeToo corpus (b) because metadata on some covariates was not recorded. Two specifications in the #MuellerReport corpus (c) and one in the COVID-19 corpus (a) are excluded because the algorithm did not converge.

Full size image
Read & annotate articles