https://academic.oup.com/pnasnexus/article/2/12/pgad385/7457939

Skip to Main Content 
Advertisement
Oxford Academic
Journals
Books

  * Search Menu
  * Information
  * Account
  * Menu
  * Information
  * Account

Navbar Search Filter [PNAS Nexus                   ] Mobile Enter
search term [                    ] Search

  * Issues
  * Advance articles
  * Subject
      + Biological, Health, and Medical Sciences
          o Browse content in Biological, Health, and Medical
            Sciences
          o Administration Of Health Services, Education, and
            Research
          o Agricultural Sciences
          o Allied Health Professions
          o Anesthesiology
          o Anthropology
          o Anthropology (Biological, Health, and Medical Sciences)
          o Applied Biological Sciences
          o Biochemistry
          o Biophysics and Computational Biology (Biological, Health,
            and Medical Sciences)
          o Biostatistics
          o Cell Biology
          o Dermatology
          o Developmental Biology
          o Ecology
          o Environmental Sciences (Biological, Health, and Medical
            Sciences)
          o Evolution
          o Genetics
          o Immunology and Inflammation
          o Internal Medicine
          o Medical Microbiology
          o Medical Sciences
          o Microbiology
          o Neurology
          o Neuroscience
          o Nutrition
          o Obstetrics and Gynecology
          o Ophthalmology
          o Pediatrics
          o Pharmacology
          o Pharmacy
          o Physiology
          o Plant Biology
          o Population Biology
          o Psychiatry
          o Psychological and Cognitive Sciences (Biological, Health,
            and Medical Sciences)
          o Public Health and Epidemiology
          o Radiation Oncology
          o Radiology
          o Rehabilitation
          o Surgery
          o Sustainability Science (Biological, Health, and Medical
            Sciences)
          o Systems Biology
      + Physical Sciences and Engineering
          o Browse content in Physical Sciences and Engineering
          o Aerospace Engineering
          o Applied Mathematics
          o Applied Physical Sciences
          o Bioengineering
          o Biophysics and Computational Biology (Physical Sciences
            and Engineering)
          o Chemical Engineering
          o Chemistry
          o Civil and Environmental Engineering
          o Computer Sciences
          o Computer Science and Engineering
          o Earth Resources Engineering
          o Earth, Atmospheric, and Planetary Sciences
          o Electric Power and Energy Systems Engineering
          o Electronics, Communications and Information Systems
            Engineering
          o Engineering
          o Environmental Sciences (Physical Sciences and
            Engineering)
          o Materials Engineering
          o Mathematics
          o Mechanical Engineering
          o Physics
          o Statistics
          o Sustainability Science (Physical Sciences and
            Engineering)
      + Social and Political Sciences
          o Browse content in Social and Political Sciences
          o Anthropology (Social and Political Sciences)
          o Demography
          o Economic Sciences
          o Environmental Sciences (Social and Political Sciences)
          o Political Sciences
          o Psychological and Cognitive Sciences (Social and
            Political Sciences)
          o Social Sciences
          o Sustainability Science (Social and Political Sciences)
  * Submit
      + Author guidelines
      + Submission site
      + Open access policy
      + Self-archiving policy
      + Why submit to PNAS Nexus
      + The PNAS portfolio
      + For reviewers
  * Alerts
  * About
      + About PNAS Nexus
      + About National Academy of Sciences
      + Editorial Board
  * Journals on Oxford Academic
  * Books on Oxford Academic

NAS Journals
PNAS Nexus
National Academy of Sciences

  * Issues
  * Advance articles
  * Subject
    All Subject Expand Expand
    Biological, Health, and Medical Sciences
    Administration Of Health Services, Education, and Research
    Agricultural Sciences
    Allied Health Professions
    Anesthesiology
    Anthropology
    Anthropology (Biological, Health, and Medical Sciences)
    Applied Biological Sciences
    Biochemistry
    Biophysics and Computational Biology (Biological, Health, and
    Medical Sciences)
    Biostatistics
    Cell Biology
    Dermatology
    Developmental Biology
    Ecology
    Environmental Sciences (Biological, Health, and Medical Sciences)
    Evolution
    Genetics
    Immunology and Inflammation
    Internal Medicine
    Medical Microbiology
    Medical Sciences
    Microbiology
    Neurology
    Neuroscience
    Nutrition
    Obstetrics and Gynecology
    Ophthalmology
    Pediatrics
    Pharmacology
    Pharmacy
    Physiology
    Plant Biology
    Population Biology
    Psychiatry
    Psychological and Cognitive Sciences (Biological, Health, and
    Medical Sciences)
    Public Health and Epidemiology
    Radiation Oncology
    Radiology
    Rehabilitation
    Surgery
    Sustainability Science (Biological, Health, and Medical Sciences)
    Systems Biology
    Physical Sciences and Engineering
    Aerospace Engineering
    Applied Mathematics
    Applied Physical Sciences
    Bioengineering
    Biophysics and Computational Biology (Physical Sciences and
    Engineering)
    Chemical Engineering
    Chemistry
    Civil and Environmental Engineering
    Computer Sciences
    Computer Science and Engineering
    Earth Resources Engineering
    Earth, Atmospheric, and Planetary Sciences
    Electric Power and Energy Systems Engineering
    Electronics, Communications and Information Systems Engineering
    Engineering
    Environmental Sciences (Physical Sciences and Engineering)
    Materials Engineering
    Mathematics
    Mechanical Engineering
    Physics
    Statistics
    Sustainability Science (Physical Sciences and Engineering)
    Social and Political Sciences
    Anthropology (Social and Political Sciences)
    Demography
    Economic Sciences
    Environmental Sciences (Social and Political Sciences)
    Political Sciences
    Psychological and Cognitive Sciences (Social and Political
    Sciences)
    Social Sciences
    Sustainability Science (Social and Political Sciences)
    Browse all content Browse content in
  * Submit
      + Author guidelines
      + Submission site
      + Open access policy
      + Self-archiving policy
      + Why submit to PNAS Nexus
      + The PNAS portfolio
      + For reviewers
  * Alerts
  * About
      + About PNAS Nexus
      + About National Academy of Sciences
      + Editorial Board

Close
Navbar Search Filter [PNAS Nexus                   ] Enter search
term [                    ] Search
Advanced Search
Search Menu
Article Navigation
Close mobile search navigation
Article Navigation
 
Issue Cover
Volume 2
Issue 12
December 2023
(In Progress)

Article Contents

  * Abstract
  * Introduction
  * Results
  * Discussion
  * Methods and materials
  * Acknowledgments
  * Supplementary Material
  * Funding
  * Author Contributions
  * Previous Presentation
  * Preprints
  * Data Availability
  * References

  * < Previous
  * Next >

Article Navigation
Article Navigation
Journal Article

Toxic comments are associated with reduced activity of volunteer
editors on Wikipedia

Ivan Smirnov,
Ivan Smirnov  
Graduate Research School, University of Technology Sydney
,
15 Broadway
, Sydney 2007,
Australia
To whom correspondence should be addressed: Email:
ivan.smirnov@uts.edu.au
ORCID logo  https://orcid.org/0000-0002-8347-6703
Search for other works by this author on:
Oxford Academic
Google Scholar
Camelia Oprea,
Camelia Oprea
Department of Computer Science, RWTH Aachen University
,
Ahornstrasse 55
, Aachen 52074,
Germany
Search for other works by this author on:
Oxford Academic
Google Scholar
Markus Strohmaier
Markus Strohmaier
Business School, University of Mannheim
,
L 15 1-6
, Mannheim 68161,
Germany
GESIS--Leibniz Institute for the Social Sciences
,
Unter Sachsenhausen 6-8
, Koln 50667,
Germany
Complexity Science Hub Vienna
,
Josefstaedter Strasse 39
, Vienna 1080,
Austria
Search for other works by this author on:
Oxford Academic
Google Scholar

Competing Interest: The authors declare no competing interest.

Author Notes
PNAS Nexus, Volume 2, Issue 12, December 2023, pgad385, https://
doi.org/10.1093/pnasnexus/pgad385
Published:
05 December 2023
Article history
Received:
30 June 2023
Accepted:
30 October 2023
Corrected and typeset:
05 December 2023
Published:
05 December 2023

  * pdfPDF
  * Split View
  *  
    Views
      + Article contents
      + Figures & tables
      + Video
      + Audio
      + Supplementary Data
  * Cite
   
    Cite

    Ivan Smirnov, Camelia Oprea, Markus Strohmaier, Toxic comments
    are associated with reduced activity of volunteer editors on
    Wikipedia, PNAS Nexus, Volume 2, Issue 12, December 2023,
    pgad385, https://doi.org/10.1093/pnasnexus/pgad385

    Select Format [Select format                  ] Download citation
    Close
  * Permissions Icon Permissions
  * Share Icon Share
      + Facebook
      + Twitter
      + LinkedIn
      + Email

Navbar Search Filter [PNAS Nexus                   ] Mobile Enter
search term [                    ] Search
Close
Navbar Search Filter [PNAS Nexus                   ] Enter search
term [                    ] Search
Advanced Search
Search Menu

Abstract

Wikipedia is one of the most successful collaborative projects in
history. It is the largest encyclopedia ever created, with millions
of users worldwide relying on it as the first source of information
as well as for fact-checking and in-depth research. As Wikipedia
relies solely on the efforts of its volunteer editors, its success
might be particularly affected by toxic speech. In this paper, we
analyze all 57 million comments made on user talk pages of 8.5
million editors across the six most active language editions of
Wikipedia to study the potential impact of toxicity on editors'
behavior. We find that toxic comments are consistently associated
with reduced activity of editors, equivalent to 0.5-2 active days per
user in the short term. This translates to multiple human-years of
lost productivity, considering the number of active contributors to
Wikipedia. The effects of toxic comments are potentially even greater
in the long term, as they are associated with a significantly
increased risk of editors leaving the project altogether. Using an
agent-based model, we demonstrate that toxicity attacks on Wikipedia
have the potential to impede the progress of the entire project. Our
results underscore the importance of mitigating toxic speech on
collaborative platforms such as Wikipedia to ensure their continued
success.

Significance Statement

While the prevalence of toxic speech online is well studied, its true
impact on the productivity of online communities remains largely
unexplored. In this study, we focus on Wikipedia, which as the
largest and most-read online reference, serves as a vital source of
knowledge for millions of users worldwide. By analyzing all comments
made over 20 years on user talk pages of 8.5 million editors across
multiple language editions, we demonstrate that toxic speech is
associated with a significant loss in the productivity of Wikipedia
editors. These findings may have broad implications for large-scale
collaborative projects and online communities, emphasizing the need
to promote healthy and sustainable communication practices to protect
crucial online information ecosystems and ensure their long-term
success.

Introduction

Wikipedia is arguably one of the most successful collaborative
projects in history. It has become the largest and most-read
reference work ever created, and it is currently the fifth most
popular website on the Internet (1). Millions of users worldwide rely
on Wikipedia as their first source of information when encountering a
new topic, for fact-checking and in-depth research (2). Even if
caution might be required when consulting less actively maintained
pages (3), numerous studies have shown that Wikipedia is a reliable
source of information in areas ranging from political science (4) to
pharmacology (5) and its accuracy is comparable to traditional
encyclopedias (6) and textbooks (7).

One of the most remarkable aspects of Wikipedia's success is that its
content is exclusively created and curated by volunteer editors,
known as Wikipedians. The English edition alone has more than 120,000
active editors (8). However, this volunteer-driven model also makes
Wikipedia susceptible to the inherent challenges associated with
maintaining such a large online community (9, 10). For example, it
has been previously observed that Wikipedia is not free of conflict,
particularly in the form of so-called edit wars (11), which impose
significant costs on the project (12) and could negatively affect the
quality of Wikipedia articles (13).

In this paper, we focus on the impact of toxic comments directed
toward editors on their activity. This aspect is less studied, but
potentially not less important, as affected by toxic comments,
Wikipedians might reduce their contributions or abandon the project
altogether, threatening the success of the platform (14).

Toxicity has been extensively studied on popular social media
websites such as Twitter (15, 16), Reddit (17, 18), and similar
platforms (19, 20). However, much of these research focuses on
automated toxicity detection and prevalence estimation rather than on
evaluating its impact (21). As an online encyclopedia, Wikipedia is
often perceived as immune to toxicity and has a strict "No personal
attacks" policy (22). Despite that, toxic speech and harassment have
been previously observed on the platform (23-27). The effects of such
behaviors on editors' contributions are, however, not well understood
nor well studied. The largest study to date relies on a voluntary
opt-in survey of the 3,845 Wikipedians conducted in 2015 (24). It
reports that 20% of users witnessing harassment have stopped
contributing for a while, 17% considered not contributing anymore and
5% stopped contributing at all.

In this paper, we analyzed all 57 million comments made on user talk
pages of editors on the six most active language editions of
Wikipedia (English, German, French, Spanish, Italian, Russian) to
understand the potential impact of toxic speech on editors'
contributions (see Methods and materials section for our definition
of toxic comments). User talk pages are a place for editors to
communicate with each other either on more personal topics or to
extend their discussion from an article's talk page. The majority of
toxic comments are left on user talk pages (28). The comments we
study were extracted from revision histories of talk pages and, thus,
include even those toxic comments that were later archived or deleted
by the page owner.

Figure 1 shows the activity of 50 randomly selected users who have
received exactly one toxic comment. While some users are seemingly
unaffected by a toxic comment, others temporarily reduce their
activity or leave the project completely. The aim of our paper is to
quantify this effect on the entire population of editors.

Fig. 1.
After receiving a toxic comment many users temporarily reduce their
activity or leave the project completely. The figure shows the
activity of 50 randomly selected users who received exactly one toxic
comment. Blue squares indicate an active day, i.e. a day when at
least one edit was done, starting from the first contribution of a
given user. Red triangles correspond to toxic comments. Note that
while some users are resilient and their activity is seemingly
unaffected by toxic comments, many users temporarily reduce their
activity or stop contributing altogether.
Open in new tabDownload slide

After receiving a toxic comment many users temporarily reduce their
activity or leave the project completely. The figure shows the
activity of 50 randomly selected users who received exactly one toxic
comment. Blue squares indicate an active day, i.e. a day when at
least one edit was done, starting from the first contribution of a
given user. Red triangles correspond to toxic comments. Note that
while some users are resilient and their activity is seemingly
unaffected by toxic comments, many users temporarily reduce their
activity or stop contributing altogether.

We estimate the number of lost active days associated with a toxic
comment by comparing the number of active days before and after
receiving a toxic comment. To account for potential baseline change,
we have matched editors that received a toxic comment with similarly
active editors who received a nontoxic comment. We have separately
studied if toxic comments increase the probability of editors leaving
the project altogether. Finally, we have used an agent-based model to
model the potential impact of an increased number of toxic comments
on Wikipedia.

Results

Loss of editor activity

To estimate the potential effect of a toxic comment, we compute the
proportion of users who were active on day X before or after
receiving a toxic comment (Fig. 2). We find that, on average, editors
are more active near the time when they receive a toxic comment, with
a peak at 24 h prior to the comment. At this time point, more than 
40% of editors were active, as shown by the red line in Fig. 2a. This
is a rather unsurprising observation since toxic comments are often
made as a reaction to an edit made by a user and, thus, users are
expected to be active around the time of a toxic comment. Note that
if the timestamps around which the curve is centered are shuffled
(black line in Fig. 2a) then this pattern disappears completely as
expected.

Fig. 2.
After receiving a toxic comment, users become less active. On
average, users are more active near the time when they receive a
toxic comment (peak at zero for the red line in panel a). Average
activity across all users who have received a toxic comment is lower
in all 100 days after the event compared to the corresponding days
before (dashed and solid red lines in panel b). This cannot be
explained by a baseline drop in activity after a nontoxic comment
(dashed and solid blue lines in panel b). Similar results hold not
only for the English edition but also for the other five editions
(c-g).
Open in new tabDownload slide

After receiving a toxic comment, users become less active. On
average, users are more active near the time when they receive a
toxic comment (peak at zero for the red line in panel a). Average
activity across all users who have received a toxic comment is lower
in all 100 days after the event compared to the corresponding days
before (dashed and solid red lines in panel b). This cannot be
explained by a baseline drop in activity after a nontoxic comment
(dashed and solid blue lines in panel b). Similar results hold not
only for the English edition but also for the other five editions
(c-g).

We also find that average activity across all users who have received
a toxic comment is lower during all 100 days after the event compared
to the corresponding days before (dashed and solid red lines in Fig. 
2b), e.g. smaller number of users is active five days after receiving
a toxic comment than five days before receiving it. To rule out the
possibility that this is due to a general drop in activity over time
or a drop in activity after any comment, we select a control group of
users who have received a nontoxic comment, and whose average
activity in the 100 days before the comment is the same as the
average activity of users who received a toxic comment (see Methods
and materials section for details).

We observe a similar characteristic peak around the nontoxic comment,
likely due to both toxic and nontoxic comments being reactions to a
contribution made by an editor. However, in contrast to a toxic
comment, a nontoxic comment does not lead to a significant decrease
in activity (dashed and solid blue lines in Fig. 2b). Similar results
hold for all six language editions that we have examined (Fig. 2c-g).

We then estimate the lost activity associated with a toxic comment by
computing the decrease in activity after a toxic comment, taking into
account a potential baseline drop, i.e. by computing D=
(Aftertoxic-Beforetoxic)-(Afternontoxic-Beforenontoxic). We find
that this loss is statistically significant for all language editions
studied (Table 1). We further explored the robustness of this result
with respect to the toxicity threshold and potential filtering of
users according to their activity. As expected, for higher toxicity
thresholds, i.e. for more severely toxic comments, the effect is
stronger (Supplementary Fig. S1). Considering only active users also
leads to higher estimates; however, here we are reporting a
conservative estimate, i.e. no filtering is used for results
presented in Fig. 2 and Table 1.

Table 1.

Lost active days in the 100 days following a toxic comment.

Edition .   D .   P-value .  Nusers . 
English    -1.207 2.6x10-66  36,332
German     -0.546 1.5x10-7   10,346
French     -1.851 4.8x10-9   2,239
Spanish    -0.563 8.6x10-3   2,446
Italian    -0.336 2.3x10-2   3,567
Russian    -1.219 7.8x10-4   1,134

Edition .   D .   P-value .  Nusers . 
English    -1.207 2.6x10-66  36,332
German     -0.546 1.5x10-7   10,346
French     -1.851 4.8x10-9   2,239
Spanish    -0.563 8.6x10-3   2,446
Italian    -0.336 2.3x10-2   3,567
Russian    -1.219 7.8x10-4   1,134

The lost active days are estimated by computing the difference
between the number of active days during 100 days after a toxic
comment and the number of active days during 100 days before a toxic
comment. This difference is then compared with the baseline drop
after a nontoxic comment, i.e. D=(Aftertoxic-Beforetoxic)-
(Afternontoxic-Beforenontoxic). The P-value is computed using
Student's t-test.

Open in new tab
Table 1.

Lost active days in the 100 days following a toxic comment.

Edition .   D .   P-value .  Nusers . 
English    -1.207 2.6x10-66  36,332
German     -0.546 1.5x10-7   10,346
French     -1.851 4.8x10-9   2,239
Spanish    -0.563 8.6x10-3   2,446
Italian    -0.336 2.3x10-2   3,567
Russian    -1.219 7.8x10-4   1,134

Edition .   D .   P-value .  Nusers . 
English    -1.207 2.6x10-66  36,332
German     -0.546 1.5x10-7   10,346
French     -1.851 4.8x10-9   2,239
Spanish    -0.563 8.6x10-3   2,446
Italian    -0.336 2.3x10-2   3,567
Russian    -1.219 7.8x10-4   1,134

The lost active days are estimated by computing the difference
between the number of active days during 100 days after a toxic
comment and the number of active days during 100 days before a toxic
comment. This difference is then compared with the baseline drop
after a nontoxic comment, i.e. D=(Aftertoxic-Beforetoxic)-
(Afternontoxic-Beforenontoxic). The P-value is computed using
Student's t-test.

Open in new tab

While these results demonstrate that our findings are not limited to
one language, they should not be used to compare effects between
language editions, as there is no guarantee that the same toxicity
threshold for the toxicity detection algorithm will have the same
meaning in different languages.

Note that given that thousands of users have received at least one
toxic comment (Supplementary Table S1), even a moderate loss per user
could result in many human-years of lost productivity for Wikipedia
in the short run. By multiplying the estimated loss per user from
Table 1 by the number of users who have received at least one toxic
comment, we could estimate the total loss of activity that is ranging
from 5 human-years for Russian Wikipedia to 265 human-years for the
English edition. The reason for the lasting effect of toxicity is
that some new users might be discouraged by a toxic comment and
choose to leave the project altogether after just a few
contributions. This means that a single toxic comment could deprive
Wikipedia of a potentially long-term contributor.

To further investigate this effect, we compare the probability of
leaving Wikipedia after receiving a toxic comment with the
probability of leaving Wikipedia after receiving a nontoxic comment.

Leaving Wikipedia

We observed that the probability of leaving Wikipedia after N
contributions declines with N. PN(leaving) is approximately
proportionate to N-a, where a ranges from 0.89 to 1.02, indicating
a long-tailed distribution. While the probability of leaving the
project after the first and only contribution is high (P1=47% for
English Wikipedia), the risk of leaving Wikipedia drops to 0.7% for
users who have made 100 contributions. To study the potential effects
of toxic comments, we separately consider contributions that are
followed by a toxic comment and contributions that are not followed
by a toxic comment (see Methods and materials section for details).
We find that the risk of an editor leaving after a toxic comment is
consistently higher for all editions and regardless of the
contribution number, see Fig. 3. We provide an analysis of the
significance of these findings in Supplementary Fig. S4.

Fig. 3.
The probability of leaving Wikipedia after receiving a toxic comment
is substantially higher than might be expected otherwise. For all six
editions the probability of leaving declines with the number of
contributions. At the same time, this probability is substantially
higher after receiving a toxic comment than might be expected
otherwise. Dots are probability estimates and solid lines are the
best linear fit on a log-log scale.
Open in new tabDownload slide

The probability of leaving Wikipedia after receiving a toxic comment
is substantially higher than might be expected otherwise. For all six
editions the probability of leaving declines with the number of
contributions. At the same time, this probability is substantially
higher after receiving a toxic comment than might be expected
otherwise. Dots are probability estimates and solid lines are the
best linear fit on a log-log scale.

Agent-based modeling

As has been demonstrated above, toxic comments increase the
likelihood of editors abandoning Wikipedia. If enough editors leave,
this could potentially impede the progress of the project as a whole.
In order to estimate the potential impact of toxic comments, we model
users' behaviors by varying the toxicity of the environment, ranging
from a nontoxic environment, where the probability of a user leaving
follows the empirically observed nontoxic probability distribution, 
PNnon (blue dots in Fig. 3), to a highly toxic environment, where the
probability of leaving corresponds to an empirically observed toxic
probability distribution, PNtox (red dots in Fig. 3). We also
consider a potential attack targeted at new users. In this scenario,
each user receives a toxic comment after their first and second
contributions, e.g. their probability of leaving after the first and
second contribution is defined by PNtox, and after that follows the
empirically observed PN.

For our modeling, we focus on a cohort of users who made their first
contribution between the 4,000th and 6,000th day from the first
recorded contribution to English Wikipedia in our dataset. We opted
for this timeframe as it reflects Wikipedia's current phase
characterized by a relatively consistent number of active editors.
This period follows the site's initial exponential growth and a
subsequent decline but comes before the anomalous increase in
activity due to the COVID-19 pandemic (see Discussion section for
details on these stages).

For our modeling, we employed an agent-based approach. Each day,
agents (representing users) join Wikipedia and make their first
contribution. The number of agents joining each day is equal to the
actual count of first-time contributors to English Wikipedia on that
particular day. After their first contribution, agents keep
contributing, following a Poisson process, i.e. in such a way that
the distance between two consecutive contributions, D, follows an
exponential distribution: D~Exp(l), where l is estimated from
empirical data. After each contribution, the agent's probability of
leaving the project is determined by the toxicity level, T, and the
empirically observed distributions PNnon and PNtox. In particular,
after N's contribution the user leaves the project with probability 
T*PNtox+(1-T)*PNnon. If the toxicity level is 0, then the
probability of leaving follows the nontoxic distribution PNtox, and
if the toxicity level is 1, then the probability of leaving follows
the toxic distribution PNtox.

After the initial 2,000 days, no new agents join the project;
however, we continue to model the behavior of the remaining agents
for the subsequent 2,000 days, for which we have available empirical
data for comparison.

Our model generally reproduces the dynamics of user activity (Fig. 4
), though, as expected, it cannot account for a later
COVID-19-induced spike in activity. We find that an extreme level of
toxicity could effectively reduce the cohort to almost no users in
the long run, compared to the sustained numbers in a nontoxic setting
or as observed in the data. Additionally, targeted attacks on
newcomers have the potential to significantly decrease the number of
active users, posing a risk to the project. The detailed results of
our modeling, showing the effects of different toxicity levels on
user count, are presented in Supplementary Fig. S6.

Fig. 4.
High levels of toxicity and targeted attacks could significantly
reduce the number of active editors. Modeling results for a cohort of
editors making their first contribution during the relatively stable
phase of Wikipedia (shaded region in the inset). The model reproduces
the general dynamics of user activity (blue line) but, as expected,
cannot capture the COVID-19-related spike in activity. An extreme
level of toxicity (red line) could reduce the cohort to virtually no
active users, contrasting with a nontoxic environment (green line) or
actual activity (blue line). Targeted attacks on newcomers (orange
line) have the potential to significantly reduce the number of active
contributors.
Open in new tabDownload slide

High levels of toxicity and targeted attacks could significantly
reduce the number of active editors. Modeling results for a cohort of
editors making their first contribution during the relatively stable
phase of Wikipedia (shaded region in the inset). The model reproduces
the general dynamics of user activity (blue line) but, as expected,
cannot capture the COVID-19-related spike in activity. An extreme
level of toxicity (red line) could reduce the cohort to virtually no
active users, contrasting with a nontoxic environment (green line) or
actual activity (blue line). Targeted attacks on newcomers (orange
line) have the potential to significantly reduce the number of active
contributors.

Discussion

We conducted a large-scale analysis, covering all comments made on
user talk pages of the six most active language editions of Wikipedia
over a period of 20 years, and found that toxic comments are
associated with a decreased activity of editors who have received
these comments and an increased risk of them leaving the project
altogether. Additionally, via agent-based modeling, we showed that
toxicity attacks on Wikipedia have the potential to impede the
progress of the entire project.

The main limitation of our study is its relatively narrow scope, as
it focuses solely on the association between toxic comments left on
user talk pages and the subsequent decrease in users' activity.
However, this approach allowed us to formulate our findings with
precision and ensure their robustness. We believe that our study
complements and extends existing studies on Wikipedia and online
communities more broadly, and may serve as a foundation for further
exploration of the effects of toxicity, as we discuss in this
section.

Conflict on Wikipedia

Conflict on Wikipedia has already been a subject of numerous studies,
with particular attention given to so-called "edit wars" (11, 29, 30
). These arise when groups of editors, disagreeing about page
content, repeatedly override each other's contributions. It has been
estimated that edit wars can impose substantial conflict and
coordination costs on Wikipedia (12). Furthermore, it has been
demonstrated that these costs increase over time and a smaller
proportion of the total work by Wikipedians directly contributes to
new article content. Conflict could also undermine content quality.
For instance, the level of conflict on discussion pages, as assessed
by raters, has been shown to negatively correlate with the quality of
the corresponding Wikipedia articles (13).

In contrast to previous studies, our focus is on comments left on
user talk pages rather than article talk pages. While this narrows
the scope of our study, it also ensures that the comments we examine
are directly addressed to a specific editor. Our approach also
mitigates potential bias that could be introduced by the topic of an
article. For instance, comments on talk pages linked to articles
about violence might be misclassified as toxic by an algorithm due to
the presence of highly negative keywords.

It is possible that toxic comments we observe on user talk pages are
not independent from a broader conflict occurring elsewhere on
Wikipedia. Therefore, it is conceivable that the effect we observe is
not purely explained by toxic comments, but also by a broader
conflict which leads both to a toxic comment on a user talk page and
decreased activity of this user. Future research is needed to address
this limitation and explore the context in which toxic comments
occur.

It is worth noting, however, that it has already been established
that toxicity on its own could lead users to stop contributing either
temporarily or permanently, as this is what editors themselves report
in surveys (24). Our study complements such studies by providing an
estimate of the potential effects while also being performed on a
scale that is not achievable by survey methods.

Stages of Wikipedia life cycle

Wikipedia has not grown linearly but has instead passed through
several stages. It began with exponential growth (31), which
subsequently slowed (32). Following that, the number of active users
declined before Wikipedia entered its current stage, characterized by
a relatively stable number of active users (33), with a slow decline
observed in some language editions. A notable exception was a
temporary spike in activity due to the COVID-19 pandemic (34). See 
Supplementary Fig. S5 for an illustration of these patterns in the
editions studied in this paper.

It has been found that the main reason for halted growth is a sharp
decline in the retention of newcomers (35). Specifically, with the
project's development, the rejection of newcomer contributions has
increased, demotivating them and driving them away. Our results
complement these findings by highlighting that newcomers are also
particularly vulnerable to toxic comments. If users receive a toxic
comment after their first or second contributions, their chances of
continuing to contribute are 1.8 times lower compared to users who
did not receive toxic comments.

Diversity of editors

Wikipedia is often considered a neutral and unbiased source of
knowledge. In fact, this is ingrained in its "Neutral point of view"
policy, which is officially one of the five fundamental principles of
Wikipedia (36). However, the claim of neutrality should not be
accepted uncritically (37). For instance, while Wikipedia mandates
that its content is supported by reliable sources, the selection of
these sources can significantly deviate from the norms of the expert
knowledge community, introducing biases to Wikipedia content (38).
Even if the content of articles is neutral, their coverage may be
biased. It is well documented, for example, that biographies of women
are underrepresented on Wikipedia (39). Wikipedia's own rules might
contribute to such biases. For instance, providing reliable sources
as required by Wikipedia for biographies of women might be
challenging because fewer sources exist on women due to historic
inequalities (40). Another case in point is the Oral Citations
project, which aimed to use oral citations for content on countries
that are underrepresented in other sources (41). However, this
initiative was met with opposition by the English Wikipedia
community.

These content biases are closely connected to the lack of diversity
among editors (38, 42). While estimates vary, the vast majority of
Wikipedians are men (43). Notably, Wikipedia did not achieve its own
goal of having at least 25% women editors by 2015 (44). This
shortfall is a significant concern for the project, as diversity can
improve the quality of content and reduce its biases (13, 45). While
multiple barriers confront women editors on Wikipedia (40, 46, 47),
toxicity is likely to be one of key factors contributing to the
observed gender imbalance. Specifically, research has shown that
while men and women are equally likely to face online harassment and
abuse, women experience more severe violations (48). They are also
more likely to be affected by such incidents and to self-censor in an
attempt to prevent potential harassment (48). This has been confirmed
in the Wikipedia context as well, where it has been demonstrated that
the psychological experiences of women and men editors differ,
leading to higher attrition rates among women (49). Similar results
were found in another survey (24), showing that women experiencing
toxicity are more likely to stop contributing in the future.

Overall, there are reasons to believe that toxicity might
significantly undermine the diversity of Wikipedia editors, which
can, in turn, compromise the quality of Wikipedia articles and
introduce biases in its coverage. This underscores the importance of
our findings. While most of the existing studies focus on the gender
gap, we want to emphasize that the Wikipedia diversity problem goes
beyond that, including racial, nonbinary, and other biases as well (
50-52). For instance, we observed that many of the toxic comments in
our data set include ethnic slurs. Future studies are needed to
better understand the experiences of minority groups on Wikipedia and
the effects that toxicity has on them.

Interventions

The Wikipedia community is well aware of the aforementioned problems,
and there have been multiple efforts to address them through various
interventions. Research into reward systems showed that while they
might work effectively for already highly productive editors, they
fail to motivate less active editors (53). Another study found no
significant effect of positive rewards in online communities (54).

To address the gender gap in Wikipedia content, numerous events
dedicated to creating entries about women were organized (46). An
analysis of such interventions, which focused on two popular feminist
interventions, confirmed that they succeeded in introducing content
about women that would otherwise be missing (55). However, there is
still a need to address the gender gap on a more systematic and
sustainable level. For instance, one study showed that most of the
women activists who attended editing workshops later chose not to
continue contributing to Wikipedia, citing safety concerns as their
primary reason (46). This issue was echoed in another study which
identified safety as a core concern for women editors (56).

A suggested solution to this problem has been the red-flagging of
harassment and harassers (46). However, the opinion that toxic
comments are negligible and should be seen as merely
over-enthusiastic participation is still present among editors (25).
Furthermore, various anti-harassment measures have been declined
multiple times by the community, as they were seen to slow the
process of content creation (57, 58). Based on our findings, we
believe there is a need to reevaluate these policies, and more
research attention is required to understand the impact of potential
interventions.

The wider role of peer-production systems

Wikipedia plays a crucial role in the global information
infrastructure, aiming to provide millions of people with access to
free, unbiased knowledge. Due to its reputation as a neutral and
comprehensive information source, it has become a trusted first
choice source of knowledge for many and its articles frequently
appear in top search engine results (59, 60). In fact, studies have
shown that Google search results rely heavily on Wikipedia, and the
quality of these results significantly diminishes without Wikipedia (
61). Beyond search engines, Wikipedia was shown to be valuable to
other online communities such as Stack Exchange and Reddit (62).

While Wikipedia is arguably the most successful peer-production
system, it is certainly not the only one. Others include hundreds of
wikis hosted by Fandom, the numerous question-and-answer communities
of Stack Exchange, and various other platforms ranging from online
maps to online learning (33). Interestingly, for these projects, the
same patterns that are typical of Wikipedia have been observed (63),
i.e. the initial growth in number of contributors is followed by a
decline characterized by a decreased retention of newcomers. This
suggests that our findings might have broader implications for
large-scale collaborative projects and online communities. It
emphasizes the need to promote healthy and sustainable communication
practices to protect crucial online information ecosystems and ensure
their long-term success.

Methods and materials

Data and preprocessing

Comments on user talk pages

The Wikimedia Foundation provides publicly accessible dumps of all
the different wikis' content.^a These dumps are updated on a regular
basis, with complete revision history dumps generated once per month.
For this paper, we used the English dump from 2021 November 1, the
German dump from 2022 August 1, the French, Italian, and Spanish
dumps from 2022 August 1, and the Russian dump from 2022 July 1. The
data was obtained from a mirror hosted by the Umea University,
Sweden.^b

From the dumps, the user talk pages were extracted. A user's talk
page is a place where other editors can communicate with the user
either on more personal topics or to extend their discussion from an
article talk page. When the comments left on the talk page are
resolved or become too old, users can choose to archive them. This
helps them keep better track of new incoming topics. Once archived,
the old comments are not displayed on the talk page anymore but are
rather linked in a subpage. Nevertheless, the entire history of the
user talk page, as of any other page on Wikipedia, can be fully seen
under the tab of revision history. The revision history records one
entry for every edit made on the page saving each time the complete
content of the page. Thus retrieving a single comment requires
performing the difference between two consecutive revisions. The
Wikimedia API does offer a method to compute the difference between
two revisions, however, applying it on a scale that was necessary for
this research was unfeasible. For that reason, we developed our own
parser to extract comments as a difference between two versions of
the page (64).

We excluded from our analysis talk pages that belong to unregistered
users, e.g. users who are represented only by an IP address rather
than a user name, because IP addresses are dynamic and it can not be
assumed that one address represents a single user throughout
Wikipedia history. Additionally, we have excluded comments made by
officially registered bots. Comments that were made by users on their
own pages are also not considered.

When extracting comments, we cleared wiki-specific formatting and
HTML markup, i.e. removed links, attachments, or other
formatting-specific sequences irrelevant to the actual content.

Contributions and active days

In order to extract information on users' contributions, i.e. edits
of Wikipedia pages made by them, we used the MediaWiki API to
retrieve timestamps for each edit made by a given user. The resulting
data set is publicly available in the project repository (64). The
timestamps of contributions were then converted into active days.
Specifically, each user i was represented as a binary vector ui=
(ai1,ai2,...,aiN), where aid=1 if user i made at least one
contribution, i.e. edited a Wikipedia page, within the 24-h period
corresponding to day d and aid=0 otherwise. N is the number of days
between the first recorded contribution in our data set and the last.
The conversion from contribution count to active days was performed
because it is hard to interpret and compare the total number of
contributions between users as one large contribution could be
equivalent to multiple smaller ones. Additionally, the size of a
contribution does not necessarily reflect the effort put into it.
While being active on a given day could still mean different levels
of activity for different users, it represents a certain level of
engagement with the project and is substantially different from not
contributing at all on a given day.

Toxicity

The automatic detection of offensive language in online communities
has been an active area of research since at least 2010 (65). Over
the past decade, researchers have focused on detecting
closely-related and intersecting types of offensive language such as
toxicity, abusive language, and hate speech (66), see (67) for an
overview of recent advancements in the field. In this paper, we use a
model from the Perspective API (68) to identify toxic comments. This
is a state-of-the-art toxicity detection algorithm that obtained
competitive results at OffensEval-2019 competition (69) without any
additional training on the contest data and is often used as a
baseline system for toxicity detection (66). Perspective API is used
across multiple platforms, including The New York Times, Der Spiegel,
Le Monde, and El Pais. It uses BERT (Bidirectional Encoder
Representations from Transformers) architecture (70) and is trained
on comments from a variety of online sources, including Wikipedia.
Each comment is labeled by 3-10 crowdsourced raters. Perspective
models provide scores for several different attributes, see 
Supplementary Table S2 for the list of attributes and their
definitions, see Supplementary Table S2 for examples of toxic
comments, and see Supplementary Table S3 for the AUC (Area Under the
Curve) scores for those languages and attributes that were used in
this paper.

We define a toxic comment as a comment that has a score of at least 
0.8 on any of the six dimensions provided by Perspective API. The 0.8
score means that on average 8 out of 10 raters would mark it as
toxic. As this threshold can be considered arbitrary, we perform
additional robustness checks using different toxicity thresholds. In
particular, we compute activity loss not only for the threshold of 
0.8 (Table 1) but for thresholds from 0.2 to 0.9. Additionally, we
applied different activity filters, e.g. we separately compute an
estimate only for those users who were active at least X days in the
past 100 days where X varies from 0 to 50. This is done in order to
ensure that the results are not exclusively driven by those users who
had made few edits and then stopped contributing to the project. We
perform this analysis for English Wikipedia as it is the largest
edition. As shown in Supplementary Fig. S1, the estimate is typically
in the range from -0.5 to -2 and significantly lower than zero for
all activity thresholds and all toxicity thresholds higher than 0.3.
Similarly, we have checked how the toxicity threshold affects the
probability of leaving the project. As might be expected, results
remain qualitatively the same for different toxicity thresholds but
higher thresholds lead to more extreme results, e.g. the probability
of leaving after a toxic comment with 0.9 score is even higher than
after a toxic comment with toxicity score of 0.8 (Supplementary Fig.
S3).

We also evaluated the robustness of our results with respect to
misclassification errors. To achieve a realistic distribution of user
activity, we repeatedly sampled 100,000 editors and their activity
histories from the English Wikipedia data set. These sampled users
were then divided into two groups: treatment and control. We
investigated two distinct scenarios: one involving an equal split
between the treatment and control groups and a second, more
realistic, scenario where the treatment group constituted 1% of the
control group.

In the treatment group, we randomly removed one active day from each
user, thereby generating a true effect of one lost active day per
user. We then introduced misclassification errors by generating false
positives (moving users from control to treatment group) and false
negatives (moving users from treatment to control group). Finally, we
compared the estimated effect, as a function of the error rate, with
the true effect.

We find that, generally, misclassification leads to the
underestimation of the true effect, becoming more pronounced with
higher error rates (Supplementary Fig. S2). The only exception is in
the case of false negatives, i.e. undetected toxic comments, in the
realistic scenario. Here, misclassification does not significantly
bias the estimate, though it does increase its variance.

Perspective API accepts texts up to 20,480 bytes. As the majority of
comments are well below this limit, we have excluded those that are
larger.

Activity loss

Users who have received at least one toxic comment constitute our
treatment group. For each user in this group, we select a random
toxic comment they have received. We then center user activity around
the timestamp, titox, of that toxic comment and convert the result
to active days by calculating

sign(|{t[?]Ti:t[?][titox+d*24*60*60,titox+(d+1)*24*60)}|),

where Ti is the set of timestamps of all contributions made by user i
, and d is a day ranging from -100 to 100. Finally, the results are
averaged over all users. We repeat the procedure of selecting a
random toxic comment 100 times and report average results. However,
since most users received only one toxic comment, there is little
variation across simulations and the average over 100 simulations is
almost identical to the result of a single simulation.

We then compare these results with a control group comprised of users
who did not receive any toxic comments. However, a direct comparison
is complicated because users who have received a toxic comment are,
on average, more active than those who have not. This is probably due
to the fact that each contribution could lead to a toxic response
with a certain probability. Hence, the more contributions a user
makes, the higher the likelihood of receiving a toxic comment and
thereby being in the treatment group.

Specifically, if each contribution can lead to a toxic comment with a
probability p, then the probability of receiving at least one toxic
comment depends on the number of contribution, N: P(gettoxiccomment)=
1-(1-p)N(1).

To ensure our control group is similarly active as the treatment
group, we randomly select users with a probability based on the
number of their contributions using formula (1). Users selected in
this manner form the control group. For these users, we then pick a
nontoxic comment at random, center their activity around its
timestamp, and follow the same procedure used for the treatment
group.

To test for the significance of the results, we compute 95%
bootstrapped confidence intervals for each estimate.

Probability of leaving

For each toxic comment, we find the closest in time contribution that
precedes that comment. We define such contributions as "contributions
followed by a toxic comment" and compare the probability of leaving
after such contributions with the probability of leaving after other
contributions. The probability of leaving after N contributions is
estimated as a fraction of users who have made exactly N
contributions among users who have made at least N contributions. As
the probability of leaving strongly depends on N, we make a
comparison separately for each contribution number N[?][1,100]. For N>
100 the number of users is too small to provide reliable estimates
for comparison.

a

https://meta.wikimedia.org/wiki/Data.dumps [accessed on 2023 January
20].

b

https://mirror.accum.se/mirror/wikimedia.org/

Acknowledgments

We acknowledge the Master's thesis by Bruckner (71), which identified
a potential pattern in data and provided an inspiration for the
design of the study presented in this paper. The initial data
collection and experiments were carried out as part of Camelia
Oprea's Master's thesis (72). We thank Liubov Tupikina and David
Garcia for their valuable discussions regarding the results presented
in this article. We thank the anonymous reviewers for their
insightful comments and suggestions.

Supplementary Material

Supplementary material is available at PNAS Nexus online.

Funding

The publication of this article was funded by the University of
Mannheim.

Author Contributions

I.S., C.O., and M.S. designed the study; I.S. and C.O. collected and
analyzed the data; I.S., C.O., and M.S. wrote the manuscript; I.S.
revised the manuscript.

Previous Presentation

These results were previously presented at International Conference
on Computational Social Science 2023.

Preprints

A preprint of this article is published at https://doi.org/10.48550/
arXiv.2304.13568

Data Availability

The data underlying this article is available in Open Science
Framework at https://osf.io/2qyxj/.

References

 
1

Semruch. 2023. Most visited websites in the world [accessed 2023
Sept]. https://www.semrush.com/website/top/.

 
2

Singer
P
, et al.
2017