Conflict A 2011 study reported a new way to measure how disputed a Wikipedia article is, and verified against 6
Indo-European language editions including
English. A 2013 article in
Physical Review Letters reported a generic
social dynamics model in a collaborative environment involving opinions, conflicts, and consensus, with a specific analogue to Wikipedia: "a peaceful article can suddenly become controversial when more people get involved in its editing." In 2014 published as a book chapter titled "The Most Controversial Topics in Wikipedia: A Multilingual and Geographical Analysis": analysed the volume of editing of articles in various language versions of Wikipedia in order to establish the most controversial topics in different languages and groups of languages. For the English version, the top three most controversial articles were
George W. Bush,
Anarchism and
Muhammad. Topics in other languages causing most controversy were Croatia (German),
Ségolène Royal (French), Chile (Spanish) and Homosexuality (Czech).
Demographics A 2007 study by
Hitwise, reproduced in
Time magazine, found that visitors to Wikipedia are almost equally split 50/50 male/female, but that 60% of edits are made by male editors. A 2010 survey found that only 13% of editors and 31% of readers were female.
Policies and guidelines A descriptive study that analyzed English language Wikipedia's policies and guidelines up to September 2007 identified a number of key statistics: • 44 official policies • 248 guidelines Even a short policy like "ignore all rules" was found to have generated a lot of discussion and clarifications: The study sampled the expansion of some key policies since their inception: •
Wikipedia:Ignore all rules: 3600% (including the additional document explaining it) •
Wikipedia:Consensus: 1557% •
Wikipedia:Copyrights: 938% •
Wikipedia:What Wikipedia is not: 929% •
Wikipedia:Deletion policy: 580% •
Wikipedia:Civility: 124% The number for "deletion" was considered inconclusive however because the policy was split in several sub-policies.
Power plays A 2007 joint peer-reviewed study conducted by researchers from the
University of Washington and
HP Labs examined how policies are employed and how contributors work towards consensus by quantitatively analyzing a sample of active talk pages. Using a November 2006
English Wikipedia database dump, the study focused on 250 talk pages in the tail of the distribution: 0.3% of all talk pages, but containing 28.4% of all talk page revisions, and more significantly, containing 51.1% of all links to policies. From the sampled pages' histories, the study examined only the months with high activity, called critical sections—sets of consecutive months where both article and talk page revisions were significant in number. The study defined and calculated a measure of policy prevalence. A critical section was considered
policy-laden if its policy factor was at least twice the average. Articles were tagged with 3
indicator variables: • controversial • featured • policy-laden All possible levels of these three factors yielded 8 sampling categories. The study intended to analyze 9 critical sections from each sampling category, but only 69 critical sections could be selected because only 6 articles (histories) were simultaneously featured, controversial, and policy laden. The study found that policies were by no means consistently applied. Illustrative of its broader findings, the report presented the following two extracts from Wikipedia talk pages in obvious contrast: • a discussion where participants decided that calculating a mean from data provided by a government agency constituted original research: • a discussion where logical deduction was used as counterargument for the original research policy: Claiming that such ambiguities easily give rise to power plays, the study identified, using the methods of
grounded theory (Strauss), 7 types of power plays: • article scope (what is off-topic in an article) • prior consensus (past decisions presented as absolute and uncontested) • power of interpretation (a sub-community claiming greater interpretive authority than another) • legitimacy of contributor (his/her expertise etc.) • threat of sanction (blocking etc.) • practice on other pages (other pages being considered models to follow) • legitimacy of source (the cited reference is disputed) Due to lack of space, the study detailed only the first 4 types of power plays that were exercised by merely interpreting policy. A fifth power play category was analyzed; it consisted of blatant violations of policy that were forgiven because the contributor was valued for his or her contributions despite his lack of respect for rules.
Article scope The study considers that Wikipedia's policies are ambiguous on scoping issues. The following
vignette is used to illustrate the claim: The study gives the following interpretation for the heated debate:
Prior consensus The study remarks that in Wikipedia consensus is never final, and what constitutes consensus can change at any time. The study finds that this temporal ambiguity is fertile ground for power plays, and places the generational struggle over consensus in larger picture of the struggle for article ownership: The study uses the following discussion snippet to illustrate this continuous struggle:
Power of interpretation A vignette illustrated how administrators overrode consensus and deleted personal accounts of users/patients with an anonymized illness (named Frupism in the study). The administrator's intervention happened as the article was being nominated to become a featured article.
Legitimacy of contributor This type of power play is illustrated by a contributor (U24) that draws on his past contributions to argue against another contributor who is accusing U24 of being unproductive and disruptive:
Explicit vie for ownership The study finds that there are contributors who consistently and successfully violate policy without sanction:
Obtaining administratorship In 2008, researchers from
Carnegie Mellon University devised a
probit model of
English Wikipedia editors who had successfully passed
the peer review process to become admins. Using only Wikipedia metadata, including the text of edit summaries, their model was 74.8% accurate in predicting successful candidates. The paper observed that despite protestations to the contrary, "in many ways election to admin is a promotion, distinguishing an elite core group from the large mass of editors." Consequently, the paper used
policy capture—a method that compares nominally important attributes to those that actually lead to promotion in a work environment. The overall success rate for promotion decreased from 75% in 2005, to 53% in 2006, and to 42% in 2007. This sudden increase in failure rate was attributed to a higher standard that recently promoted administrators had to meet, and supported by
anecdotal evidence from another recent study quoting some early admins who have expressed doubt that they would pass muster if their election (RfA) were held recently. In light of these developments the study argued that: Contrary to expectations, "running" for administrator multiple times is detrimental to the candidate's chance of success. Each subsequent attempt has a 14.8% lower chance of success than the previous one. Length of participation in the project makes only a small contribution to the chance of a successful RfA. Another significant finding of the paper is that one Wikipedia policy edit or WikiProject edit is worth ten article edits. A related observation is that candidates with experience in multiple areas of the site stood better chance of election. This was measured by the
diversity score, a simple count of the number of areas that the editor has participated in. The paper divided Wikipedia in 16 areas: article, article talk, articles/categories/templates for deletion (XfD), (un)deletion review, etc.
(see paper for full list). For instance, a user who has edited articles, her own user page, and posted once at (un)deletion review would have a diversity score of 3. Making a single edit in any additional region of Wikipedia correlated with a 2.8% increased likelihood of success in gaining administratorship. Making minor edits also helped, although the study authors consider that this may be so because minor edits correlate with experience. In contrast, each edit to an Arbitration or Mediation committee page, or a
Wikiquette notice, all of which are venues for dispute resolution, decreases the likelihood of success by 0.1%. Posting messages to administrator noticeboards had a similarly deleterious effect. The study interpreted this as evidence that editors involved in escalating or protracting conflicts lower their chances of becoming administrators. Saying "thanks" or variations thereof in edit summaries, and pointing out point of view ("POV") issues (also only in edit summaries because the study only analyzed metadata) were of minor benefit, contributing to 0.3% and 0.1% to candidate's chances in 2006–2007, but did not reach statistical significance before. A few factors that were found to be irrelevant or marginal at best: • Editing user pages (including one's own) does not help. Somewhat surprisingly, user talk page edits also do not affect the likelihood of administratorship. • Welcoming newcomers or saying "please" in edit summaries had no effect. • Participating in consensus-building, such as RfA votes or the village pump, does not increase the likelihood of becoming admin. The study admits however that participation in consensus was measured quantitatively but not qualitatively. • Vandal-fighting as measured by the number of edits to the vandalism noticeboard had no effect. Every thousand edits containing variations of "revert" was positively correlated (7%) with adminship for 2006–2007, but did not attain statistical significance unless one is willing to lower the threshold to
p < .1). More confusingly, before 2006 the number of reverts was negatively correlated (−6.8%) with adminship success, against without attaining statistical significance even at
p < .1. This may be because of the introduction of a policy known as "3RR" in 2006 to reduce reverts. The study suggests that some of the 25% unexplained variability in outcomes may be due to factors that were not measured, such as quality of edits or participation in off-site coordination, such as the (explicitly cited) secret
mailing list reported in
The Register. The paper concludes: Subsequent research by another group probed the sensemaking activities of individuals during their contributions to RfA decisions. This work establishes that decisions about RfA candidates is based on a shared interpretation of evidence in the wiki and histories of prior interactions. ==Readership==