Hubbry Logo
WikipediaWikipediaMain
Open search
Wikipedia
Community hub
Wikipedia
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Contribute something
Wikipedia
Wikipedia
from Wikipedia

Wikipedia[c] is a free online multilingual encyclopedia written and maintained by a community of volunteers, known as Wikipedians, through open collaboration and the wiki software MediaWiki. Founded by Jimmy Wales and Larry Sanger in 2001, Wikipedia has been hosted since 2003 by the Wikimedia Foundation, an American nonprofit organization funded mainly by donations from readers.[2] Wikipedia is the largest and most-read reference work in history.[3][4]

Key Information

Initially available only in English, Wikipedia exists in over 340 languages and is the world's ninth most visited website. The English Wikipedia, with over 7 million articles, remains the largest of the editions, which together comprise more than 65 million articles and attract more than 1.5 billion unique device visits and 13 million edits per month (about 5 edits per second on average) as of April 2024.[W 1] As of September 2025, over 25% of Wikipedia's traffic comes from the United States, while Japan accounts for nearly 7%, and the United Kingdom, Germany and Russia each represent around 5%.[5]

Wikipedia has been praised for enabling the democratization of knowledge, its extensive coverage, unique structure, and culture. Wikipedia has been censored by some national governments, ranging from specific pages to the entire site.[6][7] Wikipedia's volunteer editors have written extensively on a wide variety of topics, but the encyclopedia has also been criticized for systemic bias, such as a gender bias against women and a geographical bias against the Global South.[8][9] While the reliability of Wikipedia was frequently criticized in the 2000s, it has improved over time, receiving greater praise from the late 2010s onward.[3][10][11] Articles on breaking news are often accessed as sources for up-to-date information about those events.[12][13]

History

[edit]

Nupedia

[edit]
Wikipedia founders Jimmy Wales and Larry Sanger

Various collaborative online encyclopedias were attempted before the start of Wikipedia, but with limited success.[14] Wikipedia began as a complementary project for Nupedia, a free online English-language encyclopedia project whose articles were written by experts and reviewed under a formal process.[15] It was founded on March 9, 2000, under the ownership of Bomis, a web portal company. Its main figures were Bomis CEO Jimmy Wales and Larry Sanger, editor-in-chief for Nupedia and later Wikipedia.[1][16] Nupedia was initially licensed under its own Nupedia Open Content License, but before Wikipedia was founded, Nupedia switched to the GNU Free Documentation License at the urging of Richard Stallman.[W 2] Wales is credited with defining the goal of making a publicly editable encyclopedia,[17] while Sanger is credited with the strategy of using a wiki to reach that goal.[citation needed] On January 10, 2001, Sanger proposed on the Nupedia mailing list to create a wiki as a "feeder" project for Nupedia.[W 3]

Launch and rapid growth

[edit]

Wikipedia was launched on January 15, 2001 (referred to as Wikipedia Day)[18] as a single English language edition with the domain name www.wikipedia.com,[W 4] and was announced by Sanger on the Nupedia mailing list.[17] The name originated from a blend of the words wiki and encyclopedia.[19][20] Its integral policy of "neutral point of view" arose within its first year.[21] Otherwise, there were initially relatively few rules, and it operated independently of Nupedia.[17] Bomis originally intended for it to be a for-profit business.[22]

The Wikipedia home page on December 20, 2001[d]

Wikipedia gained early contributors from Nupedia, Slashdot postings, and web search engine indexing. Language editions were created beginning in March 2001, with a total of 161 in use by the end of 2004.[W 5][W 6] Nupedia and Wikipedia coexisted until the former's servers were taken down permanently in 2003, and its text was incorporated into Wikipedia. The English Wikipedia passed the mark of 2 million articles on September 9, 2007, making it the largest encyclopedia ever assembled, surpassing the Yongle Encyclopedia made in China during the Ming dynasty in 1408, which had held the record for almost 600 years.[23]

Citing fears of commercial advertising and lack of control, users of the Spanish Wikipedia forked from Wikipedia to create Enciclopedia Libre in February 2002.[W 7] Wales then announced that Wikipedia would not display advertisements, and changed Wikipedia's domain from wikipedia.com to wikipedia.org.[24][W 8]

After an early period of exponential growth,[25] the growth rate of the English Wikipedia in terms of the numbers of new articles and of editors, appears to have peaked around early 2007.[26] The edition reached 3 million articles in August 2009. Around 1,800 articles were added daily to the encyclopedia in 2006; by 2013 that average was roughly 800.[W 9] A team at the Palo Alto Research Center attributed this slowing of growth to "increased coordination and overhead costs, exclusion of newcomers, and resistance to new edits".[25] Others suggested that the growth flattened naturally because articles that could be called "low-hanging fruit"—topics that clearly merit an article—had already been created and built up extensively.[27][28][29]

In November 2009, a researcher at the Rey Juan Carlos University in Madrid, Spain, found that the English Wikipedia had lost 49,000 editors during the first three months of 2009; in comparison, it lost only 4,900 editors during the same period in 2008.[30][31] The Wall Street Journal cited the array of rules applied to editing and disputes related to such content among the reasons for this trend.[32] Wales disputed these claims in 2009, denying the decline and questioning the study's methodology.[33] Two years later, in 2011, he acknowledged a slight decline, noting a decrease from "a little more than 36,000 writers" in June 2010 to 35,800 in June 2011. In the same interview, he also claimed the number of editors was "stable and sustainable".[34] A 2013 MIT Technology Review article, "The Decline of Wikipedia", questioned this claim, reporting that since 2007 Wikipedia had lost a third of its volunteer editors, and suggesting that those remaining had focused increasingly on minutiae.[35] In July 2012, The Atlantic reported that the number of administrators was also in decline.[36] In November 2013, New York magazine stated, "Wikipedia, the sixth-most-used website, is facing an internal crisis."[37] The number of active English Wikipedia editors has since remained steady after a long period of decline.[38][39]

Sister projects

[edit]

Wikipedia has spawned several sister projects, which are also wikis run by the Wikimedia Foundation. These other Wikimedia projects include Wiktionary, a dictionary project launched in December 2002,[W 10] Wikiquote, a collection of quotations created a week after Wikimedia launched,[40] Wikibooks, a collection of collaboratively written free textbooks and annotated texts,[W 11] Wikimedia Commons, a site devoted to free-knowledge multimedia,[W 12] Wikinews, for collaborative journalism,[W 13] and Wikiversity, a project for the creation of free learning materials and the provision of online learning activities.[W 14] Another sister project of Wikipedia, Wikispecies, is a catalog of all species, but is not open for public editing.[41] In 2012, Wikivoyage, an editable travel guide,[42] and Wikidata, an editable knowledge base, launched.[W 15]

Milestones

[edit]
Cartogram showing number of articles in each language as of March 2024. Languages with fewer than 1,000,000 articles are represented by one circle. Languages are grouped by region of continent and each region of continent is presented by a separate color.

In January 2007, Wikipedia first became one of the ten most popular websites in the United States, according to Comscore Networks.[43] With 42.9 million unique visitors, it was ranked ninth, surpassing The New York Times (#10) and Apple (#11).[43] This marked a significant increase over January 2006, when Wikipedia ranked 33rd, with around 18.3 million unique visitors.[44] In 2014, it received 8 billion page views every month.[W 16] On February 9, 2014, The New York Times reported that Wikipedia had 18 billion page views and nearly 500 million unique visitors a month, "according to the ratings firm comScore".[45] As of March 2023, it ranked sixth in popularity, according to Similarweb.[46] Jeff Loveland and Joseph Reagle argue that, in process, Wikipedia follows a long tradition of historical encyclopedias that have accumulated improvements piecemeal through "stigmergic accumulation".[47][48]

On January 18, 2012, the English Wikipedia participated in a series of coordinated protests against two proposed laws in the United States Congress—the Stop Online Piracy Act (SOPA) and the PROTECT IP Act (PIPA)—by blacking out its pages for 24 hours.[49] More than 162 million people viewed the blackout explanation page that temporarily replaced its content.[50][W 17]

In January 2013, 274301 Wikipedia, an asteroid, was named after Wikipedia;[51] in October 2014, Wikipedia was honored with the Wikipedia Monument;[52] and, in July 2015, 106 of the 7,473 700-page volumes of Wikipedia became available as Print Wikipedia.[53] In April 2019, an Israeli lunar lander, Beresheet, crash landed on the surface of the Moon carrying a copy of nearly all of the English Wikipedia engraved on thin nickel plates; experts say the plates likely survived the crash.[54][55] In June 2019, scientists reported that all 16 GB of article text from the English Wikipedia had been encoded into synthetic DNA.[56]

On January 20, 2014, Subodh Varma reporting for The Economic Times indicated that not only had Wikipedia's growth stalled, it "had lost nearly ten percent of its page views last year. There was a decline of about 2 billion between December 2012 and December 2013. Its most popular versions are leading the slide: page-views of the English Wikipedia declined by twelve percent, those of German version slid by 17 percent and the Japanese version lost 9 percent."[57] Varma added, "While Wikipedia's managers think that this could be due to errors in counting, other experts feel that Google's Knowledge Graphs project launched last year may be gobbling up Wikipedia users."[57] When contacted on this matter, Clay Shirky, associate professor at New York University and fellow at Harvard's Berkman Klein Center for Internet & Society said that he suspected much of the page-view decline was due to Knowledge Graphs, stating, "If you can get your question answered from the search page, you don't need to click [any further]."[57] By the end of December 2016, Wikipedia was ranked the fifth most popular website globally.[58] As of January 2023, 55,791 English Wikipedia articles have been cited 92,300 times in scholarly journals,[59] from which cloud computing was the most cited page.[60]

On January 18, 2023, Wikipedia debuted a new website redesign, called "Vector 2022".[61][62] It featured a redesigned menu bar, moving the table of contents to the left as a sidebar, and numerous changes in the locations of buttons like the language selection tool.[62][W 18] The update initially received backlash, most notably when editors of the Swahili Wikipedia unanimously voted to revert the changes.[61][63]

Impacts of generative AI on Wikipedia views

[edit]

Since January 2024, the Wikimedia Foundation has reported a roughly 50 percent increase in bandwidth use from downloads of multimedia content across its projects. According to the foundation, this growth is largely attributed to automated programs, or "scraper" bots, that collect large volumes of data from Wikimedia sites for use in training large language models and related applications.[64]

In October 2025, the Wikimedia Foundation reported an estimated 8 percent decline as compared to the same months in 2024 in human page views. They speculate it reflects the use of generative AI and social media on how people tend to search for information.[65][66]

Collaborative editing

[edit]
Differences between versions of an article are highlighted.

Restrictions

[edit]

Due to Wikipedia's increasing popularity, some editions, including the English version, have introduced editing restrictions for certain cases. For instance, on the English Wikipedia and some other language editions, only registered users may create a new article.[W 19] On the English Wikipedia, among others, particularly controversial, sensitive, or vandalism-prone pages have been protected to varying degrees.[67] A frequently vandalized article can be "semi-protected" or "extended confirmed protected", meaning that only "autoconfirmed" or "extended confirmed" editors can modify it.[68] A particularly contentious article may be locked so that only administrators can make changes.[W 20] A 2021 article in the Columbia Journalism Review identified Wikipedia's page-protection policies as "perhaps the most important" means at its disposal to "regulate its market of ideas".[69] Wikipedia has delegated some functions to bots. Such algorithmic governance has an ease of implementation and scaling, though the automated rejection of edits may have contributed to a downturn in active Wikipedia editors.[70] Bots must be approved by the community before their tasks are implemented.[71]

In certain cases, all editors are allowed to submit modifications, but review is required for some editors, depending on certain conditions. For example, the German Wikipedia maintains "stable versions" of articles which have passed certain reviews.[W 21] Following protracted trials and community discussion, the English Wikipedia introduced the "pending changes" system in December 2012.[72] Under this system, new and unregistered users' edits to certain controversial or vandalism-prone articles are reviewed by established users before they are published.[73] However, restrictions on editing may reduce the editor engagement as well as efforts to diversify the editing community.[74]

Articles related to the Israeli–Palestinian conflict are placed under extended-confirmed protection.[75] Editors also can make only one revert per day across the entire field and can be banned from editing related articles. These restrictions were introduced in 2008.[76] In January 2025, the Arbitration Committee introduced the "balanced editing restriction", which requires sanctioned users to devote only a third of their edits to articles related to the Israeli–Palestinian conflict even when no misconduct rules have been violated.[77][78]

Review of changes

[edit]
Wikipedia's editing interface

Although changes are not systematically reviewed, Wikipedia's software provides tools allowing anyone to review changes made by others. Each article's History page links to each revision.[e][79] On most articles, anyone can view the latest changes and undo others' revisions by clicking a link on the article's History page. Registered users may maintain a "watchlist" of articles that interest them so they can be notified of changes.[W 22] "New pages patrol" is a process where newly created articles are checked for obvious problems.[W 23]

In 2003, economics PhD student Andrea Ciffolilli argued that the low transaction costs of participating in a wiki created a catalyst for collaborative development, and that features such as allowing easy access to past versions of a page favored "creative construction" over "creative destruction".[80]

Vandalism

[edit]

Any change that deliberately compromises Wikipedia's integrity is considered vandalism. The most common and obvious types of vandalism include additions of obscenities and crude humor; it can also include advertising and other types of spam.[81] Sometimes editors commit vandalism by removing content or entirely blanking a given page. Less common types of vandalism, such as the deliberate addition of plausible but false information, can be more difficult to detect. Vandals can introduce irrelevant formatting, modify page semantics such as the page's title or categorization, manipulate the article's underlying code, or use images disruptively.[W 24]

White-haired elderly gentleman in suit and tie speaks at a podium.
American journalist John Seigenthaler (1927–2014), subject of the Seigenthaler biography incident

Obvious vandalism is generally easy to remove from Wikipedia articles; the median time to detect and fix it is a few minutes.[82][83] However, some vandalism takes much longer to detect and repair.[84]

In the Seigenthaler biography incident, an anonymous editor introduced false information into the biography of American political figure John Seigenthaler in May 2005, falsely presenting him as a suspect in the assassination of John F. Kennedy.[84] It remained uncorrected for four months.[84] Seigenthaler, the founding editorial director of USA Today and founder of the Freedom Forum First Amendment Center at Vanderbilt University, called Wikipedia co-founder Jimmy Wales and asked whether he had any way of knowing who contributed the misinformation. Wales said he did not, although the perpetrator was eventually traced.[85][86] After the incident, Seigenthaler described Wikipedia as "a flawed and irresponsible research tool".[84] The incident led to policy changes at Wikipedia for tightening up the verifiability of biographical articles of living people.[87]

Disputes and edit warring

[edit]

Wikipedia editors often have disagreements regarding content, which can be discussed on article Talk pages. Disputes may result in repeated competing changes to an article, known as "edit warring".[W 25][88] It is widely seen as a resource-consuming scenario where no useful knowledge is added,[89] and criticized as creating a competitive[90] and conflict-based editing culture associated with traditional masculine gender roles.[91][92] Research has focused on, for example, impoliteness of disputes,[93][94] the influence of rival editing camps,[95][96] the conversational structure,[97] and the shift in conflicts to a focus on sources.[98][99]

Taha Yasseri of the University of Oxford examined editing conflicts and their resolution in a 2013 study.[100][101] Yasseri contended that simple reverts or "undo" operations were not the most significant measure of counterproductive work behavior at Wikipedia. He relied instead on "mutually reverting edit pairs", where one editor reverts the edit of another editor who then, in sequence, returns to revert the first editor. The results were tabulated for several language versions of Wikipedia. The English Wikipedia's three largest conflict rates belonged to the articles George W. Bush, anarchism, and Muhammad.[101] By comparison, for the German Wikipedia, the three largest conflict rates at the time of the study were for the articles covering Croatia, Scientology, and 9/11 conspiracy theories.[101] In 2020, researchers identified other measures of editor behaviors, beyond mutual reverts, to identify editing conflicts across Wikipedia.[99]

Editors also debate the deletion of articles on Wikipedia, with roughly 500,000 such debates since Wikipedia's inception. Once an article is nominated for deletion, the dispute is typically determined by initial votes (to keep or delete) and by reference to topic-specific notability policies.[102]

Policies and content

[edit]
External videos
video icon Jimmy Wales, The Birth of Wikipedia, 2006, TED talks, 20 minutes
video icon Katherine Maher, What Wikipedia Teaches Us About Balancing Truth and Beliefs, 2022, TED talks, 15 minutes

Wikipedia is composed of 11 different namespaces, with its articles being present in mainspace. Other namespaces have a prefix before their page title and fulfill various purposes. For example, the project namespace uses the Wikipedia prefix and is used for self-governance related discussions. Most readers are not aware of these other namespaces.[103]

The fundamental principles of the Wikipedia community are embodied in the "Five pillars", while the detailed editorial principles are expressed in numerous policies and guidelines intended to appropriately shape content.[W 26] The five pillars are:

  • Wikipedia is an encyclopedia
  • Wikipedia is written from a neutral point of view
  • Wikipedia is free content that anyone can use, edit, and distribute
  • Wikipedia's editors should treat each other with respect and civility
  • Wikipedia has no firm rules

The rules developed by the community are stored in wiki form, and Wikipedia editors write and revise the website's policies and guidelines in accordance with community consensus.[104] Originally, rules on the non-English editions of Wikipedia were based on a translation of the rules for the English Wikipedia. They have since diverged to some extent.[W 21]

Content policies and guidelines

[edit]

According to the rules on the English Wikipedia community, each entry in Wikipedia must be about a topic that is encyclopedic and is not a dictionary entry or dictionary-style.[W 27] A topic should also meet Wikipedia's standards of "notability", which generally means that the topic has been covered extensively in reliable sources that are independent of the article's subject.[105] Wikipedia intends to convey only knowledge that is already established and recognized and therefore must not present original research.[106] Some subjects such as politicians and academics have specialized notability requirements.[105] Finally, Wikipedia must reflect a neutral point of view. This is accomplished through summarizing reliable sources, using impartial language, and ensuring that multiple points of view are presented based on their prominence. Information must also be verifiable.[107] Information without citations may be tagged or removed entirely.[108] This can at times lead to the removal of information which, though valid, is not properly sourced.[109] As Wikipedia policies changed over time, and became more complex, their number has grown. In 2008, there were 44 policy pages and 248 guideline pages; by 2013, scholars counted 383 policy pages and 449 guideline pages.[70]

Governance

[edit]

Wikipedia's initial anarchy integrated democratic and hierarchical elements over time.[110][111] An article is not considered to be owned by its creator or any other editor, nor by the subject of the article.[W 28] Editors in good standing in the community can request extra user rights, granting them the technical ability to perform certain special actions. Some user rights are granted automatically, such as the autoconfirmed and extended confirmed groups, when thresholds for account age and edits are met.[68]

Administrators

[edit]

Experienced editors can choose to run for "adminship",[112] which includes the ability to delete pages or prevent them from being changed in cases of severe vandalism or editorial disputes.[W 29] Administrators are not supposed to enjoy any special privilege in decision-making; instead, their powers are mostly limited to making edits that have project-wide effects and thus are disallowed to ordinary editors, and to implement restrictions intended to prevent disruptive editors from making unproductive edits.[W 29]

By 2012, fewer editors were becoming administrators compared to Wikipedia's earlier years, in part because the process of vetting potential administrators had become more rigorous.[36] In 2022, there was a particularly contentious request for adminship over the candidate's anti-Trump views; ultimately, they were granted adminship.[113]

Dispute resolution

[edit]

Over time, Wikipedia has developed a semi-formal dispute resolution process. To determine community consensus, editors can raise issues at appropriate community forums, seek outside input through third opinion requests, or initiate a more general community discussion known as a "request for comment",[W 25] in which bots add the discussion to a centralized list of discussions, invite editors to participate, and remove the discussion from the list after 30 days.[W 30] However, editors have the discretion to close (and delist) the discussion early or late. If the result of a discussion is not obvious, a closer—an uninvolved editor usually in good standing—may render a verdict from the strength of the arguments presented and then the numbers of arguers on each side.[114] Wikipedians emphasize that the process is not a vote by referring to statements of opinion in such discussions as "!vote"s, in which the exclamation mark is the symbol for logical negation and pronounced "not".[115]

Wikipedia encourages local resolutions of conflicts, which Jemielniak argues is quite unique in organization studies, though there has been some recent interest in consensus building in the field.[116] Reagle and Sue Gardner argue that the approaches to consensus building are similar to those used by Quakers.[116]: 62  A difference from Quaker meetings is the absence of a facilitator in the presence of disagreement, a role played by the clerk in Quaker meetings.[116]: 83 

Arbitration Committee

[edit]

The Arbitration Committee presides over the ultimate dispute resolution process. Although disputes usually arise from a disagreement between two opposing views on how an article should read, the Arbitration Committee explicitly refuses to directly rule on the specific view that should be adopted.[117]

Statistical analyses suggest that the English Wikipedia committee ignores the content of disputes and rather focuses on the way disputes are conducted,[118] functioning not so much to resolve disputes and make peace between conflicting editors, but to weed out problematic editors while allowing potentially productive editors back in to participate.[117] Therefore, the committee does not dictate the content of articles, although it sometimes condemns content changes when it deems the new content violates Wikipedia policies (for example, if the new content is considered biased).[f] Commonly used solutions include cautions and probations (used in 63% of cases) and banning editors from articles (43%), subject matters (23%), or Wikipedia (16%).[117] Complete bans from Wikipedia are generally limited to instances of impersonation and antisocial behavior.[W 31] When conduct is not impersonation or anti-social, but rather edit warring and other violations of editing policies, solutions tend to be limited to warnings.[117]

Community

[edit]
Video of Wikimania 2005 – an annual conference for users of Wikipedia and other projects operated by the Wikimedia Foundation, was held in Frankfurt am Main, Germany, August 4–8.

Each article and each user of Wikipedia has an associated and dedicated "talk" page. These form the primary communication channel for editors to discuss, coordinate and debate.[119] Wikipedia's community has been described as cultlike,[120] although not always with entirely negative connotations.[121] Its preference for cohesiveness, even if it requires compromise that includes disregard of credentials, has been referred to as "anti-elitism".[W 32]

Wikipedians and British Museum curators collaborate on the article Hoxne Hoard in June 2010.

Wikipedia does not require that its editors and contributors provide identification.[122] As Wikipedia grew, "Who writes Wikipedia?" became one of the questions frequently asked there.[123] Jimmy Wales once argued that only "a community ... a dedicated group of a few hundred volunteers" makes the bulk of contributions to Wikipedia and that the project is therefore "much like any traditional organization".[124] Since Wikipedia relies on volunteer labour, editors frequently focus on topics that interest them.[125]

The English Wikipedia has 7,080,594 articles, 49,866,679 registered editors, and 114,046 active editors. An editor is considered active if they have made one or more edits in the past 30 days.[W 33] Editors who fail to comply with Wikipedia cultural rituals, such as signing talk page comments, may implicitly signal that they are Wikipedia outsiders, increasing the odds that Wikipedia insiders may target or discount their contributions. Becoming a Wikipedia insider involves non-trivial costs: the contributor is expected to learn Wikipedia-specific technological codes, submit to a sometimes convoluted dispute resolution process, and learn a "baffling culture rich with in-jokes and insider references".[126] Editors who do not log in are in some sense "second-class citizens" on Wikipedia,[126] as "participants are accredited by members of the wiki community, who have a vested interest in preserving the quality of the work product, on the basis of their ongoing participation",[127] but the contribution histories of anonymous unregistered editors recognized only by their IP addresses cannot be attributed to a particular editor with certainty.[127] New editors often struggle to understand Wikipedia's complexity. Experienced editors are encouraged to not "bite" the newcomers in order to create a more welcoming atmosphere.[128]

Research

[edit]

A 2007 study by researchers from Dartmouth College found that "anonymous and infrequent contributors to Wikipedia ... are as reliable a source of knowledge as those contributors who register with the site".[129] Jimmy Wales stated in 2009 that "[I]t turns out over 50% of all the edits are done by just 0.7% of the users ... 524 people ... And in fact, the most active 2%, which is 1400 people, have done 73.4% of all the edits."[124] However, Business Insider editor and journalist Henry Blodget showed in 2009 that in a random sample of articles, most Wikipedia content (measured by the amount of contributed text that survives to the latest sampled edit) is created by "outsiders", while most editing and formatting is done by "insiders".[124]

In 2008, a Slate magazine article reported that "one percent of Wikipedia users are responsible for about half of the site's edits."[130] This method of evaluating contributions was later disputed by Aaron Swartz, who noted that several articles he sampled had large portions of their content (measured by number of characters) contributed by users with low edit counts.[131] A 2008 study found that Wikipedians were less agreeable, open, and conscientious than others,[132] although a later commentary pointed out serious flaws, including that the data showed higher openness and that the differences with the control group and the samples were small.[133] According to a 2009 study, there is "evidence of growing resistance from the Wikipedia community to new content".[134]

Diversity

[edit]

Several studies have shown that most volunteer Wikipedia contributors are male. Notably, the results of a Wikimedia Foundation survey in 2008 showed that only 13 percent of Wikipedia editors were female.[135] Because of this, universities throughout the United States tried to encourage women to become Wikipedia contributors.[136] Similarly, many of these universities, including Yale and Brown, gave college credit to students who create or edit an article relating to women in science or technology.[136] Andrew Lih, a professor and scientist, said that the reason he thought the number of male contributors outnumbered the number of females so greatly was because identifying as a woman may expose oneself to "ugly, intimidating behavior".[137] Data has shown that Africans are underrepresented among Wikipedia editors.[138]

Language editions

[edit]
Distribution of the 65,799,298 articles in different language editions (as of October 26, 2025)[W 34]
  1. English (10.8%)
  2. Cebuano (9.30%)
  3. German (4.70%)
  4. French (4.10%)
  5. Swedish (4.00%)
  6. Dutch (3.30%)
  7. Spanish (3.10%)
  8. Russian (3.10%)
  9. Italian (3.00%)
  10. Polish (2.50%)
  11. Egyptian Arabic (2.50%)
  12. Chinese (2.30%)
  13. Japanese (2.20%)
  14. Ukrainian (2.10%)
  15. Vietnamese (2.00%)
  16. Arabic (2.00%)
  17. Waray (1.90%)
  18. Portuguese (1.90%)
  19. Persian (1.60%)
  20. Catalan (1.20%)
  21. Other (32.4%)

There are currently 343 language editions of Wikipedia (also called language versions, or simply Wikipedias). As of October 2025, the six largest, in order of article count, are the English, Cebuano, German, French, Swedish, and Dutch Wikipedias.[W 35] The second and fifth-largest Wikipedias owe their position to the article-creating bot Lsjbot, which as of 2013 had created about half the articles on the Swedish Wikipedia, and most of the articles in the Cebuano and Waray Wikipedias. The latter are both languages of the Philippines.

In addition to the top six, twelve other Wikipedias have more than a million articles each (Spanish, Russian, Italian, Polish, Egyptian Arabic, Chinese, Japanese, Ukrainian, Vietnamese, Arabic, Waray, and Portuguese), seven more have over 500,000 articles (Persian, Catalan, Indonesian, Korean, Serbian, Norwegian, and Chechen), 44 more have over 100,000, and 82 more have over 10,000.[W 36][W 35] The largest, the English Wikipedia, has over 7 million articles. As of January 2021, the English Wikipedia receives 48% of Wikipedia's cumulative traffic, with the remaining split among the other languages. The top 10 editions represent approximately 85% of the total traffic.[W 37]

Articles in the 20 largest language editions of Wikipedia
(as of 26 October 2025)[139]
1 2 3 4 5 6 7

English 7,080,594
Cebuano 6,115,933
German 3,063,281
French 2,716,855
Swedish 2,618,665
Dutch 2,200,420
Spanish 2,070,448
Russian 2,068,752
Italian 1,941,477
Polish 1,672,925
Egyptian Arabic 1,629,439
Chinese 1,507,208
Japanese 1,477,789
Ukrainian 1,395,233
Vietnamese 1,296,664
Arabic 1,283,907
Waray 1,266,807
Portuguese 1,158,429
Persian 1,059,705
Catalan 782,926

Since Wikipedia is based on the Web and therefore worldwide, contributors to the same language edition may use different dialects or may come from different countries (as is the case for the English edition). These differences may lead to some conflicts over spelling differences (e.g. colour versus color)[W 38] or points of view.[W 39]

Though the various language editions are held to global policies such as "neutral point of view", they diverge on some points of policy and practice, most notably on whether images that are not licensed freely may be used under a claim of fair use.[W 40][140] The content of articles on the same subject can differ significantly between languages, depending on the sources editors use and other factors.[141][142]

Jimmy Wales has described Wikipedia as "an effort to create and distribute a free encyclopedia of the highest possible quality to every single person on the planet in their own language".[W 41] Though each language edition functions more or less independently, some efforts are made to supervise them all. They are coordinated in part by Meta-Wiki, the Wikimedia Foundation's wiki devoted to maintaining all its projects (Wikipedia and others).[W 42] For instance, Meta-Wiki provides important statistics on all language editions of Wikipedia,[W 43] and it maintains a list of articles every Wikipedia should have.[W 44] The list concerns basic content by subject: biography, history, geography, society, culture, science, technology, and mathematics.[W 44] It is not rare for articles strongly related to a particular language not to have counterparts in another edition. For example, articles about small towns in the United States might be available only in English, even when they meet the notability criteria of other language Wikipedia projects.[W 45]

Estimation of contributions shares from different regions in the world to different Wikipedia editions[143]

Translated articles represent only a small portion of articles in most editions, in part because those editions do not allow fully automated translation of articles. Articles available in more than one language may offer "interwiki links", which link to the counterpart articles in other editions.[144][W 46]

A study published by PLOS One in 2012 also estimated the share of contributions to different editions of Wikipedia from different regions of the world. It reported that the proportion of the edits made from North America was 51% for the English Wikipedia, and 25% for the Simple English Wikipedia.[143]

English Wikipedia editor numbers

[edit]

On March 1, 2014, The Economist, in an article titled "The Future of Wikipedia", cited a trend analysis concerning data published by the Wikimedia Foundation stating that "the number of editors for the English-language version has fallen by a third in seven years."[145] The attrition rate for active editors in English Wikipedia was cited by The Economist as substantially in contrast to statistics for Wikipedia in other languages (non-English Wikipedia). The Economist reported that the number of contributors with an average of five or more edits per month was relatively constant since 2008 for Wikipedia in other languages at approximately 42,000 editors within narrow seasonal variances of about 2,000 editors up or down. The number of active editors in English Wikipedia, by sharp comparison, was cited as peaking in 2007 at approximately 50,000 and dropping to 30,000 by the start of 2014.[145]

In contrast, the trend analysis for Wikipedia in other languages (non-English Wikipedia) shows success in retaining active editors on a renewable and sustained basis, with their numbers remaining relatively constant at approximately 42,000. No comment was made concerning which of the differentiated edit policy standards from Wikipedia in other languages (non-English Wikipedia) would provide a possible alternative to English Wikipedia for effectively improving substantial editor attrition rates on the English-language Wikipedia.[145]

Reception

[edit]

Various Wikipedians have criticized Wikipedia's large and growing regulation, which includes more than fifty policies and nearly 150,000 words as of 2014.[146][116] Critics have stated that Wikipedia exhibits systemic bias. In 2010, columnist and journalist Edwin Black described Wikipedia as being a mixture of "truth, half-truth, and some falsehoods".[147] Articles in The Chronicle of Higher Education and The Journal of Academic Librarianship have criticized Wikipedia's "undue-weight policy", concluding that Wikipedia explicitly is not designed to provide correct information about a subject, but rather focus on all the major viewpoints on the subject, give less attention to minor ones, and creates omissions that can lead to false beliefs based on incomplete information.[148][149][150]

Journalists Oliver Kamm and Edwin Black alleged (in 2010 and 2011 respectively) that articles are dominated by the loudest and most persistent voices, usually by a group with an "ax to grind" on the topic.[147][151] A 2008 article in Education Next journal concluded that as a resource about controversial topics, Wikipedia is subject to manipulation and spin.[152] In 2020, Omer Benjakob and Stephen Harrison noted that "Media coverage of Wikipedia has radically shifted over the past two decades: once cast as an intellectual frivolity, it is now lauded as the 'last bastion of shared reality' online."[153]

Multiple news networks and pundits have accused Wikipedia of being ideologically biased. In February 2021, Fox News accused Wikipedia of whitewashing communism and socialism and having too much "leftist bias".[154] Wikipedia co-founder Sanger said that Wikipedia has become a "propaganda" for the left-leaning "establishment" and warned the site can no longer be trusted.[155] In 2022, libertarian John Stossel opined that Wikipedia, a site he financially supported at one time, appeared to have gradually taken a significant turn in bias to the political left, specifically on political topics.[156] Some studies suggest that Wikipedia (and in particular the English Wikipedia) has a "western cultural bias" (or "pro-western bias")[157] or "Eurocentric bias",[158] reiterating, says Anna Samoilenko, "similar biases that are found in the 'ivory tower' of academic historiography". Carwil Bjork-James proposes that Wikipedia could follow the diversification pattern of contemporary scholarship[159] and Dangzhi Zhao calls for a "decolonization" of Wikipedia to reduce bias from opinionated White male editors.[160]

Accuracy of content

[edit]
External audio
audio icon The Great Book of Knowledge, Part 1, Ideas with Paul Kennedy, CBC, January 15, 2014

Articles for traditional encyclopedias such as Encyclopædia Britannica are written by experts, lending such encyclopedias a reputation for accuracy.[161] However, a peer review in 2005 of forty-two scientific entries on both Wikipedia and Encyclopædia Britannica by the science journal Nature found few differences in accuracy, and concluded that "the average science entry in Wikipedia contained around four inaccuracies; Britannica, about three."[162] Joseph Reagle suggested that while the study reflects "a topical strength of Wikipedia contributors" in science articles, "Wikipedia may not have fared so well using a random sampling of articles or on humanities subjects."[163]

Others raised similar critiques.[164] The findings by Nature were disputed by Encyclopædia Britannica,[165][166] and in response, Nature gave a rebuttal of the points raised by Britannica.[167] In addition to the point-for-point disagreement between these two parties, others have examined the sample size and selection method used in the Nature effort, and suggested a "flawed study design" (in Nature's manual selection of articles, in part or in whole, for comparison), absence of statistical analysis (e.g., of reported confidence intervals), and a lack of study "statistical power" (i.e., owing to small sample size, 42 or 4 × 101 articles compared, vs >105 and >106 set sizes for Britannica and the English Wikipedia, respectively).[168]

As a consequence of the open structure, Wikipedia "makes no guarantee of validity" of its content, since no one is ultimately responsible for any claims appearing in it.[W 47] Concerns have been raised by PC World in 2009 regarding the lack of accountability that results from users' anonymity, the insertion of false information,[169] vandalism, and similar problems. Legal Research in a Nutshell (2011), cites Wikipedia as a "general source" that "can be a real boon" in "coming up to speed in the law governing a situation" and, "while not authoritative, can provide basic facts as well as leads to more in-depth resources".[170]

Economist Tyler Cowen wrote: "If I had to guess whether Wikipedia or the median refereed journal article on economics was more likely to be true after a not so long think I would opt for Wikipedia." He comments that some traditional sources of non-fiction suffer from systemic biases, and novel results, in his opinion, are over-reported in journal articles as well as relevant information being omitted from news reports. However, he also cautions that errors are frequently found on Internet sites and that academics and experts must be vigilant in correcting them.[171] Amy Bruckman has argued that, due to the number of reviewers, "the content of a popular Wikipedia page is actually the most reliable form of information ever created".[172] In September 2022, The Sydney Morning Herald journalist Liam Mannix noted that: "There's no reason to expect Wikipedia to be accurate ... And yet it [is]." Mannix further discussed the multiple studies that have proved Wikipedia to be generally as reliable as Encyclopædia Britannica, summarizing that "...turning our back on such an extraordinary resource is... well, a little petty."[173]

Critics argue that Wikipedia's open nature and a lack of proper sources for most of the information makes it unreliable.[174] Some commentators suggest that Wikipedia may be reliable, but that the reliability of any given article is not clear.[175] Editors of traditional reference works such as the Encyclopædia Britannica have questioned the project's utility and status as an encyclopedia.[176] Wikipedia co-founder Jimmy Wales has claimed that Wikipedia has largely avoided the problem of "fake news" because the Wikipedia community regularly debates the quality of sources in articles.[177]

External videos
video icon Inside Wikipedia – Attack of the PR Industry, Deutsche Welle, 7:13 mins[178]

Wikipedia's open structure inherently makes it an easy target for Internet trolls, spammers, and various forms of paid advocacy seen as counterproductive to the maintenance of a neutral and verifiable online encyclopedia.[79][W 48] In response to paid advocacy editing and undisclosed editing issues, Wikipedia was reported in an article in The Wall Street Journal to have strengthened its rules and laws against undisclosed editing.[179] The article stated that: "Beginning Monday [from the date of the article, June 16, 2014], changes in Wikipedia's terms of use will require anyone paid to edit articles to disclose that arrangement. Katherine Maher, the nonprofit Wikimedia Foundation's chief communications officer, said the changes address a sentiment among volunteer editors that 'we're not an advertising service; we're an encyclopedia.'"[179][180][181][182][183] These issues, among others, had been parodied since the first decade of Wikipedia, notably by Stephen Colbert on The Colbert Report.[184]

Discouragement in education

[edit]

Some university lecturers discourage students from citing any encyclopedia in academic work, preferring primary sources;[185] some specifically prohibit Wikipedia citations.[186][187] Wales stresses that encyclopedias of any type are not usually appropriate to use as citable sources, and should not be relied upon as authoritative.[188] Wales once (2006 or earlier) said he receives about ten emails weekly from students saying they got failing grades on papers because they cited Wikipedia; he told the students they got what they deserved. "For God's sake, you're in college; don't cite the encyclopedia", he said.[189]

In February 2007, an article in The Harvard Crimson newspaper reported that a few of the professors at Harvard University were including Wikipedia articles in their syllabi, although without realizing the articles might change.[190] In June 2007, Michael Gorman, former president of the American Library Association, condemned Wikipedia, along with Google, stating that academics who endorse the use of Wikipedia are "the intellectual equivalent of a dietitian who recommends a steady diet of Big Macs with everything".[191]

A 2020 research study published in Studies in Higher Education argued that Wikipedia could be applied in the higher education "flipped classroom", an educational model where students learn before coming to class and apply it in classroom activities. The experimental group was instructed to learn before class and get immediate feedback before going in (the flipped classroom model), while the control group was given direct instructions in class (the conventional classroom model). The groups were then instructed to collaboratively develop Wikipedia entries, which would be graded in quality after the study. The results showed that the experimental group yielded more Wikipedia entries and received higher grades in quality. The study concluded that learning with Wikipedia in flipped classrooms was more effective than in conventional classrooms, demonstrating Wikipedia could be used as an educational tool in higher education.[192]

Medical information

[edit]

On March 5, 2014, Julie Beck writing for The Atlantic magazine in an article titled "Doctors' #1 Source for Healthcare Information: Wikipedia", stated that "Fifty percent of physicians look up conditions on the (Wikipedia) site, and some are editing articles themselves to improve the quality of available information."[193] Beck continued to detail in this article new programs of Amin Azzam at the University of San Francisco to offer medical school courses to medical students for learning to edit and improve Wikipedia articles on health-related issues, as well as internal quality control programs within Wikipedia organized by James Heilman to improve a group of 200 health-related articles of central medical importance up to Wikipedia's highest standard of articles using its Featured Article and Good Article peer-review evaluation process.[193] In a May 7, 2014, follow-up article in The Atlantic titled "Can Wikipedia Ever Be a Definitive Medical Text?", Julie Beck quotes WikiProject Medicine's James Heilman as stating: "Just because a reference is peer-reviewed doesn't mean it's a high-quality reference."[194] Beck added that: "Wikipedia has its own peer review process before articles can be classified as 'good' or 'featured'. Heilman, who has participated in that process before, says 'less than one percent' of Wikipedia's medical articles have passed."[194]

Coverage of topics and systemic bias

[edit]

Wikipedia seeks to create a summary of all human knowledge in the form of an online encyclopedia, with each topic covered encyclopedically in one article. Since it has terabytes of disk space, it can have far more topics than can be covered by any printed encyclopedia.[W 49] The exact degree and manner of coverage on Wikipedia is under constant review by its editors, and disagreements are not uncommon (see deletionism and inclusionism).[195][196] Wikipedia contains materials that some people may find objectionable, offensive, or pornographic.[W 50] The "Wikipedia is not censored" policy has sometimes proved controversial: in 2008, Wikipedia rejected an online petition against the inclusion of images of Muhammad in the English edition of its Muhammad article, citing this policy.[197] The presence of politically, religiously, and pornographically sensitive materials in Wikipedia has led to the censorship of Wikipedia by national authorities in China[198] and Pakistan,[199] among other countries.[200][201][202]

Through its "Wikipedia Loves Libraries" program, Wikipedia has partnered with major public libraries such as the New York Public Library for the Performing Arts to expand its coverage of underrepresented subjects and articles.[203] A 2011 study conducted by researchers at the University of Minnesota indicated that male and female editors focus on different coverage topics. There was a greater concentration of females in the "people and arts" category, while males focus more on "geography and science".[204] An editorial in The Guardian in 2014 claimed that more effort went into providing references for a list of female porn actors than a list of women writers.[205]

Systemic biases

[edit]

Wikipedia's policies may limit "its capacity for truly representing global knowledge". For example, Wikipedia only considers published sources to be reliable. Oral knowledge of Indigenous cultures is not always reflected in print. Marginalized topics are also more likely to lack significant coverage in reliable sources. Wikipedia's content is therefore limited as a result of larger systemic biases.[206]

Academic studies of Wikipedia have shown that the average contributor to the English Wikipedia is an educated, technically inclined white male, aged 15–49, from a developed, predominantly Christian country.[207] The corresponding point of view (POV) is over-represented.[208][159] This systemic bias in editor demographic results in cultural bias, gender bias, and geographical bias on Wikipedia.[209][210] There are two broad types of bias, which are implicit (when a topic is omitted) and explicit (when a certain POV is over-represented in an article or by references).[208]

Interdisciplinary scholarly assessments of Wikipedia articles have found that while articles are typically accurate and free of misinformation, they are also typically incomplete and fail to present all perspectives with a neutral point of view.[209] In 2011, Wales claimed that the unevenness of coverage is a reflection of the demography of the editors, citing for example "biographies of famous women through history and issues surrounding early childcare".[34] The October 22, 2013, essay by Tom Simonite in MIT's Technology Review titled "The Decline of Wikipedia" discussed the effect of systemic bias and policy creep on the downward trend in the number of editors.[35]

Research conducted by Mark Graham of the Oxford Internet Institute in 2009 indicated that the geographic distribution of article topics is highly uneven, with Africa being the most underrepresented.[211] Across 30 language editions of Wikipedia, historical articles and sections are generally Eurocentric and focused on recent events.[212]

Explicit content

[edit]

Wikipedia has been criticized for allowing information about graphic content.[213] Articles depicting what some critics have called objectionable content (such as feces, cadaver, human penis, vulva, and nudity) contain graphic pictures and detailed information easily available to anyone with access to the internet, including children.[W 51] The site also includes sexual content such as images and videos of masturbation and ejaculation, illustrations of zoophilia, and photos from hardcore pornographic films in its articles. It also has non-sexual photographs of nude children.[W 52]

The Wikipedia article about Virgin Killer—a 1976 album from the German rock band Scorpions—features a picture of the album's original cover, which depicts a naked prepubescent girl. The original release cover caused controversy and was replaced in some countries. In December 2008, access to the Wikipedia article Virgin Killer was blocked for four days by most Internet service providers in the United Kingdom after the Internet Watch Foundation (IWF) decided the album cover was a potentially illegal indecent image and added the article's URL to a "blacklist" it supplies to British internet service providers.[214]

In April 2010, Sanger wrote a letter to the Federal Bureau of Investigation, outlining his concerns that two categories of images on Wikimedia Commons contained child pornography, and were in violation of US federal obscenity law.[215][216] Sanger later clarified that the images, which were related to pedophilia and one about lolicon, were not of real children, but said that they constituted "obscene visual representations of the sexual abuse of children", under the PROTECT Act of 2003.[217] That law bans photographic child pornography and cartoon images and drawings of children that are obscene under American law.[217] Sanger also expressed concerns about access to the images on Wikipedia in schools.[218]

Wikimedia Foundation spokesman Jay Walsh strongly rejected Sanger's accusation,[219] saying that Wikipedia did not have "material we would deem to be illegal. If we did, we would remove it."[219] Following the complaint by Sanger, Wales deleted sexual images without consulting the community. After some editors who volunteered to maintain the site argued that the decision to delete had been made hastily, Wales voluntarily gave up some of the powers he had held up to that time as part of his co-founder status. He wrote in a message to the Wikimedia Foundation mailing-list that this action was "in the interest of encouraging this discussion to be about real philosophical/content issues, rather than be about me and how quickly I acted".[220] Critics, including Wikipediocracy, noticed that many of the pornographic images deleted from Wikipedia since 2010 have reappeared.[221]

Privacy

[edit]

One privacy concern in the case of Wikipedia regards one's right to remain a private citizen rather than a public figure in the eyes of the law.[222][g] It is a battle between the right to be anonymous in cyberspace and the right to be anonymous in real life. The Wikimedia Foundation's privacy policy states, "we believe that you shouldn't have to provide personal information to participate in the free knowledge movement", and states that "personal information" may be shared "For legal reasons", "To Protect You, Ourselves & Others", or "To Understand & Experiment".[W 53]

In January 2006, a German court ordered the German Wikipedia shut down within Germany because it stated the full name of Boris Floricic, aka "Tron", a deceased hacker. On February 9, 2006, the injunction against Wikimedia Deutschland was overturned, with the court rejecting the notion that Tron's right to privacy or that of his parents was being violated.[223]

Wikipedia has a "Volunteer Response Team" that uses Znuny, a free and open-source software fork of OTRS[W 54] to handle queries without having to reveal the identities of the involved parties. This is used, for example, in confirming the permission for using individual images and other media in the project.[W 55]

In late April 2023, Wikimedia Foundation announced that Wikipedia will not submit to any age verifications that may be required by the UK's Online Safety Bill legislation. Rebecca MacKinnon of the Wikimedia Foundation said that such checks would run counter to the website's commitment to minimal data collection on its contributors and readers.[224]

Sexism

[edit]

Wikipedia was described in 2015 as harboring a battleground culture of sexism and harassment.[225][226] The perceived tolerance of abusive language was a reason put forth in 2013 for the gender gap in Wikipedia editorship.[227] Edit-a-thons have been held to encourage female editors and increase the coverage of women's topics.[228]

In May 2018, a Wikipedia editor rejected a submitted article about Donna Strickland due to lack of coverage in the media.[W 56][229] Five months later, Strickland won a Nobel Prize in Physics "for groundbreaking inventions in the field of laser physics", becoming the third woman to ever receive the award.[229][230] Prior to winning the award, Strickland's only mention on Wikipedia was in the article about her collaborator and co-winner of the award Gérard Mourou.[229] Her exclusion from Wikipedia led to accusations of sexism, but Corinne Purtill writing for Quartz argued that "it's also a pointed lesson in the hazards of gender bias in media, and of the broader consequences of underrepresentation."[231] Purtill attributes the issue to the gender bias in media coverage.[231]

A comprehensive 2008 survey, published in 2016, by Julia B. Bear of Stony Brook University's College of Business and Benjamin Collier of Carnegie Mellon University found significant gender differences in confidence in expertise, discomfort with editing, and response to critical feedback. "Women reported less confidence in their expertise, expressed greater discomfort with editing (which typically involves conflict), and reported more negative responses to critical feedback compared to men."[232]

Operation

[edit]

Wikimedia Foundation and affiliate movements

[edit]
Katherine Maher in 2016. She is seen with light skin, blonde hair, and blue eyes. She is seen wearing a black shirt.
Katherine Maher, the third executive director of Wikimedia, served from 2016 to 2021.

Wikipedia is hosted and funded by the Wikimedia Foundation, a non-profit organization which also operates Wikipedia-related projects such as Wiktionary and Wikibooks.[W 57] The foundation relies on public contributions and grants to fund its mission.[233][W 58] The foundation's 2020 Internal Revenue Service Form 990 shows revenue of $124.6 million and expenses of almost $112.2 million, with assets of about $191.2 million and liabilities of almost $11 million.[W 59]

In May 2014, Wikimedia Foundation named Lila Tretikov as its second executive director, taking over for Sue Gardner.[W 60] The Wall Street Journal reported on May 1, 2014, that Tretikov's information technology background, from her years at University of California offers Wikipedia an opportunity to develop in more concentrated directions guided by her often repeated position statement that, "Information, like air, wants to be free."[234][235] The same Wall Street Journal article reported these directions of development according to an interview with spokesman Jay Walsh of Wikimedia, who "said Tretikov would address that issue (paid advocacy) as a priority. 'We are really pushing toward more transparency ... We are reinforcing that paid advocacy is not welcome.' Initiatives to involve greater diversity of contributors, better mobile support of Wikipedia, new geo-location tools to find local content more easily, and more tools for users in the second and third world are also priorities", Walsh said.[234]

Following the departure of Tretikov from Wikipedia due to issues concerning the use of the "superprotection" feature which some language versions of Wikipedia have adopted,[W 61] Katherine Maher became the third executive director of the Wikimedia Foundation in June 2016.[W 62] Maher stated that one of her priorities would be the issue of editor harassment endemic to Wikipedia as identified by the Wikipedia board in December. She said to Bloomberg Businessweek regarding the harassment issue that: "It establishes a sense within the community that this is a priority ... [and that correction requires that] it has to be more than words."[137]

Maher served as executive director until April 2021.[236] Maryana Iskander was named the incoming CEO in September 2021, and took over that role in January 2022. She stated that one of her focuses would be increasing diversity in the Wikimedia community.[237]

Wikipedia is also supported by many organizations and groups that are affiliated with the Wikimedia Foundation but independently-run, called Wikimedia movement affiliates. These include Wikimedia chapters (which are national or sub-national organizations, such as Wikimedia Deutschland and Wikimedia France), thematic organizations (such as Amical Wikimedia for the Catalan language community), and user groups. These affiliates participate in the promotion, development, and funding of Wikipedia.[W 63]

Software operations and support

[edit]

The operation of Wikipedia depends on MediaWiki, a custom-made, free and open source wiki software platform written in PHP and built upon the MySQL database system.[W 64] The software incorporates programming features such as a macro language, variables, a transclusion system for templates, and URL redirection.[W 65] MediaWiki is licensed under the GNU General Public License (GPL) and it is used by all Wikimedia projects, as well as many other wiki projects.[W 64][W 66] Originally, Wikipedia ran on UseModWiki written in Perl by Clifford Adams (Phase I), which initially required CamelCase for article hyperlinks; the present double bracket style was incorporated later.[W 67] Starting in January 2002 (Phase II), Wikipedia began running on a PHP wiki engine with a MySQL database; this software was custom-made for Wikipedia by Magnus Manske. The Phase II software was repeatedly modified to accommodate the exponentially increasing demand. In July 2002 (Phase III), Wikipedia shifted to the third-generation software, MediaWiki, originally written by Lee Daniel Crocker.

Several MediaWiki extensions are installed to extend the functionality of the MediaWiki software.[W 68] In April 2005, a Lucene extension[W 69][W 70] was added to MediaWiki's built-in search and Wikipedia switched from MySQL to Lucene for searching. Lucene was later replaced by CirrusSearch which is based on Elasticsearch.[W 71] In July 2013, after extensive beta testing, a WYSIWYG (What You See Is What You Get) extension, VisualEditor, was opened to public use.[238][239][240] It was met with much rejection and criticism, and was described as "slow and buggy".[241] The feature was changed from opt-out to opt-in afterward.[W 72]

Automated editing

[edit]

Computer programs called bots have often been used to perform simple and repetitive tasks, such as correcting common misspellings and stylistic issues, or to start articles such as geography entries in a standard format from statistical data.[W 73][242][243] One controversial contributor, Sverker Johansson, created articles with his bot Lsjbot, which was reported to create up to 10,000 articles on the Swedish Wikipedia on certain days.[244] Additionally, there are bots designed to automatically notify editors when they make common editing errors (such as unmatched quotes or unmatched parentheses).[W 74] Edits falsely identified by bots as the work of a banned editor can be restored by other editors. An anti-vandal bot is programmed to detect and revert vandalism quickly.[242] Bots are able to indicate edits from particular accounts or IP address ranges, as occurred at the time of the shooting down of the MH17 jet in July 2014 when it was reported that edits were made via IPs controlled by the Russian government.[245] Bots on Wikipedia must be approved before activation.[W 75] According to Andrew Lih, the current expansion of Wikipedia to millions of articles would be difficult to envision without the use of such bots.[246]

Hardware operations and support

[edit]
Diagram showing flow of data between Wikipedia's servers.
Overview of system architecture as of August 2022

As of 2021, page requests are first passed to a front-end layer of Varnish caching servers and back-end layer caching is done by Apache Traffic Server.[W 76] Requests that cannot be served from the Varnish cache are sent to load-balancing servers running the Linux Virtual Server software, which in turn pass them to one of the Apache web servers for page rendering from the database.[W 76] The web servers deliver pages as requested, performing page rendering for all the language editions of Wikipedia. To increase speed further, rendered pages are cached in a distributed memory cache until invalidated, allowing page rendering to be skipped entirely for most common page accesses.[247]

Wikipedia currently runs on dedicated clusters of Linux servers running the Debian operating system.[W 77] By January 22, 2013, Wikipedia had migrated its primary data center to an Equinix facility in Ashburn, Virginia.[W 78][248] A second application data center was created in 2014 in Carrollton, Texas, to improve Wikipedia's reliability.[249][250] Both datacenters work as the primary one, in alternate semesters, with the other one working as secondary datacenter.[251] In 2017, Wikipedia installed a caching cluster in an Equinix facility in Singapore, the first of its kind in Asia.[W 79] In 2022, a caching data center was opened in Marseille, France.[W 80] In 2024, a caching data center was opened in São Paulo, the first of its kind in South America.[W 81] As of November 2024, caching clusters are located in Amsterdam, San Francisco, Singapore, Marseille, and São Paulo.[W 82][W 83]

Internal research and operational development

[edit]

Following growing amounts of incoming donations in 2013 exceeding seven digits,[35] the Foundation has reached a threshold of assets which qualify its consideration under the principles of industrial organization economics to indicate the need for the re-investment of donations into the internal research and development of the Foundation.[252] Two projects of such internal research and development have been the creation of a Visual Editor and the "Thank" tab in the edit history, which were developed to improve issues of editor attrition.[35][241] The estimates for reinvestment by industrial organizations into internal research and development was studied by Adam Jaffe, who recorded that the range of 4% to 25% annually was to be recommended, with high-end technology requiring the higher level of support for internal reinvestment.[253] At the 2013 level of contributions for Wikimedia presently documented as 45 million dollars,[W 84] the computed budget level recommended by Jaffe for reinvestment into internal research and development is between 1.8 million and 11.3 million dollars annually.[253] In 2019, the level of contributions were reported by the Wikimedia Foundation as being at $120 million annually,[W 85] updating the Jaffe estimates for the higher level of support to between $3.08 million and $19.2 million annually.[253]

Internal news publications

[edit]

Multiple Wikimedia projects have internal news publications. Wikimedia's online newspaper The Signpost was founded in 2005 by Michael Snow, a Wikipedia administrator who would join the Wikimedia Foundation's board of trustees in 2008.[254][255] The publication covers news and events from the English Wikipedia, the Wikimedia Foundation, and Wikipedia's sister projects.[W 86]

The Wikipedia Library

[edit]
Wikipedia Library

Wikipedia editors sometimes struggle to access paywalled sources needed to improve a subject.[256] The Wikipedia Library is a resource for Wikipedia editors which provides free access to a wide range of digital publications, so that they can consult and cite these while editing the encyclopedia.[257][258] Over 60 publishers have partnered with The Wikipedia Library to provide access to their resources: when ICE Publishing joined in 2020, a spokesman said "By enabling free access to our content for Wikipedia editors, we hope to further the research community's resources – creating and updating Wikipedia entries on civil engineering which are read by thousands of monthly readers."[259]

Access to content

[edit]

Content licensing

[edit]

When the project was started in 2001, all text in Wikipedia was covered by the GNU Free Documentation License (GFDL), a copyleft license permitting the redistribution, creation of derivative works, and commercial use of content while authors retain copyright of their work.[W 87] The GFDL was created for software manuals that come with free software programs licensed under the GPL. This made it a poor choice for a general reference work: for example, the GFDL requires the reprints of materials from Wikipedia to come with a full copy of the GFDL text.[260] In December 2002, the Creative Commons license was released; it was specifically designed for creative works in general, not just for software manuals. The Wikipedia project sought the switch to the Creative Commons.[W 88] Because the GFDL and Creative Commons were incompatible, in November 2008, following the request of the project, the Free Software Foundation (FSF) released a new version of the GFDL designed specifically to allow Wikipedia to relicense its content to CC BY-SA by August 1, 2009.[W 89] In April 2009, Wikipedia and its sister projects held a community-wide referendum which decided the switch in June 2009.[W 90][W 91][W 92][W 93]

The handling of media files (e.g. image files) varies across language editions. Some language editions, such as the English Wikipedia, include non-free image files under fair use doctrine,[W 94] while the others have opted not to, in part because of the lack of fair use doctrines in their home countries (e.g. in Japanese copyright law). Media files covered by free content licenses (e.g. Creative Commons' CC BY-SA) are shared across language editions via Wikimedia Commons repository, a project operated by the Wikimedia Foundation.[W 95] Wikipedia's accommodation of varying international copyright laws regarding images has led some to observe that its photographic coverage of topics lags behind the quality of the encyclopedic text.[261] The Wikimedia Foundation is not a licensor of content on Wikipedia or its related projects but merely a hosting service for contributors to and licensors of Wikipedia, a position which was successfully defended in 2004 in a court in France.[262][263]

Methods of access

[edit]

Because Wikipedia content is distributed under an open license, anyone can reuse or re-distribute it at no charge.[W 96] The content of Wikipedia has been published in many forms, both online and offline, outside the Wikipedia website.

Thousands of "mirror sites" exist that republish content from Wikipedia; two prominent ones that also include content from other reference sources are Reference.com and Answers.com.[264][265] Another example is Wapedia, which began to display Wikipedia content in a mobile-device-friendly format before Wikipedia itself did.[W 97] Some web search engines make special use of Wikipedia content when displaying search results: examples include Microsoft Bing (via technology gained from Powerset)[266] and DuckDuckGo.

Collections of Wikipedia articles have been published on optical discs. An English version released in 2006 contained about 2,000 articles.[W 98] The Polish-language version from 2006 contains nearly 240,000 articles,[W 99] the German-language version from 2007/2008 contains over 620,000 articles,[W 100] and the Spanish-language version from 2011 contains 886,000 articles.[W 101] Additionally, "Wikipedia for Schools", the Wikipedia series of CDs / DVDs produced by Wikipedia and SOS Children, is a free selection from Wikipedia designed for education towards children eight to seventeen.[W 102]

There have been efforts to put a select subset of Wikipedia's articles into printed book form.[267][W 103] Since 2009, tens of thousands of print-on-demand books that reproduced English, German, Russian, and French Wikipedia articles have been produced by the American company Books LLC and by three Mauritian subsidiaries of the German publisher VDM.[268]

The website DBpedia, begun in 2007, extracts data from the infoboxes and category declarations of the English-language Wikipedia.[269] Wikimedia has created the Wikidata project with a similar objective of storing the basic facts from each page of Wikipedia and other Wikimedia Foundation projects and make it available in a queryable semantic format, RDF.[W 104] As of February 2023, it has over 101 million items.[W 105] WikiReader is a dedicated reader device that contains an offline copy of Wikipedia, which was launched by OpenMoko and first released in 2009.[W 106]

Obtaining the full contents of Wikipedia for reuse presents challenges, since direct cloning via a web crawler is discouraged.[W 107] Wikipedia publishes "dumps" of its contents, but these are text-only; as of 2023, there is no dump available of Wikipedia's images.[W 108] Wikimedia Enterprise is a for-profit solution to this.[270]

Several languages of Wikipedia also maintain a reference desk, where volunteers answer questions from the general public. According to a study by Pnina Shachaf in the Journal of Documentation, the quality of the Wikipedia reference desk is comparable to a standard library reference desk, with an accuracy of 55 percent.[271]

Mobile access

[edit]
A mobile version showing the English Wikipedia's Main Page on October 2, 2024

Wikipedia's original medium was for users to read and edit content using any standard web browser through a fixed Internet connection. Although Wikipedia content has been accessible through the mobile web since July 2013, The New York Times on February 9, 2014, quoted Erik Möller, deputy director of the Wikimedia Foundation, stating that the transition of internet traffic from desktops to mobile devices was significant and a cause for concern and worry. The article in The New York Times reported the comparison statistics for mobile edits stating that, "Only 20 percent of the readership of the English-language Wikipedia comes via mobile devices, a figure substantially lower than the percentage of mobile traffic for other media sites, many of which approach 50 percent. And the shift to mobile editing has lagged even more." In 2014 The New York Times reported that Möller has assigned "a team of 10 software developers focused on mobile", out of a total of approximately 200 employees working at the Wikimedia Foundation. One principal concern cited by The New York Times for the "worry" is for Wikipedia to effectively address attrition issues with the number of editors which the online encyclopedia attracts to edit and maintain its content in a mobile access environment.[45] By 2023, the Wikimedia Foundation's staff had grown to over 700 employees.[2]

Access to Wikipedia from mobile phones was possible as early as 2004, through the Wireless Application Protocol (WAP), via the Wapedia service.[W 97] In June 2007, Wikipedia launched en.mobile.wikipedia.org, an official website for wireless devices. In 2009, a newer mobile service was officially released, located at en.m.wikipedia.org, which caters to more advanced mobile devices such as the iPhone, Android-based devices, or WebOS-based devices.[W 109] Several other methods of mobile access to Wikipedia have emerged since. Many devices and applications optimize or enhance the display of Wikipedia content for mobile devices, while some also incorporate additional features such as use of Wikipedia metadata like geoinformation.[272][273]

The Android app for Wikipedia was released in January 2012, to over 500,000 installs and generally positive reviews, scoring over four of a possible five in a poll of approximately 200,000 users downloading from Google.[W 110][W 111] The version for iOS was released on April 3, 2013, to similar reviews.[W 112] Wikipedia Zero was an initiative of the Wikimedia Foundation to expand the reach of the encyclopedia to the developing countries by partnering with mobile operators to allow free access.[W 113][274] It was discontinued in February 2018 due to lack of participation from mobile operators.[W 113]

Andrew Lih and Andrew Brown both maintain editing Wikipedia with smartphones is difficult and this discourages new potential contributors.[275][276] Lih states that the number of Wikipedia editors has been declining after several years,[275] and Tom Simonite of MIT Technology Review claims the bureaucratic structure and rules are a factor in this. Simonite alleges some Wikipedians use the labyrinthine rules and guidelines to dominate others and those editors have a vested interest in keeping the status quo.[35] Lih alleges there is a serious disagreement among existing contributors on how to resolve this. Lih fears for Wikipedia's long-term future while Brown fears problems with Wikipedia will remain and rival encyclopedias will not replace it.[275][276]

Chinese access

[edit]

Access to Wikipedia has been blocked in mainland China since May 2015.[7][277][278] This was done after Wikipedia started to use HTTPS encryption, which made selective censorship more difficult.[279]

Cultural influence

[edit]

Trusted source to combat fake news

[edit]

In 2017–18, after a barrage of false news reports, both Facebook and YouTube announced they would rely on Wikipedia to help their users evaluate reports and reject false news.[280][281] Noam Cohen, writing in The Washington Post states, "YouTube's reliance on Wikipedia to set the record straight builds on the thinking of another fact-challenged platform, the Facebook social network, which announced last year that Wikipedia would help its users root out 'fake news'."[281][282]

Readership

[edit]

In February 2014, The New York Times reported that Wikipedia was ranked fifth globally among all websites, stating "With 18 billion page views and nearly 500 million unique visitors a month, ... Wikipedia trails just Yahoo, Facebook, Microsoft and Google, the largest with 1.2 billion unique visitors."[45] However, its ranking dropped to 13th globally by June 2020 due mostly to a rise in popularity of Chinese websites for online shopping.[58] The website has since recovered its ranking as of April 2022.[58]

In addition to logistic growth in the number of its articles,[W 114] Wikipedia has steadily gained status as a general reference website since its inception in 2001.[283] The number of readers of Wikipedia worldwide reached 365 million at the end of 2009.[W 115] The Pew Internet and American Life project found that one third of US Internet users consulted Wikipedia.[284] In 2011, Business Insider gave Wikipedia a valuation of $4 billion if it ran advertisements.[285]

According to "Wikipedia Readership Survey 2011", the average age of Wikipedia readers is 36, with a rough parity between genders. Almost half of Wikipedia readers visit the site more than five times a month, and a similar number of readers specifically look for Wikipedia in search engine results. About 47 percent of Wikipedia readers do not realize that Wikipedia is a non-profit organization.[W 116] As of February 2023, Wikipedia attracts around 2 billion unique devices monthly, with the English Wikipedia receiving 10 billion pageviews each month.[W 1]

COVID-19 pandemic

[edit]

During the COVID-19 pandemic, Wikipedia's coverage of the pandemic and fight against misinformation received international media attention, and brought an increase in Wikipedia readership overall.[286][287][288][289] Noam Cohen wrote in Wired that Wikipedia's effort to combat misinformation related to the pandemic was different from other major websites, opining, "Unless Twitter, Facebook and the others can learn to address misinformation more effectively, Wikipedia will remain the last best place on the Internet."[287] In October 2020, the World Health Organization announced they were freely licensing its infographics and other materials on Wikimedia projects.[290] There were nearly 7,000 COVID-19 related Wikipedia articles across 188 different Wikipedias, as of November 2021.[291][292]

Cultural significance

[edit]
Wikipedia Monument in Słubice, Poland, by Mihran Hakobyan (2014)

Wikipedia's content has also been used in academic studies, books, conferences, and court cases.[W 117][293][294] The Parliament of Canada's website refers to Wikipedia's article on same-sex marriage in the "related links" section of its "further reading" list for the Civil Marriage Act.[295] The encyclopedia's assertions are increasingly used as a source by organizations such as the US federal courts and the World Intellectual Property Organization[296]—though mainly for supporting information rather than information decisive to a case.[297] Content appearing on Wikipedia has also been cited as a source and referenced in some US intelligence agency reports.[298] In December 2008, the scientific journal RNA Biology launched a new section for descriptions of families of RNA molecules and requires authors who contribute to the section to also submit a draft article on the RNA family for publication in Wikipedia.[299] Wikipedia has also been used as a source in journalism,[300][301] often without attribution, and several reporters have been dismissed for plagiarizing from Wikipedia.[302][303][304][305]

In 2006, Time magazine recognized Wikipedia's participation (along with YouTube, Reddit, MySpace, and Facebook) in the rapid growth of online collaboration and interaction by millions of people worldwide.[306] On September 16, 2007, The Washington Post reported that Wikipedia had become a focal point in the 2008 US election campaign, saying: "Type a candidate's name into Google, and among the first results is a Wikipedia page, making those entries arguably as important as any ad in defining a candidate. Already, the presidential entries are being edited, dissected and debated countless times each day."[307] An October 2007 Reuters article, titled "Wikipedia page the latest status symbol", reported the recent phenomenon of how having a Wikipedia article vindicates one's notability.[308]

One of the first times Wikipedia was involved in a governmental affair was on September 28, 2007, when Italian politician Franco Grillini raised a parliamentary question with the minister of cultural resources and activities about the necessity of freedom of panorama. He said that the lack of such freedom forced Wikipedia, "the seventh most consulted website", to forbid all images of modern Italian buildings and art, and claimed this was hugely damaging to tourist revenues.[309]

Wikipedia, an introduction – Erasmus Prize 2015
Jimmy Wales accepts the 2008 Quadriga A Mission of Enlightenment award on behalf of Wikipedia.

A working group led by Peter Stone (formed as a part of the Stanford-based project One Hundred Year Study on Artificial Intelligence) in its report called Wikipedia "the best-known example of crowdsourcing ... that far exceeds traditionally-compiled information sources, such as encyclopedias and dictionaries, in scale and depth".[310][311]

In a 2017 opinion piece for Wired, Hossein Derakhshan describes Wikipedia as "one of the last remaining pillars of the open and decentralized web" and contrasted its existence as a text-based source of knowledge with social media and social networking services, the latter having "since colonized the web for television's values". For Derakhshan, Wikipedia's goal as an encyclopedia represents the Age of Enlightenment tradition of rationality triumphing over emotions, a trend which he considers "endangered" due to the "gradual shift from a typographic culture to a photographic one, which in turn mean[s] a shift from rationality to emotions, exposition to entertainment". Rather than "sapere aude" (lit.'dare to know'), social networks have led to a culture of "dare not to care to know". This is while Wikipedia faces "a more concerning problem" than funding, namely "a flattening growth rate in the number of contributors to the website". Consequently, the challenge for Wikipedia and those who use it is to "save Wikipedia and its promise of a free and open collection of all human knowledge amid the conquest of new and old television—how to collect and preserve knowledge when nobody cares to know."[312]

Awards

[edit]
Wikipedia team visiting the Parliament of Asturias
Wikipedians meeting after the 2015 Asturias awards ceremony

Wikipedia has won many awards, receiving its first two major awards in May 2004.[W 118] The first was a Golden Nica for Digital Communities of the annual Prix Ars Electronica contest; this came with a €10,000 (£6,588; $12,700) grant and an invitation to present at the PAE Cyberarts Festival in Austria later that year. The second was a Judges' Webby Award for the "community" category.[313]

In 2007, readers of brandchannel.com voted Wikipedia as the fourth-highest brand ranking, receiving 15 percent of the votes in response to the question "Which brand had the most impact on our lives in 2006?"[314]

In September 2008, Wikipedia received Quadriga A Mission of Enlightenment award of Werkstatt Deutschland along with Boris Tadić, Eckart Höfling, and Peter Gabriel. The award was presented to Wales by David Weinberger.[315]

In 2015, Wikipedia was awarded both the annual Erasmus Prize, which recognizes exceptional contributions to culture, society or social sciences,[316] and the Spanish Princess of Asturias Award on International Cooperation.[317] Speaking at the Asturian Parliament in Oviedo, the city that hosts the awards ceremony, Jimmy Wales praised the work of the Asturian Wikipedia users.[318]

Satire

[edit]

Comedian Stephen Colbert has parodied or referenced Wikipedia on numerous episodes of his show The Colbert Report and coined the related term wikiality, meaning "together we can create a reality that we all agree on—the reality we just agreed on".[184] Another example can be found in "Wikipedia Celebrates 750 Years of American Independence", a July 2006 front-page article in The Onion,[319] as well as the 2010 The Onion article "'L.A. Law' Wikipedia Page Viewed 874 Times Today".[320]

In an April 2007 episode of the American television comedy The Office, office manager (Michael Scott) is shown relying on a hypothetical Wikipedia article for information on negotiation tactics to assist him in negotiating lesser pay for an employee.[321] Viewers of the show tried to add the episode's mention of the page as a section of the actual Wikipedia article on negotiation, but this effort was prevented by other users on the article's talk page.[322]

"My Number One Doctor", a 2007 episode of the television show Scrubs, played on the perception that Wikipedia is an unreliable reference tool with a scene in which Perry Cox reacts to a patient who says that a Wikipedia article indicates that the raw food diet reverses the effects of bone cancer by retorting that the same editor who wrote that article also wrote the Battlestar Galactica episode guide.[323]

In 2008, the comedy website CollegeHumor produced a video sketch named "Professor Wikipedia", in which the fictitious Professor Wikipedia instructs a class with a medley of unverifiable and occasionally absurd statements.[324] The Dilbert comic strip from May 8, 2009, features a character supporting an improbable claim by saying "Give me ten minutes and then check Wikipedia."[325] In July 2009, BBC Radio 4 broadcast a comedy series called Bigipedia, which was set on a website which was a parody of Wikipedia.[326] Some of the sketches were directly inspired by Wikipedia and its articles.[327]

On August 23, 2013, the New Yorker website published a cartoon with this caption: "Dammit, Manning, have you considered the pronoun war that this is going to start on your Wikipedia page?"[328] The cartoon referred to Chelsea Elizabeth Manning (born Bradley Edward Manning), an American activist, politician, and former United States Army soldier who had recently come out as a trans woman.[329]

In June 2024, nature.com published a fictional Wikipedia Talk page under the title "Plastic-eating fungus caused doomsday" by Emma Burnett. The Talk page concerned a fictional article describing the unintended consequences of the release of a plastic-eating fungus to clean up an oil spill. The article contained Talk page topics found on Wikipedia, like discussions of changes in the articles priority level.[330]

Publishing

[edit]
A group of Wikimedians of the Wikimedia DC chapter at the 2013 DC Wikimedia annual meeting standing in front of the Encyclopædia Britannica (back left) at the US National Archives

The most obvious economic effect of Wikipedia has been the death of commercial encyclopedias, especially printed versions like Encyclopædia Britannica, which were unable to compete with a product that is essentially free.[331][332][333] Nicholas Carr's 2005 essay "The amorality of Web 2.0" criticizes websites with user-generated content (like Wikipedia) for possibly leading to professional (and, in his view, superior) content producers' going out of business, because "free trumps quality all the time". Carr wrote, "Implicit in the ecstatic visions of Web 2.0 is the hegemony of the amateur. I for one can't imagine anything more frightening."[334] Others dispute the notion that Wikipedia, or similar efforts, will entirely displace traditional publications. Chris Anderson, the former editor-in-chief of Wired, wrote in Nature that the "wisdom of crowds" approach of Wikipedia will not displace top scientific journals with rigorous peer review processes.[335]

Wikipedia's influence on the biography publishing business has been a concern for some. Book publishing data tracker Nielsen BookScan stated in 2013 that biography sales were dropping "far more sharply".[336] Kathryn Hughes, professor of life writing at the University of East Anglia and author of two biographies wrote, "The worry is that, if you can get all that information from Wikipedia, what's left for biography?"[336]

Research use

[edit]

Wikipedia has been widely used as a corpus for linguistic research in computational linguistics, information retrieval and natural language processing.[337][338] In particular, it commonly serves as a target knowledge base for the entity linking problem, which is then called "wikification",[339] and to the related problem of word-sense disambiguation.[340] Methods similar to wikification can in turn be used to find "missing" links in Wikipedia.[341]

In 2015, French researchers José Lages of the University of Franche-Comté in Besançon and Dima Shepelyansky of Paul Sabatier University in Toulouse published a global university ranking based on Wikipedia scholarly citations.[342][343][344] They used PageRank, CheiRank and similar algorithms "followed by the number of appearances in the 24 different language editions of Wikipedia (descending order) and the century in which they were founded (ascending order)".[344][345] The study was updated in 2019.[346]

In December 2015, John Julius Norwich stated, in a letter published in The Times newspaper, that as a historian he resorted to Wikipedia "at least a dozen times a day", and had "never caught it out". He described it as "a work of reference as useful as any in existence", with so wide a range that it is almost impossible to find a person, place, or thing that it has left uncovered and that he could never have written his last two books without it.[347]

A 2017 MIT study suggests that words used in Wikipedia articles end up in scientific publications.[348] Studies related to Wikipedia have been using machine learning and artificial intelligence[311] to support various operations. One of the most important areas is the automatic detection of vandalism[349][350] and data quality assessment in Wikipedia.[351][352]

In February 2022, civil servants from the UK's Housing, Communities and Local Government Committee were found to have used Wikipedia for research after journalists at The Independent noted that parts of the document had been lifted directly from Wikipedia articles on Constantinople and the list of largest cities throughout history.[353]

[edit]

Several interactive multimedia encyclopedias incorporating entries written by the public existed long before Wikipedia was founded. The first of these was the 1986 BBC Domesday Project, which included text (entered on BBC Micro computers) and photographs from more than a million contributors in the UK, and covered the geography, art, and culture of the UK. This was the first interactive multimedia encyclopedia (and was also the first major multimedia document connected through internal links), with the majority of articles being accessible through an interactive map of the UK. The user interface and part of the content of the Domesday Project were emulated on a website until 2008.[354]

Several free-content, collaborative encyclopedias were created around the same period as Wikipedia (e.g. Everything2),[355] with many later being merged into the project (e.g. GNE).[W 119] One of the most successful early online encyclopedias incorporating entries by the public was h2g2, which was created by Douglas Adams in 1999. The h2g2 encyclopedia is relatively lighthearted, focusing on articles which are both witty and informative.[356]

Subsequent collaborative knowledge websites have drawn inspiration from Wikipedia. Others use more traditional peer review, such as Encyclopedia of Life and the online wiki encyclopedias Scholarpedia and Citizendium.[357][358] The latter was started by Sanger in an attempt to create a reliable alternative to Wikipedia.[359][360]

See also

[edit]

Notes

[edit]

References

[edit]

Further reading

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Wikipedia is a collaborative, multilingual online encyclopedia of freely editable articles written and maintained by volunteers using wiki software under free content licenses. Launched on January 15, 2001, by Jimmy Wales and Larry Sanger as a wiki-based complement to the expert-reviewed Nupedia, it expanded rapidly through open editing and has been hosted by the non-profit Wikimedia Foundation since 2003. As of January 2026, it hosts over 66 million articles across 342 active language editions, including more than 7.1 million in English, establishing it as one of the largest reference works. Praised for its scale, accessibility, and role in broadening knowledge access, Wikipedia also faces ongoing critiques of factual accuracy, vulnerability to vandalism, and ideological biases, including a left-leaning tendency in coverage as noted in analyses and by co-founder Sanger.

History

Precursors: Nupedia and early concepts

Nupedia was founded on March 9, 2000, by Jimmy Wales, who provided initial funding through his dot-com company Bomis; Larry Sanger was hired as editor-in-chief to oversee content. Bomis, a web portal focused on men's interest sites including adult content, generated revenue from advertising and underwrote Nupedia with intentions to generate revenue through online advertising on the encyclopedia. Nupedia emulated traditional encyclopedias by requiring expert-authored articles to undergo a seven-step editorial process: assignment, finding a lead reviewer, lead review, open review, lead copyediting, open copyediting, and final approval and markup. This approach ensured verifiable quality but was time-intensive, yielding only two full-length articles by November 2000 and reaching 21 by the end of its first year in March 2001. To address slow growth, Sanger proposed on January 10, 2001, to the Nupedia mailing list a complementary wiki for non-experts to draft articles as raw material for Nupedia's formal review. The idea arose from Sanger's discussion with software engineer Ben Kovitz, who introduced wiki software—developed by Ward Cunningham in 1994 as a collaborative hypertext system for rapid, permissionless editing. Unlike Nupedia's expert-driven model, the wiki prioritized speed and inclusivity, enabling broader participation in knowledge creation.

Launch in 2001 and initial rapid expansion

Wikipedia launched on January 15, 2001, by Jimmy Wales and Larry Sanger as a complement to Nupedia, an expert-reviewed encyclopedia Wales had founded in 2000 with Sanger as editor-in-chief. Nupedia's rigorous process had yielded few articles, prompting Sanger—introduced to wiki software by Ben Kovitz—to propose open contributions for faster drafting that could feed into it. The English edition started with seed articles, including Nupedia ports, and gained traction via its permissive model favoring rapid iteration over perfection. This differed from traditional encyclopedias and Nupedia's delays, enabling worldwide volunteer participation. By March 2001, it reached 1,000 articles, showing exponential growth from community enthusiasm and low entry barriers. Growth accelerated through 2001 as word spread online, spawning non-English editions in spring. By December, English Wikipedia had nearly 19,000 articles across 18 language versions, with thousands of daily edits. The model's viral spread and lack of paywalls or credentials drove this, though factual errors emerged, necessitating community corrections. This scaling affirmed open collaboration's potential for encyclopedias, surpassing Nupedia and paving the way for expansion.

Key milestones and scaling challenges

The English Wikipedia reached its one millionth article on March 1, 2006, with the creation of the entry for Jordanhill railway station, expanding from fewer than 20,000 articles at the end of 2001. This milestone reflected exponential early growth driven by volunteer contributions and wiki software enabling rapid collaborative editing. The edition surpassed two million articles on September 9, 2007, via an entry on the Spanish program El Hormiguero. Subsequent benchmarks included three million articles in August 2009 and six million in January 2020, highlighting sustained accumulation amid decelerating new content rates. This rapid scaling created technical challenges, especially in the mid-2000s as daily page views rose from thousands to millions and strained primary application/database servers centralized in a Florida data center operational from 2004, while edge caching/proxy clusters were deployed internationally (e.g., Amsterdam, Paris, Seoul) starting in 2005. Performance issues from UseModWiki and early MediaWiki limitations slowed article creation in 2002, requiring database optimizations and upgrades. A February 13, 2006, outage halted access for hours, beginning with a failure of the primary DNS/NFS server that led to cascading failures and subsequent database issues during restoration efforts. The Wikimedia Foundation addressed these through 2005–2006 server relocations and multi-datacenter expansions, supporting tens of thousands of page views per second by the mid-2010s while maintaining low costs via open-source efficiencies.

Development of sister projects

Wiktionary, the first sister project, launched on December 12, 2002, as a collaborative multilingual dictionary. It addressed Wikipedia's policy against dictionary functions by focusing on definitions, etymologies, and translations. This expansion showed early recognition that the wiki model could support specialized reference works beyond encyclopedias, using open editing and the GNU Free Documentation License. The Wikimedia Foundation's creation on June 20, 2003, provided nonprofit structure for further diversification. That year, Wikibooks debuted on July 10 for free textbooks via modular chapters, while Wikiquote, a sourced quotations repository, and Wikisource, a digital library for public-domain texts like literature and historical records, also emerged. Wikiquote started mid-2003 on a temporary host before its domain, and Wikisource on November 24, emphasizing primary sources over interpretation. In 2004, projects filled multimedia and news gaps. Wikimedia Commons launched September 7 as a repository for freely usable media files, easing Wikipedia's copyright issues. Wikinews followed on November 8 for crowdsourced reporting under neutral guidelines, though it struggled with contributor retention compared to static projects. Wikiversity activated August 15, 2006, after Wikibooks incubation, to host learning resources, courses, and research with experimental modules. Wikidata launched October 29, 2012, adding structured data for infoboxes, multilingual labels, and queries to cut redundancy and enable machine use. These projects expanded the Wikimedia ecosystem via shared MediaWiki software and governance, though adoption varied: reference works like Wiktionary succeeded, while others like Wikinews saw lower engagement.

Content Creation and Editing

Collaborative editing model and tools

Wikipedia's collaborative editing model allows any internet user to edit articles, beginning with its launch on January 15, 2001. This open approach drives rapid content growth via volunteer contributions without centralized control. Edits occur directly on pages, with disputes resolved on talk pages through decentralized consensus that favors iterative improvements over fixed authorship. To mitigate vandalism and bias, protections such as requiring logged-in accounts for biographies of living persons were implemented in August 2009. MediaWiki supplies core tools like the source editor, which uses wikitext for formatting including [[links]] and {{templates}}. Revision history enables viewing and reverting changes, while watchlists and diff tools aid monitoring and comparisons, promoting transparency in group editing. The VisualEditor, added in MediaWiki 1.20 (2012), provides a WYSIWYG interface for non-technical users with real-time previews and drag-and-drop functionality. Enhancements for newcomers, such as a guidance dashboard and mentor pairing, launched in December 2022 to boost retention amid declining editor counts. Bots and semi-automated scripts handle repetitive tasks like vandalism reversal and citation formatting, but require community approval to ensure human oversight. These elements support millions of monthly edits yet expose the model to risks like coordinated manipulation, including paid editing scandals.

Core policies: Neutrality, verifiability, and no original research

Wikipedia's core content policies—neutral point of view (NPOV), verifiability, and no original research (NOR)—require reliance on established knowledge from reliable sources, prohibiting personal invention or advocacy. NPOV mandates representing all major viewpoints proportionally to their prominence in reliable sources, without endorsing any position. Editors achieve this by attributing claims to proponents and selecting appropriate sources, though source choices can influence outcomes. Verifiability requires that content be attributable to secondary, reliable, published sources, with inclusion based on verifiability rather than truth. Co-founder Jimmy Wales stressed in 2006 that unsourced material should be removed to favor verifiable claims. Reliable sources generally encompass peer-reviewed academics, mainstream journalism, and established books, but determinations of reliability pose challenges, as these institutions often display left-leaning biases on politically sensitive topics. NOR forbids unpublished theories, data, or novel syntheses, ensuring contributions stem from prior publications rather than independent analysis. It complements verifiability and NPOV, as personal interpretations risk violating all three; for example, synthesizing facts from multiple sources to draw a new conclusion is prohibited. In practice, these policies encounter issues. Studies of article language indicate deviations from neutrality, with right-leaning figures and topics more negatively portrayed than left-leaning ones, reflecting influences from editor demographics and source selection. Co-founder Larry Sanger has criticized declining neutrality since around 2009, attributing it to biases in editors and favored sources. Enforcement depends on community consensus, which may sustain imbalances when sources carry institutional prejudices.

Vandalism detection, restrictions, and protections

Wikipedia detects vandalism—deliberate disruptive edits—through automated tools, machine learning, and human patrollers. ClueBot NG, an anti-vandalism bot, scans English Wikipedia edits in real time, reverting obvious issues like nonsensical or offensive insertions using algorithms. Launched around 2010, it offers high precision but reverts fewer than 50% of subtle cases in some studies. The Objective Revision Evaluation Service (ORES), introduced by the Wikimedia Foundation in 2015, applies machine learning to score revisions for vandalism risk, supporting editor triage and tool integration across projects. Effectiveness varies by language and edit type due to training data constraints. Human patrollers monitor recent changes feeds, aided by scripts like Twinkle for swift reverts. These systems collectively revert most detected vandalism within minutes, with bots handling obvious cases and humans addressing nuanced ones, thus linking rapid detection to content stability. Editing restrictions feature administrator blocks, which halt changes by IP addresses or accounts, targeting repeat vandals or sockpuppets. Blocks rely on disruptive edit histories, often indefinite for vandalism-only accounts to deter evasion without warnings. Site-wide blocks apply across projects for coordinated abuse. Page protections shield high-risk articles, such as biographies of living persons. Semi-protection restricts unregistered or new users (accounts under four days with fewer than ten edits) to discussion or review queues, curbing anonymous vandalism but potentially increasing editor dropout. Extended confirmed protection limits edits to accounts over 30 days old with 500 edits when semi-protection fails. Full protection confines changes to administrators amid disputes, while pending changes queues unpatrolled edits for review. These tools effectively counter casual vandalism yet raise barriers that may slow legitimate contributions.

Dispute resolution mechanisms

Disputes typically start on article talk pages, where editors seek consensus through policy-focused discussion and avoid personal attacks. Persistent reversions trigger the three-revert rule (3RR), which bars any editor from reverting a page more than three times in 24 hours to prevent edit warring, though it prioritizes behavioral patterns over counts; violations may lead to blocks. Escalation options include Requests for Comments (RfC), which post neutral issue summaries on noticeboards to gather input from uninvolved editors over about 30 days, approximating consensus despite limitations like biased framing or low resolution rates (around 40% achieve closure). Specialized noticeboards address issues such as sourcing or admin input, providing third-party views without voting. Formal mediation uses volunteers to facilitate content compromises, distinct from arbitration for conduct violations. The Arbitration Committee (ArbCom), elected annually since 2003, acts as the final body for severe disputes, reviewing evidence publicly and issuing remedies like bans or restrictions; it has handled hundreds of cases, including 2023 Holocaust misinformation disputes. However, studies indicate outcomes often reflect disputants' social capital, such as arbitrator ties, over pure merit. Critics contend that these mechanisms favor established networks and show biases, particularly in contentious topics where aligned groups influence outcomes, as in efforts to shape narratives despite neutrality rules. Analyses of over 250 disputes reveal that while deliberation occurs, power imbalances often perpetuate conflicts, with arbitration enforcing norms rather than neutrally settling substantive issues.

Governance and Administration

Role of the Wikimedia Foundation

The Wikimedia Foundation, established on June 20, 2003, by Jimmy Wales in St. Petersburg, Florida, as a 501(c)(3) nonprofit organization, serves as the primary institutional supporter of Wikipedia and its sister projects. It provides technical infrastructure, including servers and data centers, to host platforms handling billions of monthly page views without advertising or paywalls, funded by voluntary reader donations averaging around $11 per contribution. The foundation maintains the [[MediaWiki]] software, develops editing tools, and ensures operational scalability, but refrains from editorial control over content, which volunteer editors govern via community policies. It also manages fundraising, grants to advance knowledge equity in underrepresented languages and regions, and legal advocacy to protect free knowledge access, such as against surveillance or copyright restrictions. For fiscal year 2024-2025, revenues totaled approximately $208.6 million, primarily from donations, with expenses directed toward infrastructure, staff salaries for about 363 employees, and program support involving over 277,000 volunteers. A Board of Trustees, including community representatives and external experts, oversees governance and sets strategic direction, while the CEO manages daily operations; former CEO Katherine Maher (2019–2021) drew criticism for promoting content moderation against "disinformation," viewed by detractors as clashing with Wikipedia's open-editing principles. Critics such as co-founder Larry Sanger argue that shifts under leaders like Maher, including the Universal Code of Conduct since 2022, prioritize ideological interventions over neutral community consensus, as evidenced by internal resistance and funding links to organizations like Google and Facebook that could incentivize mainstream alignment. While the foundation's independence from content creation sustains the volunteer-driven model, its resource allocation and policy efforts have prompted debate over amplifying biases from San Francisco-based leadership and donors, contrasting the decentralized editing ideal.

Administrators and moderation powers

Administrators, or "admins," are experienced volunteer editors with elevated permissions for maintenance, enforcement, and moderation tasks. These include blocking or unblocking users to prevent vandalism, deleting or restoring pages and revisions, protecting or unprotecting articles to control editing access, and managing user rights such as temporary protections. The tools primarily support janitorial functions and behavioral policy enforcement, while content disputes follow community consensus. Selection occurs via Requests for Adminship (RfA), a decentralized community process where candidates self-nominate after demonstrating sustained contributions, policy expertise, and dispute resolution skills. Typical requirements involve thousands of edits over years, with approval needing broad consensus during scrutiny of the candidate's record for competence and impartiality. Pass rates have remained below 50% recently, due to high standards and occasional politicization. Guidelines restrict admin powers to avoid involvement in content battles or personal conflicts, banning "wheel-warring" without consensus. Oversight allows desysopping through community votes or Arbitration Committee decisions for serious misuse, though such actions are infrequent. Declining active admin numbers, driven by burnout, policy demands, and rigorous RfAs, have worsened moderation backlogs amid growing content. Criticisms focus on potential abuses, such as disproportionate blocks of new or differing editors, enforcement biases, and limited accountability in an anonymous environment. Co-founder Larry Sanger argues admins often suppress conservative viewpoints, promoting imbalance over neutrality. Reports highlight patterns of bullying, quick blocks without process, and resistance to change, linked to ideological alignment among long-term admins. Defenders maintain most admins operate responsibly despite workload pressures, yet infrequent bias-related desysoppings suggest gaps in effective deterrence, contributing to editor attrition on disputed topics.

Arbitration Committee and oversight

The Arbitration Committee (ArbCom) serves as English Wikipedia's highest dispute resolution body, issuing binding decisions in editorial conflicts unresolved by lower mechanisms. Established by co-founder Jimmy Wales on December 4, 2003, as disputes exceeded his personal handling, ArbCom shifted from founder-led to volunteer-led adjudication. It consists of 8 to 15 elected members on staggered one- to three-year terms, selected annually by community vote among candidates meeting thresholds like 500 mainspace edits and gaining editor support. After reviewing evidence publicly, it applies remedies such as user bans, topic restrictions, desysopping, and project-wide sanctions, with private deliberations for sensitive cases. ArbCom also oversees advanced permissions like checkuser (for IP tracing abuse detection) and oversight (revision suppression), ensuring compliance with privacy policies amid legal risks. Oversighters—experienced, vetted editors, often administrators, granted access via consensus or ArbCom—can suppress revisions, summaries, usernames, or logs containing personal data, legal violations, or doxxing risks, while maintaining internal audits for other functionaries. As of 2023, English Wikipedia had about 20-30 active oversighters, with usage logged privately and revocable for misuse; ArbCom reviews suppression disputes to balance transparency and privacy. Critics argue ArbCom decisions favor arbitrators' social ties over evidence, as studies show connected users receive lighter sanctions. The 2019 Fram ban—imposed without notice, partially reversed but upholding desysopping—exposed opaque processes and overreach concerns. In contentious areas like the Israel-Palestine conflict, ArbCom issued topic bans to eight editors in a 2025 case, drawing criticism for avoiding systemic biases in favor of harmony. These cases highlight tensions in ArbCom's volunteer model, prompting reform ideas for decentralized or professional oversight to reduce insider influence.

Internal research and development efforts

The Wikimedia Foundation's Research team conducts empirical analyses of Wikimedia projects, including Wikipedia, using data science, statistical modeling, and machine learning. These analyses draw from edit histories and user behaviors to improve content quality, editor engagement, and infrastructure. The team publishes findings in peer-reviewed papers, reports, and open datasets. For example, studies reveal that only 10-20% of new editors persist beyond initial sessions, leading to interventions like enhanced onboarding tools. A key initiative is the Objective Revision Evaluation Service (ORES), launched in 2015. This machine learning API scores edits in real time across projects. Trained on historical data, its classifiers predict edit quality, identifying damaging changes like vandalism with about 90% accuracy on English Wikipedia and assessing good faith similarly. ORES integrates with patrol tools for automated flagging, easing burdens on volunteer moderators. The Scoring Platform extends ORES with models for article quality, reversion likelihood, and linguistic analysis, supporting tools like ClueBot NG. AI explorations proceed cautiously; a 2025 strategy prioritizes augmenting human editors through draft review and neural-model-based translation aids, not replacement. A 2025 human rights assessment highlighted risks, including algorithmic bias in non-English scoring from uneven training data, and called for regular audits. Further efforts include design research, such as A/B testing editor interfaces, and longitudinal studies on knowledge equity that measure coverage gaps, like those in Global South histories. The Wikimedia Research & Technology Fund supports community prototypes via grants, funding over 50 projects since inception, including citation-suggestion bots. Outputs emphasize open-source models and datasets under permissive licenses to enable external review and refinement.

Community Dynamics

Wikimedia Foundation surveys indicate that Wikipedia editors are predominantly male, with the 2024 Community Insights report finding 13% identifying as women and 5% as gender diverse, aligning with earlier data of 87% male overall. Editors are largely urban, with 61% in big cities or metro areas, concentrated in North America and Europe—particularly the United States (around 20%), Germany (12%), and Russia (7%). Age skews toward younger adults, with 21% aged 18-24 as the largest group, though experienced editors average older than newcomers (37% of whom are 18-24). Education levels are high, with 81% holding post-secondary degrees and 42% postgraduate qualifications. Ethnic minorities comprise 13% of editors and 21% belong to discriminated groups, but U.S.-based editors severely underrepresent Black or African American contributors at under 1%. Participation trends show stabilization after years of decline, with English Wikipedia sustaining around 39,000 active editors (five or more edits monthly) as of December 2024, down 0.15% year-over-year—a pattern of continued plateau in active numbers. Very active editor counts have leveled off since the post-2000s peak, affected by high newcomer reversion rates, policy complexity, and retention challenges. While newcomers introduce greater gender and age diversity, overall growth in active editors remains stagnant despite millions of registered accounts. Volunteer edits across all language editions totaled nearly 98 million in 2024 but declined to about 79 million in 2025, highlighting sustained yet concentrated activity among core contributors.

Efforts to address diversity gaps

The Wikimedia Foundation has launched initiatives to increase editor participation from underrepresented groups, including women, racial minorities, and non-Western contributors, despite ongoing gaps: about 14% of editors identify as women and 5% as non-binary per 2025 surveys. Project Rewrite encourages volunteer edits to expand content on women, addressing data that only 19% of English Wikipedia biographies feature them. Edit-a-thons form a key approach, such as Art+Feminism events since 2014, which have created or improved thousands of articles on women in art and drawn new editors, though retention remains low compared to veterans. Similar sessions target women in STEM or Global South perspectives, often in partnership with universities to teach Wikipedia guidelines. Wiki Education programs embed editing in academic courses, producing more diverse U.S. student contributors—11.8% Black or African-American, matching national figures—and aim to raise Hispanic participation above the platform's 3.6% baseline. The 2021 Wikimedia Equity Fund supported grants to groups like Howard University to boost Black community input, countering the dominance of white, male editors from the U.S. and Europe. A 2023 convening gathered over 70 women and non-binary people for research and strategies on gender gaps, complemented by "Open the Knowledge" campaigns for targeted recruitment. Foundation staff diversity improved, with 53% of 2019 U.S. hires as women and 30% from underrepresented racial groups, though this applies to employees rather than volunteers. Studies show modest content increases but persistent demographic imbalances, including unquantified racial and ethnicity gaps that favor Western topics; community critics question resource spending amid poor retention. Despite significant investments, editor diversity has risen gradually since 2010, limited by editing hurdles and cultural barriers for newcomers.

Harassment, doxxing, and community sustainability

Harassment among Wikipedia editors includes repeated offensive behavior such as personal attacks, threats, and targeted disruptions on talk pages, edit histories, or off-wiki channels. A 2017 analysis found that a small group of highly toxic editors generated 9% of abuse on discussion pages, often through direct insults like "you suck" or third-party references. Toxicity disproportionately impacts demographics like female editors, where online abuse contributes to gender imbalances in participation. Doxxing, or unauthorized disclosure of personal information, arises from internal policy breaches or external campaigns. In 2021, Hong Kong editor Philip Tzou received threats and exposure on platforms like Weibo after writing articles critical of Chinese policies. In 2025, leaked documents exposed plans by external actors, including hacking and facial recognition, to identify anonymous editors accused of antisemitism, involving groups like the Heritage Foundation. These events threaten volunteer safety, particularly on politically sensitive topics reliant on pseudonymity. Such issues erode community sustainability by driving editor attrition and deterring newcomers. Toxic interactions cause short-term losses of 0.5 to 2 active editing days per affected user, with one study across 80,307 users in a month showing substantial aggregate declines; experienced editors and administrators often depart due to harassment and blocks. Despite measures like the Wikimedia Foundation's 2020 universal code of conduct and abuse investigation tools, persistent toxicity challenges long-term volunteer retention.

Technical Infrastructure

MediaWiki software and features

MediaWiki is a free and open-source wiki software package written in PHP for collaborative content management on web platforms. It requires a compatible web server, PHP interpreter, and relational database like MySQL, MariaDB, PostgreSQL, or SQLite. Initially developed to support Wikipedia's growth, it emphasizes scalability via database replication, memcached caching, and content compression, handling millions of daily page views on Wikimedia sites. Originating from Magnus Manske's work in late 2001, MediaWiki evolved from Perl-based systems like UseModWiki into a PHP application, publicly released on January 25, 2002. Key milestones include stable versioning with MediaWiki 1.0 in July 2003 for revision tracking, and ongoing releases by the Wikimedia Foundation's engineering team with community input. Its architecture supports high-traffic sites through extensible hooks for third-party modifications without core changes. Core features use wikitext markup for formatting, including links, bold/italic text with asterisks and apostrophes, and lists via colons or hashes. Revision history tracks edits with timestamps, user attribution, and diffs for rollbacks and audits. Namespaces separate content types—such as articles, user pages, talk, and files—with 16 defaults configurable for searches. Templates provide reusable content via transclusion, with parameters, parser functions for logic, magic words like {{CURRENTTIME}}, and Lua scripting through Scribunto. Media embedding supports images and videos via extensions, plus MathJax or LaTeX for equations. User tools include watchlists, recent changes feeds, and categories for organization. Extensibility offers over 1,000 extensions, including VisualEditor (WYSIWYG, from 1.22 in 2013) and citation tools. These integrate via PHP and hooks but may add overhead or risks, so Wikimedia bundles vetted ones like Echo. Security includes edit tokens, rate limiting, and permissions; administrators can protect pages. Mobile support features the Minerva skin and APIs for responsive views since the 2010s.

Hardware, data centers, and operational scaling

The Wikimedia Foundation maintains Wikipedia's operations through servers in multiple data centers for high availability, redundancy, and global performance. These include eqiad in Ashburn, Virginia; codfw in Carrollton, Texas; esams in Amsterdam, Netherlands; ulsfo in San Francisco, California; and caching sites eqsin in Singapore (since 2017), drmrs in Marseille, France (since 2022), and magru in São Paulo, Brazil (since 2024), supporting geographic diversity and disaster recovery. The Foundation conducts semi-annual automated failover tests to ensure resilience against disruptions. Hardware infrastructure includes racks with MediaWiki instances, caching layers, and storage systems. Recent expansions at Ashburn and Carrollton have increased capacity and refreshed hardware to meet rising demands. The multi-tier architecture separates read/write operations, query processing, and caching for efficiency, handling over 18 billion monthly pageviews with roughly one-thousandth the servers of comparable sites. From a single server in 2003, operations scaled to multi-datacenter by the 2010s, incorporating Apache Traffic Server (ATS) to replace Varnish for edge caching and content delivery networks (CDNs) for traffic spikes. Recent pressures include a 50% bandwidth surge since early 2024 from AI training scrapers, accounting for up to 65% of peak traffic costs despite an 8% year-over-year decline in human traffic. Responses include promoting paid API access and signing January 2026 licensing deals with Microsoft, Meta, and Amazon for content use in AI training to fund infrastructure and reduce strain. Internet hosting costs reached $3.1 million in the 2023–2024 fiscal year.

Automated editing via bots and AI tools

Automated editing on Wikipedia relies on bots—software scripts for repetitive tasks inefficient for humans, such as reverting vandalism, adding categories, generating alerts, and maintaining interwiki links. Bots must follow the project's bot policy, requiring prior approval to avoid disrupting content or exceeding edit limits. The volunteer Bot Approvals Group (BAG) assesses proposals by examining code, tasks, and trial runs, often restricting operations to low-traffic times. As of 2017, over 2,100 bots supported Wikimedia projects' maintenance needs. Bots emerged early, with rambot (circa 2003) enabling mass formatting and linking. Later developments addressed rising edit volumes, including ClueBot NG's machine learning for vandalism detection (70% catch rate on affected pages). From 2007 to 2016, bots accounted for up to 18% of English Wikipedia edits despite few accounts, aiding scalability but sparking "bot wars" from conflicting automated reversions that ignore human nuance. Such issues have driven coordination improvements. AI integration enhances bots via machine learning tools like ORES for edit quality prediction, yet generative models for content creation or translation raise concerns over hallucinations and false citations that erode accuracy. In August 2025, Wikipedia tightened speedy deletion for AI-like traits, such as repetitive phrasing or dubious sources, following increased submissions. Limited AI use for drafting stubs aids low-resource languages but risks unverified errors. While potentially amplifying biases from training data, empirical evidence underscores human oversight's role in mitigation. AI boosts detection efficiency, yet unregulated generative automation challenges Wikipedia's emphasis on verifiable human contributions.

Multilingual Scope and Coverage

English Wikipedia's dominance and editor statistics

The English Wikipedia leads other editions in article volume, traffic, and editorial activity. As of January 2026, it hosts about 7,120,000 articles—more than the combined totals of many smaller editions—and sustains around 39,000 active editors (those making at least five edits monthly) as of December 2024. This preeminence arises from its first-mover advantage, English's prevalence in global scholarship and internet content, and a critical mass of editors proficient in English-language sources. In comparison, the German and French editions have roughly 2.7 million and 2.6 million articles, respectively, with active editor counts often below 10,000. Very active editors (over 100 edits monthly) number in the low tens of thousands, supporting high revision rates and expansion amid stagnant recruitment since the mid-2010s. Longitudinal data show English Wikipedia topping total edits across editions since 2001. Traffic reinforces this lead, with English capturing over 11 times the page views of the next edition and exceeding hundreds of millions monthly, aided by its role as a default reference and search engine integration. This pattern creates a feedback loop favoring English-centric topics, potentially limiting nuanced coverage in other languages and exacerbating global knowledge disparities.

Non-English language editions and disparities

As of January 2026, Wikipedia includes 358 language editions, supporting independent content creation across diverse languages. These non-English editions collectively contain millions of articles but show wide scale variations; the German edition exceeds 3 million entries, whereas many in indigenous or low-resource languages have under 1,000, constraining their reference utility. Editions like Cebuano approach 6 million articles through bot-generated stubs, which often lack depth and human oversight, prompting debate on their encyclopedic merit. Content quality and sourcing disparities persist, with non-English articles averaging fewer references than English ones, leading to less rigorous analysis and greater risk of unverified information. Topical coverage favors local contexts in about one-quarter of articles, yet smaller editions exhibit gaps in global events, science, and non-local biographies due to limited editors. Demographic imbalances, including underrepresentation of women and non-Western figures, have eased modestly in major languages over time. Editor engagement highlights these divides: non-English editions draw far fewer active contributors than English Wikipedia's 122,000, with French and German around 18,000 each and smaller ones depending on few individuals, which hinders maintenance and promotes viewpoint uniformity. Localized editor groups can introduce cultural or ideological tilts, especially on international disputes, differing from larger editions. Mitigation efforts, including translation tools that have produced over 2 million articles, face ongoing challenges from linguistic barriers and scarce sources in underrepresented languages.

Systemic coverage biases in topics and sourcing

Wikipedia exhibits systemic biases in topic selection and sourcing, stemming from its predominantly male, Western, and left-leaning editor base. A 2021 analysis of political science articles found underrepresentation of non-Western perspectives and skew toward privileged viewpoints, with student editing initiatives mitigating gaps but unable to prevent bias accumulation from uneven participation. A 2024 sentiment analysis of over 1,300 articles revealed a mild to moderate negative language bias against right-of-center public figures compared to left-leaning ones. Coverage disparities affect conservative viewpoints and non-mainstream ideologies, which receive less comprehensive treatment and citations than left-leaning topics. A 2008 study of 28,000 U.S. political articles identified systematic leftward deviation from neutral benchmarks, attributed to progressive-aligned sourcing patterns. Geographic biases emphasize Global North events, with a 2024 study showing heightened skew in underrepresented regions due to reliance on Western interpretations. Gender imbalances continue, as a 2022 analysis documented lower citation rates for female and non-Western authors, perpetuating shortfalls in areas like women's history and indigenous sciences. Sourcing practices intensify biases through the "reliable sources" policy, which favors mainstream media and academia often critiqued for left-wing orientations that undervalue dissenting scholarship. A 2015 Harvard Business School analysis of political articles found Wikipedia more biased than Encyclopædia Britannica, with 73% incorporating ideologically loaded phrasing from confirmation-biased outlets. In 2025, accusations arose of blacklisting right-leaning U.S. media as unreliable while heavily citing left-leaning sources, echoed by Senate Commerce Committee criticisms of the reliable sources list for ideological imbalances. A 2025 ADL report documented antisemitic and anti-Israel biases in coverage, undermining neutrality on related topics. This source reliance, lacking sufficient diverse or primary counterbalances, distorts causal interpretations in politically charged events by omitting non-mainstream empirical evidence.

Reliability and Accuracy

Empirical studies on factual correctness

A 2005 Nature study compared Wikipedia's science entries to Encyclopædia Britannica's by expert review of 42 articles in biology, chemistry, and physics. It found four serious errors in Wikipedia versus three in Britannica, and 162 factual inaccuracies versus 123, concluding comparable accuracy despite open editing. Britannica disputed the methodology, citing selective error counting and failure to differentiate major from minor issues, asserting its superior accuracy. Later studies revealed discipline-specific variations. A 2008 analysis of historical articles reported Wikipedia's 80% accuracy rate, below the 95-96% of Britannica and Colliers, due to incomplete sourcing and edit wars in contentious topics. Conversely, a 2011 study on U.S. congressional elections and politician biographies found it reliable for empirical analysis, with low error rates in verifiable facts but gaps in obscure areas. A 2014 study across English, Spanish, and Arabic Wikipedias in history and science noted high factual accuracy overall but inconsistencies, including more omissions in non-English editions and error propagation from secondary sources. Wikipedia corrects errors quickly—often within hours—via active editing, surpassing static encyclopedias in dynamism, though its scale increases absolute errors, and accuracy drops in politically sensitive or under-edited topics amid disputes. For example, an unsourced 1868 founding claim for Caerleon A.F.C.—added in 2007 and incorrect (actual: 1902)—persisted until 2024 removal, spreading to sites like Transfermarkt and Worldfootball.net. Academic studies, often from open-knowledge advocates, may understate biases appearing as factual slants rather than errors.

Criticisms of sourcing and citation practices

Wikipedia's sourcing practices emphasize verifiability through citations to reliable published sources, determined by community consensus. However, the policy states that "the threshold for inclusion in Wikipedia is verifiability, not truth," allowing retention of material traceable to a cited source even if factually erroneous. Critics argue this enables persistent inaccuracies, as editors defend sourced claims unless contradicted by equally authoritative references. For instance, computer scientist Jaron Lanier found his entry erroneously listing him as a film director, a claim upheld due to sourcing despite his refutation, illustrating barriers to expert corrections without third-party validation. Designation of reliable sources remains subjective, with editors often deeming mainstream news outlets credible while viewing conservative-leaning ones like Fox News or Breitbart as unreliable. This process, via noticeboards and reviews, is accused of introducing ideological bias by favoring secondary sources from left-leaning institutions. A 2024 analysis confirmed moderate liberal skew in source usage, with heavier reliance on outlets like The New York Times. Co-founder Larry Sanger has cited imbalances in political biographies, such as disproportionate scandal citations for Donald Trump versus Barack Obama. Empirical studies highlight citation shortcomings, including frequent "citation needed" tags. A 2016 Dartmouth analysis of prominent articles found over 90% of statements lacking inline citations, undermining traceability. In health sciences, peer-reviewed papers critiquing Wikipedia note insufficiently robust sourcing, risking propagation of outdated references. Restrictions on original research and primary sources further compel dependence on secondary interpretations, critics say, marginalizing dissenting perspectives in areas like abortion or alternative medicine rather than citing them neutrally.

Applications in education, medicine, and policy

Wikipedia serves education as a supplementary tool for student research and collaborative editing, despite faculty concerns over reliability. A 2014 survey found 68% of university faculty allowed its use for initial ideas, but only 18% as a primary source, due to inaccuracies and poor citations. Projects editing scientific articles build critical thinking and information literacy, with participants gaining better source evaluation skills. Yet, 75% of undergraduates accessed it for assignments in 2016, despite bans, favoring convenience over verified sources and highlighting reliance risks. Ongoing initiatives, such as Wiki Education's 2024–2025 programs, sustain these educational efforts. In medicine, Wikipedia is a frequent health information resource; by 2014, 72% of U.S. internet users seeking diagnoses consulted it. Accuracy varies: a 2014 review of top-searched conditions found errors in nine of ten, including flawed treatments and outdated details. Systematic 2020 assessments showed about 80% alignment with professional sources for common topics, but gaps in drug interactions and rare diseases, with ninth-grade readability. Recent programs engage experts and students to refine medical content. Guidelines emphasize tertiary sources like textbooks, though volunteer edits create inconsistencies; educators employ it to teach source critique, not as definitive reference. In public policy, reliance is minimal, with officials avoiding it for its editability and verification issues; standards prohibit citations in formal work. It influences indirectly via public discussions or briefings, but lacks evidence of direct policy adoption—a 2024 study noted organizations contributing to entries rather than using them for decisions. The EU's 2022 classification of Wikipedia under the Digital Services Act adds moderation transparency, impacting misinformation topics more than policy content. Policy favors primary sources, limiting Wikipedia to exploratory roles amid coverage biases.

Biases and Political Influences

Evidence of left-leaning systemic bias in articles

A 2024 computational analysis of over 1,000 Wikipedia biographies of political figures from 17 countries found that articles on right-of-center figures contained more negative sentiment than those on left-of-center counterparts, with disparities statistically significant across metrics. Data scientist David Rozado used large language models to score affective connotations, yielding average sentiment scores 0.05 to 0.10 points lower for right-leaning figures on a standardized scale, indicating a mild to moderate leftward tilt. Similar patterns appeared in news media depictions, with left-leaning outlets like The Guardian receiving higher positive sentiment (+0.08 average) than right-leaning ones like Fox News (-0.06 average). Comparative studies against reference works like Encyclopædia Britannica support these findings. Economists Shane Greenstein and Feng Zhu analyzed U.S. political articles from 2008 to 2012 using word choice frequencies (e.g., associations with "war" vs. "taxes") and found Wikipedia's left-leaning slant exceeded Britannica's by 9-11% in aggregate scores, with conservative topics drawing more negative phrasing. Later analyses showed collaborative editing reduced some variance but not the overall skew, linked to editor demographics and sourcing from liberal-leaning outlets. This bias manifests in topic coverage and framing of politically charged issues. Sentiment analysis of economic policy articles revealed right-associated terms like "free market" paired with 15% more negative modifiers (e.g., "unregulated," "exploitative") than left-associated ones like "social welfare." Sourcing guidelines favor legacy media, 70-80% of which independent raters (e.g., AllSides Media Bias Chart) classify as center-left or left, embedding progressive framing while sidelining contrarian studies from conservative sources absent mainstream corroboration. Critics, including Wikipedia co-founder Larry Sanger, contend this perpetuates biases from academia and journalism, where surveys indicate 80-90% liberal self-identification among professors. Edit war data and user surveys provide further evidence. A 2023 review of 500 U.S. politician pages showed reversions of neutrality-flagged edits 2.5 times more likely to retain left-favorable language, such as downplaying Democratic scandals versus amplifying Republican ones, tied to editor clusters in urban, liberal-leaning areas. Although the neutral point of view policy seeks balance, enforcement patterns suggest it aligns with dominant editor ideologies, leading to underrepresentation of evidence challenging progressive views on issues like minimum wage or immigration economics.

Allegations of systemic bias in coverage of Israel and the Israeli-Palestinian conflict

Reports have documented allegations of coordinated anti-Israel editing campaigns influencing Wikipedia's coverage of Israel and the Israeli-Palestinian conflict. A March 2025 Anti-Defamation League (ADL) analysis of edits since 2002 identified a group of approximately 30 editors responsible for over 1 million edits to more than 10,000 related articles, showing coordination through 18 times more internal communications than control groups, tandem editing patterns, and biased consensus voting (e.g., 90% retention of critical content). The report cited instances of removing reputable sources on Palestinian violence and antisemitism while introducing narratives perceived as antisemitic or anti-Israel. A November 2025 Institute for National Security Studies (INSS) publication characterized English Wikipedia as an arena for anti-Israeli advocacy, highlighting agenda-driven editors' use of tools like Discord in groups such as "Tech for Palestine," policy applications (e.g., the 500/30 rule restricting new articles), and source disqualifications disadvantaging pro-Israel perspectives. Examples include reframing Zionism in articles to emphasize "colonization" and selective casualty reporting that allegedly omits context favoring Israel. These claims attribute biases to editor demographics and Wikimedia's "Knowledge Equity" priorities, paralleling broader systemic influence concerns.

Political editing scandals and external pressures

IP addresses linked to the United States Congress edited Wikipedia articles in 2006 to remove critical details on politicians, such as deleting references to Representative Bob Ney's involvement in the Abramoff scandal and softening Senator Strom Thurmond's segregationist history. WikiScanner revealed these conflicts of interest, prompting temporary blocks by administrators. By 2014, persistent vandalism and neutrality violations from House networks led to a two-week ban on edits from those IPs, affecting over 1,000 prior changes. Similar issues arose internationally: in 2007, Belgian Prime Minister's office staff excised negative political details, fueling debates on affiliated editing; 2020 saw coordinated proxy edits inserting biased content for U.S. candidates; and in 2026, PR firm Portland Communications faced allegations of commissioning favorable changes via subcontractors for clients including Qatar. In January 2025, the Arbitration Committee banned eight editors—six associated with pro-Palestinian views and two with pro-Israel views—from editing articles related to the Arab-Israeli conflict due to disruptive behavior, including personal attacks and source misrepresentation. State pressures include Russian fines on the Wikimedia Foundation for retaining Ukraine conflict content, Belarusian detentions of editors on related topics, and China's access block, with cases limited to organizational fines and isolated arrests. Post-2024 U.S. elections, Republican investigations in 2025 examined alleged coordinated bias, while conservative groups like the Heritage Foundation targeted perceived biased editors and officials threatened regulation over disputes. These events illustrate conflicts between Wikipedia's volunteer model and demands for alignment with partisan or official narratives.

Responses to bias allegations and reform attempts

The Wikimedia Foundation addresses systemic bias allegations mainly through its Neutral Point of View (NPOV) policy, which mandates fair, proportionate representation of viewpoints without endorsement. Enforced by volunteer editors, this policy functions as a self-correcting mechanism via community consensus and revisions, according to foundation officials. In March 2025, it launched a working group of active editors to standardize NPOV across projects, reaffirming neutrality amid scrutiny. Critics maintain these efforts fall short, as editor demographics sustain underrepresentation of conservative views. Co-founder Jimmy Wales has defended Wikipedia against political bias claims, stating in October 2025 interviews that content aligns with reliable sources and corrects via collective editing. He rejected notions of a "woke" or left-leaning tilt as exaggerated, highlighting the site's libertarian origins and volunteer basis to avert systemic capture. Responding to critics like Elon Musk, Wales called for greater editor diversity without policy overhauls, attributing minor leftward shifts—such as 9-11% more liberal language than Britannica—to source availability, not intent. Former co-founder Larry Sanger has advocated reforms to counter perceived left-wing bias, including a legislative process for policy changes to avoid unilateral foundation actions, as detailed in his October 2025 manifesto. His nine theses propose ending consensus-based decisions, allowing competing articles, lifting source blacklists, restoring the original neutrality policy, and repealing "Ignore all rules," cautioning that absent these, external regulation may follow. Sanger has testified before Congress and urged measures like mandatory neutrality audits, arguing Wikipedia's neutrality eroded since the mid-2000s. U.S. Republicans initiated investigations in August 2025 into alleged organized bias, including anti-Israel content and source manipulation. In October 2025, Senator Ted Cruz queried Wikipedia's funding and editing for bias transparency. These actions spurred calls for legislative oversight, which the foundation resists to preserve independence and open knowledge. Criticism of former CEO Katherine Maher's equity-over-truth statements has intensified demands for accountability, with Sanger decrying their policy influence. Proposed reforms have seen limited adoption, with debates persisting on whether volunteer governance adequately counters evident slant.

Major Controversies

Early scandals: Seigenthaler incident and PR manipulations

On May 26, 2005, an anonymous editor added false claims to the Wikipedia article on John Seigenthaler Sr., a former journalist and aide to Attorney General Robert F. Kennedy, stating he was a suspect in the assassinations of John F. Kennedy and Robert F. Kennedy and had lived in the Soviet Union before returning to the U.S. Created by Brian Chase as a prank against a colleague, the hoax remained uncorrected for four months until September 2005, despite partial revisions. Seigenthaler found the errors in late 2005 and criticized Wikipedia's open editing in a November 29 USA Today op-ed, highlighting risks of defamation and questioning its reliability, which drew media attention to hoax vulnerabilities. Investigative reporting identified Chase in December 2005; he apologized, noting the entry was meant as an internal joke. The case exemplified risks of malicious edits, spurring discussions on verification improvements. Jimmy Wales defended Wikipedia's self-correction while recognizing harm to victims. It prompted temporary bans on anonymous edits to living persons' biographies and greater focus on reliable sources. Around the same time, in early 2006, U.S. Congress staff edited articles from official IP addresses to remove scandals or inflate achievements in politicians' biographies, such as that of Congressman Marty Meehan. IP tracking revealed these conflicts of interest, leading to blocks on congressional edits and policies against undisclosed affiliated changes. These events challenged perceptions of neutrality, exposing limits in volunteer-based oversight.

Content disputes over sensitive topics

Content disputes over sensitive topics on Wikipedia often involve protracted edit wars, with editors reverting changes to enforce competing views on neutrality and sourcing. A 2013 analysis of over 10 million edits identified hotspots in articles on God, Jesus, George W. Bush, abortion, and global warming, accumulating thousands of reverts from ideological clashes over phrasing and viewpoints. Similar patterns occur cross-linguistically in religion (e.g., Muhammad), historical figures, and social issues like circumcision, where sustained battles exceed routine maintenance. These conflicts hinge on the neutral point of view policy, requiring balanced reliable sources, but intensify when mainstream media—often seen as left-leaning—is favored over primary data or dissenting academics. A specific instance concerns articles on the Holocaust in Poland. A 2023 study by historians Shira Klein and Jan Grabowski, published in the peer-reviewed Journal of Holocaust Research, analyzed dozens of Wikipedia articles and hundreds of related talk pages and dispute forums. The study alleged systematic distortions, including downplaying Polish societal involvement in the Holocaust, exaggerating Polish aid to Jews, and promoting stereotypes of Jewish complicity or collaboration, such as claims of disproportionate Jewish roles in communist security forces persecuting Poles or Jewish participation in pogroms. Examples cited included persistent inaccuracies like a hoax claiming most Auschwitz gas chamber victims were Poles rather than Jews. The research, stemming from resistance to a student's sourced edit on Jewish life in postwar Poland, received widespread media coverage, amplifying debates on Wikipedia's handling of Holocaust history. Some distortions have been corrected, but many remain. In response, Wikipedia's Arbitration Committee sanctioned editors in related disputes for disruptive conduct, such as overriding consensus and personal attacks, without adjudicating the accuracy of the content itself. Political and biographical articles on leaders and events see disputes over whitewashing or demonization, with revert wars documented for U.S. presidents and foreign policy. Recent examples include 2025 Arbitration Committee bans of eight editors—six pro-Palestinian and two pro-Israel—from Israeli-Palestinian conflict articles for disruptive behavior, such as overriding revisions, personal insults, and misrepresenting sources. Cycles of additions and deletions in topics like anarchism or U.S. history reflect broader divides, sometimes triggering page protections. Co-founder Larry Sanger contends that predominantly urban, educated, progressive editors systematically exclude conservative perspectives, framing disputes as censorship where alternative sources are deemed unreliable. Jimmy Wales responds that Wikipedia's strength is aggregating reputable sources without endorsing fringes, though edit wars on hot-button issues require stricter sourcing to curb POV pushing. Social controversies around gender, race, and identity have intensified, with edit wars over pronouns and classifications pitting self-identification guidelines against verifiability demands, resulting in locked articles and arbitrations. Sanger highlights topics like race and intelligence or feminism for one-sided coverage, where empirical challenges to progressive views face source blacklisting and harassment, eroding trust. These issues reveal tensions between open editing and quality control, as low barriers enable interest-group campaigns while reliance on biased sources provokes conflicts over corrections.

Recent developments: 2024–2026 investigations and credibility threats

In early 2024, resurfaced statements from former Wikimedia Foundation CEO Katherine Maher described Wikipedia's "free and open" ethos as a "white male Westernized construct" perpetuating power imbalances. Co-founder Larry Sanger cited these views, along with Maher's coordination with U.S. agencies on disinformation, as signs of ideological bias suppressing dissenting edits on sensitive topics. These issues spurred congressional scrutiny. In May 2024, the U.S. House Committee on Energy and Commerce requested Maher's testimony amid probes into bias in public broadcasting. By October 2025, the Senate Committee on Commerce, Science, and Transportation questioned Wikimedia CEO Maryana Iskander on whether policies favored ideological conformity over impartiality, highlighting reliance on left-leaning sources. Reports of foreign manipulation compounded concerns. In January 2024, Iranian state actors were found coordinating edits on Farsi Wikipedia to censor regime criticism and promote official narratives on human rights and conflicts. An investigation by Ashley Rindsberg in 2024–2025 exposed coordinated editing by approximately 40 users in the Discord group Tech for Palestine (TFP), who sought to advance pro-Palestinian perspectives and denigrate Israel. Reported alterations included omitting references to Hamas's original 1988 charter, mitigating mentions of Iran's human rights abuses, and establishing a new article depicting Zionism through racial lenses. A prominent case involved over 1,000 edits to the Hajj Amin al-Husseini article, aimed at diminishing his collaboration with Nazi Germany and excising related imagery and details. The Anti-Defamation League identified this as indicative of anti-Israel bias efforts. One key TFP-affiliated editor faced potential sanctions, including a lifetime ban, from Wikipedia's Arbitration Committee amid the unfolding controversy. Further reports of foreign manipulation emerged in January 2026, when Estonian and Lithuanian media and volunteers identified coordinated edits to English Wikipedia articles on Baltic history and biographies. These changes included altering birthplaces of prominent figures, such as Kaja Kallas and Gitanas Nausėda, from "Estonia" or "Lithuania" to "Estonian SSR, Soviet Union" or "Lithuanian SSR," with one user modifying nearly 600 Estonian profiles in extended sessions exceeding 21 hours. Observed by journalists like Ronald Liive and Wikimedia Estonia's Robert Treufeldt, as well as Lithuanian outlets, the edits were described as systematic efforts to promote pro-Soviet narratives, potentially legitimizing the occupation of the Baltics—a period viewed by Estonia and Lithuania as illegal despite Soviet claims. Reversion attempts faced challenges, including editor bans for alleged nationalism and article locks. In August 2025, House Oversight Committee Republicans investigated organized manipulation, targeting antisemitic, anti-Israel, and pro-adversary content via suspicious accounts and thousands of edits. Wikimedia cited editor privacy in resisting full disclosure of IP addresses and histories. Conservative groups responded to perceived biases. In January 2025, the Heritage Foundation planned to expose editors involved in antisemitic or ideological distortions, amid criticisms of sourcing practices. This drew accusations of threatening anonymity, while proponents sought reforms like editor verification. Wikimedia's 2025 WikiCredCon focused on harassment defenses rather than bias. On October 27, 2025, xAI launched Grokipedia, an AI-generated encyclopedia challenging Wikipedia on bias grounds. In January 2026, The Bureau of Investigative Journalism reported that London PR firm Portland Communications had outsourced favorable Wikipedia edits for clients including governments and billionaires.

Cultural and Societal Impact

Wikipedia's readership grew significantly since its inception, peaking in the early 2020s, but has declined recently due to generative AI tools and search engine changes that offer direct summaries, reducing site visits. In 2023, the English Wikipedia recorded about 92 billion page views; all Wikimedia projects totaled 296 billion in 2024. Traffic metrics show a downward trend: daily visits to Wikipedia.org fell over 14 percent from 2022 to 2025, with user numbers dropping 16.5 percent, while 2025 saw an 8 percent decline in human-generated views linked to AI integrations in search and social platforms. December 2024 recorded 11 billion total page views across editions, a 1.38 percent monthly decrease. These shifts indicate greater reliance on alternative knowledge sources, though Wikipedia ranks among the top 10 global websites by traffic. Access primarily occurs via web browsers, with mobile devices driving most traffic (about 76.86 percent of views versus 23.14 percent from desktops). Optimized desktop and mobile sites exist, the latter with responsive design for touch and low bandwidth. iOS and Android apps enable offline reading, search, and multimedia, but browser access predominates; 2023 estimates showed over 46 million daily mobile accesses. APIs support third-party integrations, bots, and services like search engines. Offline options include data dumps and mirrors for analysis without connectivity. Voice assistants reference Wikipedia via partnerships, but direct visits remain central. In 2025, AI-generated encyclopedias like xAI's Grokipedia emerged, using large language models to generate and rewrite articles, with humans mainly providing error feedback. This contrasts with Wikipedia's volunteer model and raises questions about AI-trained-on-Wikipedia systems handling authorship.

Influence on public perception and misinformation

Wikipedia's high visibility in search engines, such as Google's featured snippets, shapes initial public perceptions by delivering concise summaries for factual queries. In December 2024, the site attracted 11 billion page views, reaching a global audience that often views entries as authoritative overviews. This influence extends to professionals; a 2022 MIT study found that edits to articles on legal cases shifted judges' citations and decisions. Systemic content biases, especially left-leaning framing on political topics, risk distorting perceptions. A 2024 Manhattan Institute analysis of over 1,000 articles identified disproportionate negative sentiment toward right-of-center figures compared to left-leaning ones. Such patterns stem from the progressive demographics of active editors, potentially reinforcing selective narratives aligned with academia and media biases. The open-editing model promotes corrections but allows misinformation to persist through edit wars and coordinated efforts, impacting discourse. Examples include Holocaust entry distortions advancing nationalist claims and propaganda insertions by foreign actors, as revealed in 2025 U.S. House investigations. Responses include rapid reversions and initiatives like the 2025 WikiCredCon conference, which updated guidelines on disinformation, alongside student editing programs enhancing accuracy. Biased articles can foster hindsight bias, altering event interpretations per experimental findings. Scraped content also trains AI models, propagating biases into automated outputs encountered by billions. While policies prohibit falsehoods, volunteer-driven oversight leaves gaps where ideological conformity limits challenges, though diversity efforts aim to foster balanced editing. The Wikimedia Foundation has faced legal challenges over user-generated content, invoking Section 230 of the Communications Decency Act for platform immunity. Courts have ordered removals in defamation cases; a 2019 German ruling required deleting a defamatory edit from an article's history, pitting right-to-be-forgotten laws against archival transparency. In February 2025, a German court dismissed a defamation suit by a Pakistani citizen against Wikimedia, shielding volunteers from forum-shopping while highlighting global liability risks. Wikimedia has contested regulations seen as infringing moderation autonomy. In July 2024, the Foundation urged the U.S. Supreme Court to invalidate Texas and Florida laws on social media content moderation, arguing they conflict with community governance and First Amendment rights. In August 2025, a UK High Court rejected Wikimedia's challenge to Online Safety Act verification requirements, which risked exposing anonymous contributors; the Foundation announced in September 2025 it would not appeal. Conversely, in September 2025, the Paris Civil Court upheld Wikimedia's position in a freedom of expression case. Privacy concerns arise from IP address logging for all edits, which enables traceability but exposes pseudonymous contributors to subpoenas, harassment, or doxxing. Tor and VPN users have reported vulnerabilities, with IP blocks potentially discouraging edits; a 2016 survey of anonymous editors found prevalent fears of exposure on sensitive topics. In response to 2025 doxxing threats, the Foundation implemented temporary accounts for logged-out editors starting late 2024, fully effective by November 2025; these attribute edits to temporary usernames rather than IPs, reducing identification risks. Policies bar off-wiki personal data disclosures, enforced via community norms, though incidents like 2013 IP-traced congressional edits demonstrate potential leaks. Explicit content policies permit sexual imagery if contextually relevant and reliably sourced, while prohibiting hardcore pornography to weigh notability against offensiveness. A 2010 Commons dispute over nude images prompted Jimmy Wales to relinquish founder privileges amid concerns over child access; subsequent rules reject blanket removals based on shock value, favoring sourcing over filters. Guidelines broadly define sexual content and urge caution for obscenity laws, but search results surfacing erotic material—including historical uploads—have sparked complaints on minor accessibility and institutional blocks. Aiming for neutrality, these approaches have drawn external critique, with data showing high views for explicit articles absent universal warnings.

Wikimedia projects overview

The Wikimedia Foundation, established in 2003 as a non-profit, hosts volunteer-edited online projects powered by the open-source MediaWiki software and licensed under Creative Commons Attribution-ShareAlike (CC BY-SA). These include encyclopedic content, multimedia repositories, linguistic resources, and databases, serving over 18 billion monthly page views in fiscal year 2024-2025. Unlike commercial platforms, they depend on community contributions without editorial gatekeeping, allowing rapid updates but risking errors, vandalism, and disputes handled via consensus policies. Central is Wikipedia, launched January 15, 2001, as a multilingual encyclopedia with over 7.1 million articles in English as of early 2026, plus editions in more than 300 languages. Linguistic projects include Wiktionary (2002), a collaborative dictionary and thesaurus with definitions, etymologies, and translations; Wikiquote (2003), compiling verifiable quotations; and Wikisource (2003), archiving public-domain texts like books and historical documents. Educational ones are Wikibooks (2003) for open textbooks and Wikiversity (2006) for learning resources and research without accreditation. Multimedia efforts feature Wikimedia Commons (2004), a repository of over 100 million freely usable files such as images and videos, shared across projects. Wikidata (2012) offers an editable database of over 120 million items for structured data, supporting queries and reducing redundancy in Wikipedia infoboxes. Niche projects encompass Wikispecies (2004) for taxonomy on over 1.5 million species and Wikivoyage (2012) for travel guides. Wikinews (2004) covers news via citizen reports under neutrality guidelines, though with lower activity. Projects interconnect, with Wikidata providing facts for Wikipedia and Commons visuals, forming a modular knowledge base. In 2025, over 220,000 active contributors participate globally, backed by Foundation grants over $9 million yearly for initiatives, amid efforts to address male and Western skews. Experimental projects like Wikifunctions (2023) support computation for Abstract Wikipedia. The model emphasizes verifiability and openness, with editable content safeguarded by histories and administrators.

Commercial and derivative uses

Wikipedia content is licensed under the Creative Commons Attribution-ShareAlike (CC BY-SA) 3.0 Unported License (or version 4.0 for newer contributions), allowing commercial use, modification, and distribution with attribution and share-alike requirements for derivatives. This framework, in place since inception, enables reuse of voluntarily contributed text and media while preventing proprietary enclosures through the share-alike clause. Forks and alternative encyclopedic projects have utilized exported Wikipedia content under these licensing terms, including Citizendium, founded in 2006 by Wikipedia co-founder Larry Sanger as an expert-reviewed encyclopedia, and Encyc, a smaller open-content project that has adapted material from Wikipedia data dumps. Commercial entities integrate Wikipedia content into search engines, virtual assistants, and databases. Google uses excerpts in its Knowledge Graph and search results; Apple incorporates data into Siri; Amazon and Facebook have tested infoboxes for fact-checking. These rely on scraping or APIs but require attribution, such as links to source articles. Wikipedia also serves as training and evaluation data for large language models and AI assistants, including chatbots and tools like Grokipedia. Such systems draw on Wikipedia text for facts and style, potentially reproducing its strengths and biases in outputs, even without attribution, which raises issues of credit, responsibility, and neutrality in proprietary models. Derivative works include offline archives, mobile apps, and printed books adapting Wikipedia material under CC BY-SA terms. Kiwix provides downloadable archives for low-connectivity areas, while commercial guides compile article summaries. In 2021, the Wikimedia Foundation launched Wikimedia Enterprise, a paid API for structured, low-latency access aimed at high-volume users like Google and the Internet Archive, to cover costs without affecting public access. Google subscribed in 2022; the service reached profitability in fiscal year 2024–2025, with a cumulative net profit of $646,000 by June 2025. License restrictions bar non-share-alike proprietary uses, such as paywalled derivatives without free access to modifications. The Foundation bans implied endorsements and systematic image hotlinking to manage bandwidth, though content remains free. Non-compliance prompts takedown requests for unauthorized mirrors, balancing utility with nonprofit sustainability.

References

  1. https://meta.wikimedia.org/wiki/Wikipedia
  2. https://meta.wikimedia.org/wiki/Wikipedia_milestones
  3. https://en.wikinews.org/wiki/Wikipedia_suffers_outage
  4. https://meta.wikimedia.org/wiki/Overall_timeline
  5. https://en.wikiquote.org/wiki/Wikiquote:History_of_Wikiquote
  6. https://commons.wikimedia.org/wiki/Commons:Welcome
  7. https://foundation.wikimedia.org/wiki/Memory:Timeline/en
  8. https://www.mediawiki.org/wiki/Extension:WikiEditor
  9. https://www.mediawiki.org/wiki/Manual:Tools
  10. https://www.mediawiki.org/wiki/VisualEditor
  11. https://www.mediawiki.org/wiki/ORES
  12. https://foundation.wikimedia.org/wiki/Home
  13. https://commons.wikimedia.org/wiki/Commons:Administrators
  14. https://meta.wikimedia.org/wiki/Wikimedia_power_structure
  15. https://meta.wikimedia.org/wiki/Research:Wikipedia_Administrator_Recruitment%2C_Retention%2C_and_Attrition/Report
  16. https://meta.wikimedia.org/wiki/Oversight_policy
  17. https://www.mediawiki.org/wiki/Article_feedback/Version_5/Help/Oversighters
  18. https://wikiconference.org/wiki/Submissions:2023/Reflections_on_Eight_Years_at_ArbCom:_Policy%2C_Structure%2C_and_Reform
  19. https://meta.wikimedia.org/wiki/Research:Revision_scoring_as_a_service/Archived
  20. https://meta.wikimedia.org/wiki/Grants:Programs/Wikimedia_Research_%2526_Technology_Fund/Wikimedia_Research_Fund
  21. https://meta.wikimedia.org/wiki/Community_Insights/Community_Insights_2024_Report
  22. https://www.mediawiki.org/wiki/Manual:MediaWiki_architecture
  23. https://www.mediawiki.org/wiki/MediaWiki_history
  24. https://www.mediawiki.org/wiki/Manual:MediaWiki_feature_list
  25. https://www.mediawiki.org/wiki/Manual:Extensions
  26. https://www.mediawiki.org/wiki/Extension:VisualEditor
  27. https://wikitech.wikimedia.org/wiki/Data_centers
  28. https://meta.wikimedia.org/wiki/Wikimedia_servers
  29. https://commons.wikimedia.org/wiki/Category:Wikipedia_editor_statistics
  30. https://commons.wikimedia.org/wiki/Help:Sexual_content
  31. https://meta.wikimedia.org/wiki/Wikimedia_projects
  32. https://foundation.wikimedia.org/wiki/Policy:Terms_of_Use
  33. https://commons.wikimedia.org/wiki/Commons:Reusing_content_outside_Wikimedia
Add your contribution
Related Hubs
Contribute something
User Avatar
No comments yet.