Research:Interpolating quality dynamics in Wikipedia and demonstrating the Keilana Effect

From Meta, a Wikimedia project coordination wiki
Jump to: navigation, search
VisualEditor - Icon - Check.svg
This page documents a completed research project.


For open, volunteer generated content like Wikipedia, quality is a prominent concern. To measure Wikipedia's quality, researchers have historically relied on expert evaluation or assessments of article quality by Wikipedians themselves. While both of these methods have proven effective for answering many questions about Wikipedia's quality and processes, they are both problematic: expert evaluation is expensive and Wikipedian quality assessments are sporadic and unpredictable. Studies that explore Wikipedia's quality level or the processes that result in quality improvements have only examined small snapshots of Wikipedia and often rely on complex propensity models to deal with the unpredictable nature of Wikipedians' own assessments. In this paper, I describe a method for measuring article quality in Wikipedia historically and at a finer granularity than was previously possible. I use this method to demonstrate an important coverage dynamic in Wikipedia (specifically, articles about Women Scientists) and offer this method, dataset, and open API to the research community studying Wikipedia quality dynamics.

Introduction[edit]

Since its inception, quality has been the most prominent concern about the future of Wikipedia. After all, how can high-quality information artifacts be produced when there's literally no restriction on who is allowed to contribute? Over the past 12 years (as of 2017), the research literature around Wikipedia has advanced our understanding of the open encyclopedia's quality and the processes by which crowds of volunteers can manage such an information artifact.

Our first major leaps in understanding of Wikipedia's quality dynamics happened around the time that Jim Giles published a report in Nature (2005)[1] that surprised the world. This seminal report showed that Wikipedia's coverage of scientific content compared favorably (and in some ways, better) to the more dominant encyclopedias that were used. Since that surprising result, researchers have been pushing toward greater understanding of how open, volunteer processes could have led to this point.

While we do know a lot about quality dynamics in Wikipedia, there are still many questions that remain. Where are Wikipedia's coverage gaps? What types of editing patterns lead to efficient quality improvements? These questions are important for the science and for the practice of Wikipedians. In this paper, we'll detail the development of a measurement strategy and the release of a public dataset that I believe will make answering these questions far easier than ever before. In the following sections, I'll summarize the state of the art with regards to quality dynamics and measurement in Wikipedia, I'll explain my measurement methodology, and I'll provide a demonstration analysis that gives novel insights in the coverage gaps and quality dynamics of articles about Women Scientists.

The quality of English Wikipedia[edit]

Giles' study set in motion a series of studies examining the quality of several different subject spaces in the encyclopedia. Mesgari et al. provides an excellent survey of this work[2], but for the purposes of this study, we'll summarize their key findings as follows:

  • Wikipedia's coverage is broad and comprehensive.
  • Wikipedia has a high level of currency -- especially with regards to topics of interest to the public.
  • Wikipedia's accuracy compares favorably with traditional encyclopedias.

Yet the story isn't all one of pure success. Given the conclusions drawn from years of research, Wikipedia's coverage of most major topics that are traditionally covered in an encyclopedia is hard to question. However, not all topics are covered as completely as others. Gaps in Wikipedia's content coverage are concerning because of the encyclopedia's dominance as an information resource. 600 million people read the encyclopedia every month[3]. With an information source of such popular use and apparent completeness, any topic that is covered less completely than others (or not at all) could imply less importance or relevance to certain types of knowledges. Based on past work, we know that Wikipedia lacks in coverage of topics about Women[4], of interest to Women[5], about rural and non-Western geographies[6][7] and cultures[8]. Thus, understanding where Wikipedia's coverage is lacking and which initiatives are effective in closing coverage gaps is critical to the long term success of the project and of critical concern for humanity.

Quality dynamics in Wikipedia[edit]

But how did Wikipedia arrive at such a high quality level? There are two major threads of research in this area: counter-vandalism and article quality dynamics.

Counter-vandalism. Due to its open nature, Wikipedia is under constant threat of vandalism and other types of damaging changes. At its most basic level, Wikipedia protects against damage by maintaining a history of all versions of every article. This allows "patrollers" to easily clean up damage whenever it is discovered. A revert[footnote 1] is a special type of edit that removes the changes of another edit -- usually by restoring the last good version of the article.

But reverts aren't the whole story of counter-vandalism in Wikipedia. There's a whole complex network of communities of practices (like the counter-vandalism unit[footnote 2]), policies (like the 3-revert-rule[footnote 3]), and highly advanced automated tools that support editors in finding and quickly removing damaging contributions to Wikipedia[9][10]. Together, this socio-technical system fills the infrastructural role of keeping out the bad stuff -- ensuring that Wikipedia's open nature does not cause the quality to decay into nonsense.

Article quality dynamics. While counter-vandalism and other type of patrolling work helps keep the bad out of the Wiki, there are other dynamics and processes at play that determine which articles will increase in quality efficiently. Stvilia et al. first hypothesized a framework for how high quality content was generated in Wikipedia[11]. Essentially, volunteers will allow their interests to drive where they contribute, and through building on to each others work, articles grow and are gradually refined. This interest-driven pattern could likely explain how Wikipedia's demographic gaps have led to coverage and quality gaps[5].

Through Kittur et al.'s work, we know that group structure and dynamics play an important role in the efficiency of article improvement. They showed that articles where few editors make most of the edits and many editors contribute few tend to increase in quality more quickly than articles where the distribution of edits per editor is more uniform[12]. Arazy & Nov extend this conclusion by showing that certain types of editor experience are critical to article improvement -- that it's not only important to have diversity in contribution rates but also diversity in the total experience level of Wikipedia editors[13]. In order to draw these conclusions, the authors of both of these studies needed to operationalize measurements of quality in Wikipedia and compare the configurations of editors and their contribution types to changes in measured quality.

Methods for measuring article quality[edit]

summary
FA Professional, outstanding, and thorough; a definitive source for encyclopedic information.
GA Useful to nearly all readers, with no obvious problems; approaching (but not equalling) the quality of a professional encyclopedia.
B Readers are not left wanting, although the content may not be complete enough to satisfy a serious student or researcher.
C Useful to a casual reader, but would not provide a complete picture for even a moderately detailed study.
Start Provides some meaningful content, but most readers will need more.
Stub Provides very little meaningful content; may be little more than a dictionary definition.

Wikipedia 1.0 (wp10) assessment rating scale. The 6 levels of assessment are provided with the "readers experience" description copied from Wikipedia[footnote 4]. Each level also has a well-defined detailed set of assessment criteria that involves discussion of formatting, coverage, and proper sourcing.

Wikipedia assessment ratings. Wikipedians assign quality assessments to articles based on a scale that was originally developed to produce an official "1.0" version of Wikipedia[footnote 5]. This scale has since been adopted by WikiProjects[footnote 6], self-organized subject-matter focused working groups on Wikipedia (e.g. WikiProject Video Games, WikiProject Medicine, and WikiProject Breakfast). #Wikipedia 1.0 (wp10) assessment rating scale shows the rating scale with defunct old grades ("A", "B+", etc.) removed. Wikipedians use this scale to track progress towards content coverage goals and to build work lists (e.g. WikiProject Medicine's tasks[footnote 7]).

Both Kittur et al.[12] and Arazy & Nov[13] used article quality assessments provided by Wikipedians to track article improvements and to find correlations with editor activity and experience level characteristics that they use to draw conclusions about successful collaboration patterns. Regretfully, the process by which Wikipedians assess and re-assess articles is unpredictable. We can be relatively sure of the quality level of an article at the time it was assessed, but there's no clear way to know when, exactly, the quality level of an article actually changed. The aforementioned studies use a set of complex propensity modeling strategies like Heckman Correction[footnote 8] to minimize the possibility that the correlations they observe were simply due to the assessment behavior of Wikipedians and not actual changes in article quality. The modeling of correction is difficult to optimize and evaluate independently of the effects that editor collaboration patterns may have had on those quality changes. Further, past work has been limited to the times at which Wikipedians were performing assessments of articles. Assessments didn't become common until 2006 (5 years after Wikipedia's inception) and the criteria by which articles are assessed has been undergoing changes since then. Articles that were "B" class in 2006 would likely now be classified as a "Start" class due to insufficient inline references. #B-class criteria comparison shows the change in B-class criteria between 2006 and 2017.

Further, the fact that assessments are sparse and unpredictable also means that the assessments are often out of date. This is an operational issue for Wikipedians too. WikiProject groups organize re-assessment drives to bring the assessments of articles under their purview into compliance, but this is a never-ending process. Since Wikipedia contributions come from anyone (member of the WikiProject or not) and re-assessments are uncommon despite regular coordinated efforts, the overall assessments of Wikipedia articles are perpetually out of date and reassessments are a never-ending source of new work.

Modeling article quality. There's great promise in using machine prediction models to measure the quality of articles automatically. There is a lot of research that has explored the development of such models. The modeling work I do for the purposes of this study and in producing this dataset is largely based on work by Warncke-Wang et al. who developed a relatively high fitness quality model based on explicitly "actionable" features[14]. I recommend reviewing their discussion of related work to learn how this research area has been developing. Warncke-Wang et al. challenge the state of the literature by defining a set actionable features as those that an editor may directly affect and draw a contrast to "inactionable" sources of signal. Many features used by Stvilia et al.[15] (who notably first published about article quality modeling in Wikipedia) like the time since the article was created or the number of editors who have worked on the article cannot be directly affected by an editor, so those features only incidentally are predictive of the quality level of an article. In an effort to target truly actionable features, Warnke-Wang et al. trained their model based on measurements of the text content of the article -- the primary artifact that editors directly affect when they contribute to Wikipedia. I agree that a useful model of quality in Wikipedia -- both for research needs and Wikipedians' operational concerns shouldn't weight predictions by incidentally predictive features and instead should be based entirely on characteristics of the article's content -- the same characteristics that Wikipedians evaluate when they make assessments.

2006[footnote 9] Has several of the elements described in "start", usually a majority of the material needed for a completed article. Nonetheless, it has significant gaps or missing elements or references, needs substantial editing for English language usage and/or clarity, balance of content, or contains other policy problems such as copyright, Neutral Point Of View (NPOV) or No Original Research (NOR). With NPOV a well written B-class may correspond to the "Wikipedia 0.5" or "usable" standard. Articles that are close to GA status but don't meet the Good article criteria should be B- or Start-class articles.
2017[footnote 10]
  1. The article is suitably referenced, with inline citations. It has reliable sources, and any important or controversial material which is likely to be challenged is cited. Any format of inline citation is acceptable: the use of <ref> tags and citation templates such as {{cite web}} is optional.
  2. The article reasonably covers the topic, and does not contain obvious omissions or inaccuracies. It contains a large proportion of the material necessary for an A-Class article, although some sections may need expansion, and some less important topics may be missing.
  3. The article has a defined structure. Content should be organized into groups of related material, including a lead section and all the sections that can reasonably be included in an article of its kind.
  4. The article is reasonably well-written. The prose contains no major grammatical errors and flows sensibly, but it does not need to be "brilliant". The Manual of Style does not need to be followed rigorously.
  5. The article contains supporting materials where appropriate. Illustrations are encouraged, though not required. Diagrams and an infobox etc. should be included where they are relevant and useful to the content.
  6. The article presents its content in an appropriately understandable way. It is written with as broad an audience in mind as possible. Although Wikipedia is more than just a general encyclopedia, the article should not assume unnecessary technical background and technical terms should be explained or avoided where possible.

B-class criteria comparison. The B-class criteria is compared between December of 2006 and April of 2017. Note the increase in detail generally and the discussion of the importance of inline citations specifically. Also note that "C" class didn't exist as of December 2006.

Modeling actionable article quality[edit]

n ROC-AUC acc (within-1) ~Stub ~Start ~C ~B ~GA ~FA
Stub 4968 98.5% 85.5% (99.2%) 4247 685 27 9 0 0
Start 4982 91.2% 64.3% (91.5%) 600 3205 754 358 58 7
C 4994 86.6% 48.9% (86.1%) 44 870 2443 986 558 93
B 4990 84.1% 40.3% (79.8%) 51 617 1258 2012 710 342
GA 5000 92.1% 62.7% (93.3%) 1 19 313 304 3135 1228
FA 4454 96.1% 77.4% (94.4%) 5 2 23 220 757 3447

ORES wp10 fitness statistics. The fitness statistics of ORES "wp10" model based on 10-fold cross-validation are shown in table format. Note that "within-1" accuracy represents the per-class accuracy measure where being off by one is considered a successful prediction. The overall accuracy is 62.9% and the within-1 accuracy is 90.7%. These figures compare favorably to [16] of 58.2% and 89.5% respectively. I've included the ROC-AUC metric to demonstrate that the probability estimates made by the model are generally useful. Like [16], I also find that "B" and "C" class show lower overall fitness than other classes.

In order to enable better understanding of the dynamics of article quality in Wikipedia, I sought to develop a method that would allow for granular analysis of the quality level of articles. Warncke-Wang et al.'s "actionable" modeling strategy seems particularly suitable to the task due to the high level of fitness they demonstrate (matching Wikipedians' own assessments) and its reliance on characteristics of the text content of the article (as opposed to external characteristics). I worked with Dr. Warncke-Wang to re-implement this model in the ORES system[footnote 11] and to implement a few minor improvements since he and his collaborators last published about improvements to the model[16]. I used the related open dataset[17] of assessments to train and test this model. For a full specification of the features used in prediction including scaling and controlling features, see the code[footnote 12]. ORES currently uses a GradientBoosting algorithm with estimators=700, max-depth=7, max-features=log2, and loss=deviance[footnote 13]. #ORES wp10 fitness statistics presents the overall prediction fitness of the re-implemented model.

Modeling quality changes. I hypothesized that, at a large enough timescale, this "actionable model" would be able to track quality dynamics -- the changes in article quality over time. After all, if the model is able to match Wikipedians' assessments, it should be able to fill in the gaps between assessments as well. I suspected that a large timescale that covered several revisions would be necessary for measuring article quality dynamics accurately, because it takes time for Wikipedia's counter-vandalism systems to work. I.e. vandalism and other damaging edits to an article might also favorably change the features extracted since they are arguably quite basic (e.g. number of headers and number of image links &endash; see [14]). It would be inappropriate to consider such an edit to have increased the quality level of the article. By waiting a substantial time period between automatic quality predictions, Wikipedians' natural quality control processes can run their course.

There are two reasons why I have concluded that this strategy is working in practice: (1) I performed a set of basic spot-checks on hand-picked articles focusing on substantial quality changes and (2) I released this prediction model publicly 2 years ago and since then many users have told my collaborators and I that they find it to be effective and useful for supporting their work. I was actually quite surprised to find that the large timescale between assessments seems to not be necessary at all. Users report that the model is accurate and useful even when used to score every version of a page and that, in general, vandalism appears as either no change or a decrease in article quality as intended[18].

"prediction": "GA",
"probability": {
  "Stub": 0.0019,
  "Start": 0.0132,
  "C": 0.1252,
  "B": 0.2090,
  "GA": 0.3345,
  "FA": 0.3162
}

Quality prediction for Murie Curie. The article quality prediction an revision of en:Murie Curie saved on April 4th, 2017 is presented. See https://ores.wikimedia.org/v2/scores/enwiki/wp10/773753742.

One additional request that I got from Wikipedia editors was to provide a mechanism for getting between-class quality predictions. So I developed a basic strategy that I refer to as the weighted sum of class predictions. I assume that the ordinal article quality scale developed by Wikipedia editors is roughly cardinal and evenly spaced. To arrive at the weighted sum measurement, multiply the prediction probability for each class by an enumeration of ordered classes starting at zero (0) for Stub and ending at five (5) for FA.


Weighted sum. This equation weights each *step* in quality the same by multiplying an enumerated integer by the prediction probability.

When the model predicts that a version of an article is a Stub with 100% confidence, this weighted sum would be 0. If the prediction were split 50% Start and 50% C, the weighted sum would be 1.5. The weighted sum of the prediction demonstrated in #Quality prediction for Murie Curie would be 3.8096 -- slightly on the "B" side of "GA". Again, my own spot checking and ORES' users confirm that this appears to be useful and consistent when applied to the history of articles.

Once I had concluded that the model operated consistently over time, I worked with my collaborators to generate a dataset that contains a predicted quality level for all articles in English Wikipedia at a monthly interval (~600m article-month predictions). This dataset[19] contains the highest probability quality class ("prediction") as well as the weighted sum measurement for each article-month ("weighed_sum").

The 3 manual assessments of the quality of en:Biology are plotted over time with a dashed line connecting them.
Biology's assessments. The 3 manual assessments of the quality of en:Biology are plotted over time with a dashed line connecting them.
The weighted_sum prediction for en:Biology as of the last revision at the beginning of each month is plotted.  Note how this plot shows much more nuanced detail about the development of the article than #Biology's assessments.
Biology's quality dynamics. The weighted_sum prediction for en:Biology as of the last revision at the beginning of each month is plotted. Note how this plot shows much more nuanced detail about the development of the article than #Biology's assessments.

Aggregated quality measures. Using this dataset, we can move up a level from articles and assess the quality of Wikipedia at an aggregate level -- as a whole or across interesting cross-sections. I employed two aggregation strategies for comparing the quality of article aggregates: the mean weighted sum, and proportions of articles falling into each predicted class. In order to give these measurements a useful denominator I use the total number of articles in the aggregate as of the the most recent month in the dataset. Since the number of articles is clearly monotonically increasing (article creations always outnumber deletions), the count in the last month is always the max number of articles for any month in the cross-section. Assuming that a Stub (the lowest quality class) is substantially more useful than no article at all so I assign a zero (0) weighted sum for all articles that have yet to be created in a particular month and increment the weighed sum for articles that exist by one (1) when generating the aggregate weighed sum. So if all articles were empty (like in the month before Jan 2001 -- inception of Wikipedia), the mean weighted sum is zero. If all articles were predicted to be 100% FA class, the mean weighted sum would be 6.

These assumptions are bold and clearly provide an incomplete view of the reality of Wikipedia. E.g. if one were to draw a cross section of Wikipedia that included only one FA-class article, that cross-section would get the maximum mean weighted sum. If in a following month, a Stub article were created, then the mean weighted sum would decrease even though there is more useful content. However, because this measure uses the number of articles in the last month as the denominator for all months, the mean weighed sum would also decrease historically, so this assumption is somewhat fair. Still, this assumption implies that Wikipedia will get no new articles after the date at which the dataset is generated. I feel that this is acceptable for comparing cross-sections/aggregates that have more than a trivial number of articles and that are also generated using normal Wikipedian processes. I'll discuss this complicated problem and propose some solutions for addressing it for #Future work. For the purposes of the demonstration analysis, I encourage you, dear reader, to judge this analysis by its interest as a demonstration of the novel measurement capabilities and how it enables us to visualize shifts in quality over time.

Demonstration: Coverage of Women Scientists in Wikipedia[edit]

When developing this measurement strategy, I found myself curious about the quality level of some cross-sections of Wikipedia where there were once known content coverage gaps. I decided to focus on coverage of Women Scientists in Wikipedia. Regretfully, Wikipedia's category[footnote 14] structure is notoriously useless for taking meaningful cross-sections of content[20]. Luckily, WikiProject Women Scientists are keen on making sure that all biographies about women scientists are tagged and tracked by their project templates[footnote 15]. This provides a useful means to identify all of the relevant biographies and thus, querying Wikimedia's replica databases for usage of their template is an efficient means for gathering this cross-section [footnote 16].

As described in the previous section, I generated aggregate metrics for both the entire Wikipedia (about 5 million) and just the articles covering Women Scientists (5,681). I then compared the aggregated measures for mean weighted sum and proportion of articles in each quality class to look for shifts between gaps (when the apparent quality of articles about women scientists is lower than the apparent quality of the rest of the encyclopedia) and surpluses.

The difference in mean weighted sum quality predictions for all wiki and articles about Women Scientists is plotted over time. Note the transition from red to blue represents the switch from a gap to a surplus.  Important dates for User:Keilana's initiatives are annotated with arrows.
Overall WikiProject Women Scientists vs all Wikipedia. The difference in mean weighted sum quality predictions for all wiki and articles about Women Scientists is plotted over time. Note the transition from red to blue represents the switch from a gap to a surplus. Important dates for User:Keilana's initiatives are annotated with arrows.
The proportion of articles falling into Empty, Stub, Start, and C-class predictions is plotted for articles tagged by WikiProject Women Scientists and all of Wikipedia.
Proportions of class prediction (empty - C). The proportion of articles falling into Empty, Stub, Start, and C-class predictions is plotted for articles tagged by WikiProject Women Scientists and all of Wikipedia.
The proportion of articles falling into B, GA, and FA-class predictions is plotted for articles tagged by WikiProject Women Scientists and all of Wikipedia.
Proportions of class prediction (B - FA). The proportion of articles falling into B, GA, and FA-class predictions is plotted for articles tagged by WikiProject Women Scientists and all of Wikipedia.

#Overall WikiProject Women Scientists vs all Wikipedia shows the difference between the mean weighted sum of article quality for all of Wikipedia and just the articles about Women Scientists. From the start of Wikipedia, a gap quickly develops and reaches the maximum of about a half a quality level around the middle of 2012. But after that point, the gap starts to rapidly close and a massive surplus begins to grow to a full quality class above the rest of the encyclopedia.

This dynamic roughly plays out in the same way for the proportions of articles in the set of predicted quality classes. #Proportions of class prediction (empty - C) and #Proportions of class prediction (B - FA) clearly show that the proportion of Start, C, and GA-class articles shows a similar gap and surplus period. Stub-class articles about Women Scientists seem to continue to fall behind the rest of the wiki, but there's still a noticeable increase in the proportion around the beginning of User:Keilana's initiatives (discussed in the next section). It seems that the growth dynamics of B-class articles is shaped very differently from the rest of the quality prediction classes. Surprisingly, there seems to be no surplus in the proportion of articles that fall into FA-class. It seems that this is the only quality class where not noticeable shift occurs -- the gap between Women Scientist bios and the rest of Wikipedia only widens over time.

Discussion[edit]

When I first saw this gap-to-surplus shift, I honestly had no idea what could have caused it. In an effort to share my analysis with a larger community, I presented the preliminary results in a prominent public forum for Wikimedia-related research projects -- the Wikimedia Research Showcase[footnote 17]. During the presentation, Wikipedia editors who ran a series of outreach initiatives to bring attention to biographies about Women Scientists reached out to us to let us know that the beginning of their initiatives corresponded to the beginning of the period of the shift. User:Keilana (aka Emily Temple-Wood) and her collaborators received substantial attention from the western media for their work to increase coverage of Women Scientists in Wikipedia[21]. At the most basic level, my analysis of content quality and coverage of this topic in Wikipedia seems to suggest that their work has had a very large effect.

I haven't been able to determine what might explain the unusual shape of the trajectory for the proportion of B-class articles. One hypothesis is related to the process by which GA and FA-class articles are assessed. Unlike the lower quality classes, GA and FA-class articles go through a formal peer review process and are promoted in other places in the wiki. It's possible that there's a large incentive to bring any B-class article to the next stage (GA) since it's only one step away and it is rewarded with public recognition.

The smaller proportion of articles that are predicted to be at the FA-class level is maybe a bit more concerning for the coverage of content about Women Scientists. One possibility is that the model is biased against articles about Women Scientists. There's a well discussed bias against coverage of Women Scientists in the type of reference work that Wikipedia tends to cite. On one had, it could be that articles about Women Scientists are inherently shorter and just look to be lower quality than other articles in the wiki. It could also be that it's hard to write a truly comprehensive article about a Woman Scientist for the same reason. It turns out that Wikipedians' own assessments suggest that there's a similar proportion of FA-class articles about Women Scientists (7/5681 = 0.12%[22]) as there are across the entire wiki (6k/5m = 0.12%[23]), so it could be possible that the model is showing a slight bias here.

Conclusions and future work[edit]

In this paper, I introduce a novel measurement strategy. By repeatedly applying the modeling strategy developed by Warncke-Wang et al to Wikipedia articles, I can examine the dynamics of their quality at a granularity that was not possible before. Using this measurement strategy, I demonstrate that topics of high importance to Wikimedia and Wikipedia editors can be examined in novel ways and new insights can be gained. I've also linked directly to a public release of open licensed data and an openly available API for generating new predictions that I hope will enable others to more easily explore similar analyses and analysis strategies.

To download the dataset, see https://doi.org/10.6084/m9.figshare.3859800
To access the prediction API, see https://ores.wikimedia.org/

Limitations[edit]

As in past work that used Wikipedia's quality assessments as an outcome measure, we're assuming that Wikipedians' notions of quality correspond to some general, fundamental true quality. But quality is a more complex concept that is arguably context- and purpose-dependent. For some users and uses of Wikipedia, a Stub may be perfectly acceptable (e.g. settling a bar bet), while for others an FA would be necessary to really get a complete overview of the subject. We've decided to model Wikipedians' own quality ratings because their context and purpose is the construction of an encyclopedia. While this may not fit all contexts and purposes, it does seem relevant to a substantial cross-section of the research literature that is concerned about the efficiency of open production processes and it is clearly useful for ORES' intended audience -- Wikipedians.

As was mentioned in #Modeling actionable article quality, measuring the completeness of a cross-section of Wikipedia is difficult when it's not clear how many articles the cross-section should have. We've chosen to operate on the assumption that the number of articles present in the cross section at the time of measurement is a useful denominator to use historically. Future work could explore the development of a more reasonable denominator by taking advantage of indexes of known notable topics that might eventually have an article in Wikipedia. Notably, Emijrp has already put substantial effort into just such an initiative[24]. He estimates that the current total number of articles in English Wikipedia (about 5 million) represents only about 5% of all of the articles that will eventually be covered in the encyclopedia.

Future work[edit]

Beyond this demonstration, I would like to use this dataset and method to explore article quality inflection points that occur in other cross sections of Wikipedia. As was discussed, I was surprised to learn about the effectiveness of Keilana's efforts since I had discovered the inflection that started around her initiatives independently. Other "Keilana Effect"s should also become evident when re-applying this analysis method. Doing so would allow the Wikimedia Foundation and other organizations that support initiatives (like the grant that supported Keilana) to bring attention and resources to efforts that are already working far better than I had expected.

Beyond looking for inflection points, we can simply look for gaps to target effort towards. In the past, determining where a coverage gap might be and how big it is, would require massive effort on the part of Wikipedians to assess thousands of articles. And worse, it would be nearly impossible to find out when the gap originated and whether or not it seemed to be widening or closing. Using this dataset, researchers can explore the largest gaps in coverage in Wikipedia and organizations like the Wikimedia Foundation can target new outreach campaigns (like Inspire campaigns[footnote 18]) to improve coverage in areas with known gaps.

This modeling approach also makes research around the nature of Wikipedia's quality dynamics easier. For example, using this modeling strategy, one could replicate or extend the modeling work performed by Kittur et al.[12] and Arazy & Nov[13] without resorting to propensity modeling, with far more observations, and in time scales where Wikipedian assessments are sparse. By developing and releasing this dataset, I intend to make exactly this type of work easier.

Acknowledgements[edit]

I would like to thank Morten Warncke-Wang who has been a thoughtful collaborator in helping us re-implement his model in the ORES system. Thanks to Amir Sarabadani who contributed substantially to the generation of the article quality prediction dataset used in this study[19]. I would like also acknowledge the Wikimedia Foundation who directly supported the research described in this study and Jonathan T. Morgan for feedback and proofreading.

Bibliography[edit]

References[edit]

  1. a b Giles, J. (2005). Internet encyclopaedias go head to head.
  2. a b Mesgari, M., Okoli, C., Mehdi, M., Nielsen, F. Å., & Lanamäki, A. (2015). “The sum of all human knowledge”: A systematic review of scholarly research on the content of Wikipedia. Journal of the Association for Information Science and Technology, 66(2), 219-245.
  3. reference: https://tools.wmflabs.org/siteviews/?platform=all-sites&source=unique-devices&start=2016-04&end=2017-03&sites=en.wikipedia.org
  4. a b Reagle, J., & Rhue, L. (2011). Gender bias in Wikipedia and Britannica. International Journal of Communication, 5, 21.
  5. a b c Lam, S. T. K., Uduwage, A., Dong, Z., Sen, S., Musicant, D. R., Terveen, L., & Riedl, J. (2011, October). WP: clubhouse?: an exploration of Wikipedia's gender imbalance. In Proceedings of the 7th international symposium on Wikis and open collaboration (pp. 1-10). ACM.
  6. a b Johnson, I. L., Lin, Y., Li, T. J. J., Hall, A., Halfaker, A., Schöning, J., & Hecht, B. (2016, May). Not at home on the range: Peer production and the urban/rural divide. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems (pp. 13-25). ACM.
  7. a b Hecht, B., & Gergle, D. (2009, June). Measuring self-focus bias in community-maintained knowledge repositories. In Proceedings of the fourth international conference on Communities and technologies (pp. 11-20). ACM.
  8. a b Graham, M., Hogan, B., Straumann, R. K., & Medhat, A. (2014). Uneven geographies of user-generated information: patterns of increasing informational poverty. Annals of the Association of American Geographers, 104(4), 746-764.
  9. a b Geiger, R. S., & Ribes, D. (2010, February). The work of sustaining order in wikipedia: the banning of a vandal. In Proceedings of the 2010 ACM conference on Computer supported cooperative work (pp. 117-126). ACM.
  10. a b Priedhorsky, R., Chen, J., Lam, S. T. K., Panciera, K., Terveen, L., & Riedl, J. (2007, November). Creating, destroying, and restoring value in Wikipedia. In Proceedings of the 2007 international ACM conference on Supporting group work (pp. 259-268). ACM.
  11. a b Stvilia, B., Twidale, M. B., Smith, L. C., & Gasser, L. (2008). Information quality work organization in Wikipedia. Journal of the Association for Information Science and Technology, 59(6), 983-1001.
  12. a b c d Kittur, A., & Kraut, R. E. (2008, November). Harnessing the wisdom of crowds in wikipedia: quality through coordination. In Proceedings of the 2008 ACM conference on Computer supported cooperative work (pp. 37-46). ACM.
  13. a b c d Arazy, O., & Nov, O. (2010, February). Determinants of wikipedia quality: the roles of global and local contribution inequality. In Proceedings of the 2010 ACM conference on Computer supported cooperative work (pp. 233-236). ACM.
  14. a b c Warncke-Wang, M., Cosley, D., & Riedl, J. (2013, August). Tell me more: an actionable quality model for Wikipedia. In Proceedings of the 9th International Symposium on Open Collaboration (p. 8). ACM.
  15. a b Stvilia, B., Twidale, M. B., Smith, L. C., & Gasser, L. (2005, November). Assessing Information Quality of a Community-Based Encyclopedia. In IQ.
  16. a b c d Warncke-Wang, M., Ayukaev, V. R., Hecht, B., & Terveen, L. G. (2015, February). The success and failure of quality improvement projects in peer production communities. In Proceedings of the 18th ACM Conference on Computer Supported Cooperative Work & Social Computing (pp. 743-756). ACM.
  17. a b Warncke-Wang, Morten (2015): English Wikipedia Quality Asssessment Dataset. figshare. https://doi.org/10.6084/m9.figshare.1375406.v1 Retrieved: 23 12, Apr 18, 2017 (GMT)
  18. a b Ross, Sage., Visualizing Article History with Structure Completeness. URL=https://wikiedu.org/blog/2016/09/16/visualizing-article-history-with-structural-completeness/ Retrieved: 2017-04-17.
  19. a b c Halfaker, Aaron; Sarabadani, Amir (2016): Monthly Wikipedia article quality predictions. figshare. https://doi.org/10.6084/m9.figshare.3859800.v3 Retrieved: 14 52, Apr 05, 2017 (GMT)
  20. a b Kittur, A., Chi, E. H., & Suh, B. (2009, April). What's in Wikipedia?: mapping topics and conflict using socially annotated category structure. In Proceedings of the SIGCHI conference on human factors in computing systems (pp. 1509-1512). ACM.
  21. a b Change, Rita. Emily Temple-Wood: A cool Wikipedian on a big mission. Wikimedia Blog. URL=https://blog.wikimedia.org/2013/10/11/emily-temple-wood-profile/ Retrieved: 2014-04-19
  22. a b Wikipedia 1.0 assessment table for Women Scientists URL=https://en.wikipedia.org/wiki/User:WP_1.0_bot/Tables/Project/Women_scientists Retrieved: 2017-04-19
  23. a b Wikipedia 1.0 assessment table for all of Wikipedia URL=https://en.wikipedia.org/wiki/Wikipedia:Version_1.0_Editorial_Team/Statistics Retrieved: 2017-04-19
  24. a b Emijrp. All Human Knowledge. URL=https://en.wikipedia.org/wiki/User:Emijrp/All_human_knowledge Retrieved: 2017-04-17.

Footnotes[edit]

  1. https://meta.wikimedia.org/wiki/Research:Revert
  2. https://en.wikipedia.org/wiki/Wikipedia:CVU
  3. https://en.wikipedia.org/wiki/Wikipedia:3RR
  4. https://enwp.org/Wikipedia:WikiProject_assessment
  5. https://en.wikipedia.org/wiki/Wikipedia:Version_1.0_Editorial_Team
  6. https://en.wikipedia.org/wiki/WikiProject
  7. https://en.wikipedia.org/wiki/Wikipedia:WikiProject_Medicine/Tools#Tasks
  8. https://en.wikipedia.org/wiki/Heckman_correction
  9. https://en.wikipedia.org/wiki/?oldid=96568092
  10. https://en.wikipedia.org/wiki/?oldid=769814730
  11. The Objective Revision Evaluation System is a machine prediction platform used by Wikipedians to detect vandalism, measure article quality, and automate other useful activities in Wikipedia. See https://meta.wikimedia.org/wiki/ORES
  12. https://github.com/wiki-ai/wikiclass/blob/950f693d789f8512e30f483f18e2d13483d13749/wikiclass/feature_lists/enwiki.py
  13. http://pythonhosted.org/revscoring/revscoring.scorer_models.html#revscoring.scorer_models.GradientBoosting
  14. See https://www.mediawiki.org/wiki/Help:Categories
  15. See https://www.mediawiki.org/wiki/Help:Templates
  16. See https://quarry.wmflabs.org/query/14033 for the query and its result.
  17. https://www.mediawiki.org/wiki/Wikimedia_Research/Showcase
  18. https://meta.wikimedia.org/wiki/Grants:IdeaLab/Inspire