Countering Underproduction of Peer Produced Goods
Abstract
Peer produced goods such as online knowledge bases and free/libre open source software rely on contributors who often choose their tasks regardless of consumer needs. These goods are susceptible to underproduction: when popular goods are relatively low quality. Although underproduction is a common feature of peer production, very little is known about how to counteract it. We use a detailed longitudinal dataset from English Wikipedia to show that more experienced contributors—including those who contribute without an account—tend to contribute to underproduced goods. A within-person analysis shows that contributors’ efforts shift toward underproduced goods over time. These findings illustrate the value of retaining contributors in peer production, including those contributing without accounts, as a means to counter underproduction.
keywords:
Peer production; knowledge gaps; computer-supported cooperative work; anonymity; privacy; collaboration; underproduction; public goods1 Introduction
Peer production is a collaborative technology-assisted process in which participants select and perform tasks in a self-directed way, then combine their work with others (Benkler,, 2006). Peer production plays a critical role in today’s information ecosystem. From web servers to programming languages, Internet infrastructure is largely peer produced (Eghbal,, 2016). Much of the content we consume online is also peer produced. Peer produced websites such as Wikipedia, Reddit, and Fandom are among the most visited sites on the Internet.111https://perma.cc/K5VM-V99E Peer produced content includes map data, search results, digital assistants responses, AI training data, and more (McMahon et al.,, 2017).
How do producers’ choices of what to make correspond to what consumers want to use? In a market-driven system, supply and demand are aligned via price. But in commons-based peer production, no such signal is available. This can lead to underproduction: the production of high-interest but low-quality goods (Warncke-Wang et al.,, 2015).
This project seeks to understand the association between contributor experience, account use, and task selection. We first examine what is known about how tasks are selected to provide the rationale for several hypotheses. We then describe the setting, data, measures, and analytical plan. Next, we share empirical results. Finally, we discuss the limitations and implications of our results before concluding.
2 Background
2.1 Commons-based Peer Production and Underproduction
Commons-based peer production is a term coined by Yochai Benkler to describe an emerging form of cooperative production made possible by new communication technology. Tasks are self-selected and combined through technology, resulting in information public goods. Examples of peer production include free/libre open source software (FLOSS) projects like GNU/Linux and Apache, Wikipedia, and OpenStreetMap (OSM). Benkler, (2006) argues that because individuals know their interests, knowledge, availability, and skills, peer production creates the potential for diminished separation between producers and consumers and that self-assignment to tasks supports efficient matching between contributors and tasks.
One way to assess the success of a peer production project is to examine how well it meets the needs of its users. Warncke-Wang et al., (2015) found that over 40% of views to English Wikipedia were to articles that were lower quality than one might expect given their popularity. Countries, religions, LGBT topics, psychology, pop and rock music, the Internet and technology, comedy, and science fiction were found to be disproportionately affected. The misalignment of quality and public interest in these topics is concerning because of their relevance to public affairs and modern culture.
To quantify misalignment, Warncke-Wang et al., (2015) propose that if a system is aligned, goods in the highest demand should be the highest quality while those in the lowest demand should be the lowest quality. When quality is low relative to demand, goods are described as underproduced. When quality is high relative to demand, they are overproduced. Although overproduction may not be harmful beyond the potential for wasted effort, underproduction can impact public knowledge goods and digital infrastructure (Champion and Hill,, 2021; Eghbal,, 2016).
2.2 Motivation, Experience, and Task Selection
To understand the sources of underproduction, we must consider why people participate in peer production. Studies have observed a range of motivations. Some are motivated to create public goods (Budhathoki and Haythornthwaite,, 2013), to address social issues (March and Dasgupta,, 2020), to help other community members (Wu et al.,, 2007), to enhance their own use of a public good (Krishnamurthy et al.,, 2014; Meng and Wu,, 2013), to enhance their reputation or qualifications (Silva et al.,, 2020; Xu and Li,, 2015; Oreg and Nov,, 2008), as an experience of self-efficacy (Yang and Lai,, 2010), as a professional responsibility (Farič and Potts,, 2014), for personal enjoyment and learning (Farič and Potts,, 2014), out of reciprocity (Xu and Li,, 2015), for a class assignment (McDowell and Vetter,, 2022; Xing and Vetter,, 2020; Coelho et al.,, 2018; Konieczny,, 2012), or as part of their job (Germonprez et al.,, 2019). The results of this body of literature suggest that contributor motivation in peer production tends to be complex and multidimensional and primarily, although not exclusively, intrinsic (Benkler et al.,, 2015; Rafaeli and Ariel,, 2008; Belenzon and Schankerman,, 2015; Kuznetsov,, 2006).
To the extent that peer production contributors do so because they are told to (e.g., by an employer or instructor in a class), extrinsic rewards and others’ interests may be primary motivators. For example, contributing to Wikipedia is an increasingly common assignment in college classes where instructors are driven by concerns about content gaps or their commitment to public goods (McDowell and Vetter,, 2022; Xing and Vetter,, 2020; Konieczny,, 2012), or by a desire to have students engage with their coursework in public (Gallagher et al.,, 2019). Firms may encourage their employees to contribute to FLOSS to enhance the firm’s use of the public good or to build their reputation (Germonprez et al.,, 2019).
Nonprofit organizations also encourage contributions to peer production projects as part of their missions. Examples include the Missing Maps project by the Red Cross and Doctors Without Borders encouraging contributions to OSM (Herfort et al.,, 2023), partnerships encouraging contributions to FLOSS projects targeting climate change,222e.g., Climage Triage https://perma.cc/9R8F-QQYY and galleries, libraries and museums partnering with Wikipedia (Karczewska,, 2023). Some non-profits seek to counter content and participation gaps in peer production. For example, WikiEdu supports instructors using Wikipedia in their course assignments to close content gaps (Wilfahrt and Michelitch,, 2022). The Outreachy initiative aims to increase diversity in open source (Ackermann,, 2023).
Within this complex motivational landscape, contributor task selection varies across users and within users over time. In their examination of OSM contributors, Budhathoki and Haythornthwaite, (2013) reported that high-volume contributors were likelier to seek community recognition, while low-volume mappers reported wanting to contribute to a free and open project. In contrast, Shah, (2006) showed that FLOSS contributors describe their first contributions as directly related to personal skills, needs, and priorities, while longer-term participants report working for the good of the project. Likewise, Bryant et al., (2005) found that while Wikipedia editors reported participating within their areas of expertise initially, they sought to build the Wikipedia community and serve the public good over time. Restivo and van de Rijt, (2014) found that the highest-volume contributors produce more after receiving informal social rewards, while lower-volume contributors did not and may even respond negatively to the same awards.
Previous work offers some additional clues to the relationship between experience and choosing to contribute to underproduced goods. New contributors initially select tasks that match their immediate needs, areas of knowledge, and skills (Preece and Shneiderman,, 2009; Bryant et al.,, 2005). Because people are interested in similar things, it seems likely that these areas of knowledge will, on average, be associated with high-interest subjects. On the other hand, contributors faced with high-quality artifacts may not see where their assistance is needed (Preece and Shneiderman,, 2009; Bryant et al.,, 2005). However, as they accumulate experience, we expect contributors’ skills to increase.
Although this literature points in conflicting directions, prior work has emphasized that skilled participants incorporate the public’s desire for information as a component of their task selection. As a result, we expect to find that the most experienced contributors are more likely to engage in the improvement of underproduced goods: H1: individuals with less experience will contribute to less underproduced goods than individuals with more experience.
In that experienced contributors may seek recognition for their work, social factors may act to disrupt the trend we describe in H1 (Oreg and Nov,, 2008). Once a participant is involved in a project, their ongoing participation may be encouraged by a desire for status (Willer,, 2009). To distinguish the extent to which social rewards drive task selection, we take advantage of the fact that receiving in-group social rewards partially depends on identifiability. Collaborators may seek to direct their responses to an attributed person. Previous research suggests that creating an account may be driven by a desire to obtain feedback or recognition (Forte et al.,, 2017). By contrast, Belenzon and Schankerman, (2015) argued that anonymous contributors to FLOSS projects were more motivated by a desire to contribute to the public interests, as measured by the fact that they were more likely to contribute to projects operated by a nonprofit than a for-profit entity, and those serving end users (rather than other developers). Because contributing without an account makes social rewards more difficult, we propose that H2: contributing without an account is associated with more underproduced goods than contributing with an account.
Finally, we consider two competing explanations for H1 and H2. An association between task selection and experience could be explained by contributor attrition (i.e., those who go on to make many contributions have always differed from those who make only a few) or by shifting motivations (i.e., persistence causes changes in task selection). In other words, if we find support for H1 and H2, is it due to differences caused by users being “born” or “made”? Support for the “born” explanation comes from past work that has found that individuals who go on to contribute at high volume seem to do so from the start (e.g., Panciera et al.,, 2009). On the other hand, contributors’ motivation appears to shift over time (e.g., Shah,, 2006; Bryant et al.,, 2005). Moreover, peer production projects have successfully invested in efforts to socialize and retain newcomers (e.g., Tan et al.,, 2020; Morgan and Halfaker,, 2018), suggesting that contributors can also be “made”, at least to some extent. Given this evidence of motivational shifts, we hypothesize within-person change as H3A: an individual will shift toward underproduced goods as they accumulate experience, and H3B: an individual not using an account will shift toward underproduced goods at lower experience levels than one using an account. Our reasoning for H3B is that because non-account-users are less likely to receive social rewards, they are even more likely to be motivated by the public interest.
3 Research Design
3.1 Empirical Setting
This study is conducted in the context of Wikipedia, the Wikimedia Foundation (WMF) website. Wikipedia is one of the most popular websites in the world and is used as reference material, a fact-checking source, and an input to machine learning and AI systems (McMahon et al.,, 2017). Wikipedia projects in 326 different language editions received 29 billion page views in January 2024.333 WMF Statistics: https://perma.cc/H3NT-RP9N, Wikipedia page about Wikipedia: https://perma.cc/RNC3-4SLN The largest Wikipedia language edition, English, has more than 6.7 million articles. With some exceptions, clicking on an ‘Edit’ tab on every Wikipedia page creates an interface for revising the existing text. For most language editions, changes are immediately visible without review.
3.2 Data
Our unit of analysis is the revision, and we use the complete revision history of Wikipedia, made available by the WMF, from its inception in January 2001 through July 2021. The complete revision history of Wikipedia contains a wide range of content, including discussion and personal pages.
Our hypotheses concern the relationship between experience and having an account with article underproduction, a function of viewership and quality. To test our hypotheses, we filtered the revision population in multiple ways before drawing our sample. We excluded work done by bots, revisions to non-article pages, vandalism and its removal, and purely administrative revisions. Because no viewership data are available before December 2007, we only considered contributions after December 2007. Details on these filters and the total revisions after each step are reported in our online supplement.
In Wikipedia, a small fraction of contributors make a large portion of all contributions. To capture variation based on editor experience, we drew two samples. For the revision sample used to test H1 and H2, we first stratified contributions by their ordinal position into increasingly large “buckets” increasing in size by (i.e., 1, 2, 4, 8, and so on). We then drew equal-sized random samples from each bucket, oversampling where the data was thinner, for a total of 192,672 revisions. The inverse of the resulting sampling proportions was used as weights during subsequent calculations (Tracy and Carkin,, 2014).
The second sample is a within-person sample, designed to extract the contribution history of editors who differed in their ultimate contribution level. To build this sample, we stratified editors based on their total contributions through July 2021, again by creating increasingly large () buckets. We then drew a random sample of editors from each. Again, we oversampled where data was thinner and retained proportions as weights. After selecting editors, we gathered all article contributions made by the selected editors for a total of 42,602,912 revisions. This approach gives us the contribution history of a diverse range of editors: those who went on to contribute a great deal and those who only contributed once or twice.
We have made our data and code available at https://doi.org/10.7910/DVN/UDQT6E.
3.3 Measures
Our dependent variable is underproduction factor. To construct this measure, we drew from elements of both Champion and Hill, (2021) and Warncke-Wang et al., (2015) and calculated the negative log of the ratio between quality rank and popularity rank. Both ranks were assigned such that high quality and high popularity receive high ranks, while low quality and low popularity receive low ones. A perfectly aligned article—one with the same popularity and quality rank—would have an underproduction factor of 0 since log(1) is 0. An overproduced article of high quality and low popularity would have a negative underproduction factor. An underproduced article of low quality and high popularity would have a positive underproduction factor.
We count views at the monthly level using data released by the WMF. We calculate quality at the monthly level using the ORES quality measure. ORES is a machine learning-based quality measure that reflects the structural characteristics of articles, such as the presence of references, section headings, and inter-wiki links (Halfaker et al.,, 2016). ORES is a continuous measure with separate predictions for six article quality levels. We treat these levels as equally spaced to create a single quality score between 0 and 5. More details are provided in the supplement.
We operationalize editor experience as revision count: the number of revisions the editor has made before and including the current revision. Therefore, before conducting the filtering described in the data and sample section, we calculated each revision’s ordinal position in its contributor’s edit history, ignoring any reverted contributions. To consider the influence of the potential for receiving social rewards, we also introduced a variable reflecting whether contributions were made by users who were not using accounts. Because these users are identified by IP address in Wikipedia, we call this measure IP-based.
3.4 Analytical Plan
We employ a hierarchical multiple regression model to test our hypotheses H1 and H2. We use revision count (logged), revision count (logged) squared, and IP-based as predictors and a random intercept term for users using the lmer package in R and a ‘raw’ specification for our polynomials.
Although our hierarchical models attempt to correct for repeated measures of users, this approach does not account for the fact that experienced users might edit different articles than less experienced users in ways that reflect underlying differences in the types of users who go on to become less or more experienced (Panciera et al.,, 2009). To answer H3A and H3B, we fit a second group of models with user-level fixed effects using the felm package in R and our within-person sample. Because we have no within-person variation in account use, we cannot include IP-based in the within-person model. Due to heteroskedasticity, we use robust standard errors in all our models.
3.5 Ethics
This study was conducted entirely using publicly available data published by WMF and does not involve any interaction or intervention with human subjects. The IRB at our institution has reviewed this type of research using these data and determined it is not human-subject research. WMF fully anonymized article view data before release.
4 Results
The results of our polynomial model evaluating the relationship between underproduction and experience (H1) and between underproduction and having an account (H2) are shown in Table 1 and visualized in a marginal effects plot in Figure 1.

Linear Model | Polynomial Model | |
---|---|---|
Intercept | ||
Revision Count (ln) | ||
Revision Count (ln2) | ||
Editor was IP-based | ||
Revision Count (ln) and IP-based | ||
Revision Count (ln2) and IP-based | ||
AIC | ||
BIC | ||
Log Likelihood | ||
Num. obs. | ||
Num. groups: editor_id_or_ip | ||
Var: editor_id_or_ip (Intercept) | ||
Var: Residual |
In the linear model for H1, we find that an increase in one log unit of experience is associated with a 0.0287 increase in the underproduction factor of the article selected for contribution (see Table 1). The quadratic specification of our model is shown alongside the linear specification and is a substantially better fit for the data (; ; ; ). The parameter estimates for all terms in both models are statistically significant, except for the interaction of revision count and IP-based. The polynomial model describes U-shaped curves with an inflection point at about 150 revisions for those contributing with an account (see Figure 1). These results provide some support for H1: contributions by low-experience individuals are to less underproduced articles than those from users with high experience levels. However, as we see from the polynomial model, the relationship is U-shaped. We also note that the highly skewed distribution of contributions is such that most contributors do not make more than a small number of contributions.

For H2, our linear model shows that contributing using an IP address is associated with a 0.0609 unit increase in the underproduction factor of the selected article. An increase in one log unit of experience is associated with a difference of -0.0198 in the same outcome. The quadratic specification of our model shows the opposite: the parameter estimate of the main effect associated with using an IP address is a -0.1511 difference in underproduction factor, with a positive interaction term (revision count and IP-based) of 0.0376. The squared interaction term (revision count squared and IP-based) is both small in magnitude and not statistically significant.
To help understand these results, we construct Figure 2, which shows a nonparametric GAM smoothed line fit to a 10% sample of the same data used to fit the models alongside the predicted values from both the linear and polynomial models. While the quadratic model is a better fit for the data, the linear model is more consistent with the behavior of IP-based contributors as shown in our nonparametric data visualization. If we use the linear model as the standard for evaluating our hypothesis about IP-based editors, we find that although they shift towards underproduced materials as they accumulate experience, they do not do so at greater levels than those contributing with an account. Overall, these results contradict our proposal in H2.
Within-Person: With Account IP-based With Account, Quadratic IP-based, Quadratic Revision Count (ln) Revision Count (ln2) Num. obs. R2 (full model) R2 (proj model) Num. groups: editor_id_or_ip
H3A and H3B relate to within-person change. The results of our model testing both parts of H3 are presented in Table 2. In H3A, we proposed a within-person shift toward underproduced goods as contributors increase in experience. All models support H3A. We observe a positive coefficient in the linear specification, suggesting an upward trend toward increasingly underproduced articles as individuals accumulate experience. In the quadratic specification, we observe a positive coefficient associated with the second-order term, suggesting a U-shaped relationship such that at higher revision counts, contributors select underproduced articles. We see mixed results for H3B related to those contributing without an account. When fit with a linear specification, the magnitude of the effect of experience is smaller for those contributing without an account (0.0219) than the coefficient associated with contributing with an account (0.0808). However, in the quadratic model, the coefficient in the squared term is slightly larger for those contributing without an account (0.0087) than those contributing with an account (0.0077).

To understand the implications of these models, we simulated task selection across experience levels, with the individual-level fixed effect set at the model estimated median. Figure 3 shows the model-predicted marginal effect of accumulating editing experience on task selection. The figure shows that contributors tend towards underproduced articles as they accumulate experience.
Overall, our results from these models provide evidence that contributors’ task selection shifts toward underproduced articles as their experience increases. This trend is present in people who contribute with and without accounts. However, some models point in the opposite direction of our hypothesis in H3B that those who contribute without an account will shift sooner than those who use an account. We offer additional interpretation of these results in our supplement.
5 Limitations
Although we measure the revision count of IP-based contributors, IP addresses change in ways that mean we cannot consistently associate individuals with IP addresses. Furthermore, some IP-based revisions are likely to be authored by experienced Wikipedians. As a result, our model may understate the influence of editing with an IP at low experience or overstate it at high experience.
Using revision count to operationalize experience introduces important limitations. For example, our measure does not distinguish between a revision that fixes a typo and one that adds a paragraph. Similarly, it does not differentiate between long-lived text and quickly deleted text. We use revision count because it is the measure of editor experience most widely used by both Wikipedians and Wikipedia researchers and because the most discussed alternative measures based on content persistence are both computationally complex and require a range of difficult decisions. Other research has shown that revision count is highly correlated with, and frequently leads to similar results as, measures that account for content persistence (e.g., Hill and Shaw,, 2021). Ultimately, assessing how contribution types vary over the lifespan of editors remains a subject for future research.
Additionally, our design does not address causal factors among quality, popularity, and task selection. Indeed, we use measures that cannot be interpreted causally due to their granularity: our measures of quality and popularity are taken at the monthly level, not at the moment of editing.
Another threat was described by the Wikipedia editing community when Warncke-Wang et al.,’s (2015) work on misalignment was first published. Community members stated that one explanation for underproduction is that some articles are more difficult to write than others and that this difficulty may be systematically distributed in a way that coincides with how people search for and consume articles.444Wikipedia community-run internal newsletter, The Signpost: https://perma.cc/GX5U-32R9 There may be factors (such as conceptual generality) driving both high popularity and low quality. In part, using ORES as a quality measure mitigates this risk because ORES measures quality based on structural characteristics independent of conceptual qualities like completeness. However, this approach does not address the fact that, as a tertiary source, Wikipedia is limited by the availability of reliable published information, including the biases in the topics these sources cover. Writing a high-quality article about a high-interest, under-reported topic may be more difficult than writing about widely documented topics.
6 Discussion
6.1 Born, Made, or Something Else?
This article offers insight into community-wide trends and within-contributor changes in task selection behaviors. Our results in H1 and H3A suggest that as they increase in experience, contributors tend to shift their attention to underproduced content. This is consistent with the “made” argument that accumulating experience is associated with editing underproduced articles and, presumably, with increased exposure to socialization. Our results in H2 and H3B call this interpretation into question. Socialization suggests a social process, but contributors who participate without an account have less opportunity to build a social identity in Wikipedia. How does this occur if these contributors also shift towards underproduced articles?
Our models suggest that IP-based contributors are initially less inclined to select underproduced articles but are comparable to account-based contributors for moderate contribution levels (i.e., in the 100-200 range). This finding opens up a new set of challenges for social research. Given that contributors without accounts also move toward underproduced goods—and hence may represent a valuable subgroup to seek to influence and retain—how should we explain their persistence and changing task selection?
One way to understand this shift in the behavior of IP-based contributors is through an extension of learning theory—e.g., Preece and Shneiderman,’s (2009) reader-to-leader perspective and Lave and Wenger,’s (1991) legitimate peripheral participation. As participants gain experience, perhaps their awareness of underproduced articles increases. Given that this process is mediated through technology in rich social computing environments, this may be evidence of what we call “technosocial learning,” rather than social learning per se.
We suggest that technosocial learning may be an important mechanism to shape behavior in peer production environments. For example, consider the general phenomenon of feedback on a contribution. Someone contributing without an account might not receive such feedback. However, this contributor could notice whether their past contribution remains present.
Platforms offer other opportunities to learn from technical traces: contributors might experience success in mastering a particular formatting trick, observing the work of others, reading documentation, or seeing feedback others receive. A technosocial learning process, illustrated in Table 3, may provide a mechanism for the changes in task selection we observe.
implicit | explicit | |
---|---|---|
social | observing the interests of others | personal feedback |
technical | noticing a good is low quality | automated feedback |
6.2 How may underproduction be countered?
Given the previous finding of widespread underproduction in peer production (Warncke-Wang et al.,, 2015; Champion and Hill,, 2021), countering it is an important challenge. One approach to addressing underproduction is to increase retention overall. Another approach is nudging contributors towards more severe areas of underproduction earlier: both when they are newer to platform, and more quickly in response to emerging needs. For example, peer production communities responded during the COVID-19 pandemic by developing FLOSS analytic and visualization tools (Wang et al.,, 2021), organizing to produce Wikipedia articles to provide reliable information at high speed (Avieson,, 2022), and producing personal protective equipment and filling supply chain gaps (Sarkar et al.,, 2023). Evaluation of recommender systems suggests they can influence contributors to improve underrepresented topics, providing the recommendation’s relevance is kept constant (Houtti et al.,, 2023). Creating interest groups within a larger project may also be an effective way to tackle underproduction. In Wikipedia, contributors have self-organized around themes of interest to identify and work to close a wide range of participation and content gaps—e.g., WikiProject Women in Red (Tripodi,, 2021), WikiProject Women Scientists (Halfaker,, 2017), and WikiProject Vital Articles (Houtti et al.,, 2022).
However, more work is needed to understand when these interventions are effective. Ford et al., (2018) found that notification of existing contributors, writ large, was insufficient to address content gaps, instead recommending recruiting people with expertise related to the gap. A computational linguistics analysis from Schmahl et al., (2020) found that gendered bias in articles changed modestly, at best, between 2006 and 2020 when many targetted interventions were conducted. Successful interventions in task selection include not only involving diverse participants but also participants with diverse motives (e.g., through course assignments and non-profit outreach activities (as per Konieczny,, 2012; Coelho et al.,, 2018; Gallagher et al.,, 2019; Xing and Vetter,, 2020; Herfort et al.,, 2023; Karczewska,, 2023)).
6.3 Problematizing retention and account creation
Our analysis points to promoting contributor retention as a means of countering underproduction. Scholarship in online community and peer production community management has also emphasized retention as a path toward building healthy online communities (e.g., Kraut et al.,, 2012). However, our findings should be read alongside the evidence against requiring account creation and the risks associated with overemphasizing the retention of existing users.
Requiring accounts has been found to carry unintended consequences in terms of diminishing overall community activity levels (e.g., Hill and Shaw,, 2021). Being prompted to create an account increases the barrier to contributing. Maintaining a low barrier to entry may be essential to the success of peer production projects. Beyond this risk, Shaw and Hill, (2014) found that governance of peer production tends to concentrate power in the hands of elite long-standing contributors, emphasizing that longevity risks further empowering elites. Additional analysis of long-term contributors appears in our supplement.
Increased retention may also magnify systemic biases reflected in peer production. For example, previous work has explored geographical biases in OSM (Thebault-Spieker et al.,, 2018), gender inequality in FLOSS participation (Qiu et al.,, 2023), gender bias in how FLOSS projects engage with contributors (Chatterjee et al.,, 2021), gender inequality in Wikipedia contributors (Hill and Shaw,, 2013) and gender bias in Wikipedia content (Adams and Brückner,, 2015). Therefore, while retaining contributors is valuable, it may reinforce inequality in the population of early contributors. Scholarship has shown that communities can make progress in closing participation gaps with interventions that address bias in sociotechnical processes through which people are retained unequally (Karczewska,, 2023; Langrock and González-Bailón,, 2022; Hilderbrand et al.,, 2020; Evans et al.,, 2015).
Along with recruiting and retaining diverse contributors, we believe there is also value in retaining those who contribute without accounts. A proposed update to Wikipedia’s software slated for 2024 describes a change to the way that users without accounts are identified by replacing visible IP addresses with temporary identifiers.555See implementation proposal: https://perma.cc/3H59-YLH8 Although details of this proposal are still in flux, this change will improve the privacy of those contributing without an account by hiding IP addresses that may reveal geographic location. Although the proposed alternative will improve the ability of researchers to track contributions from an individual across locations, it will also make it impossible to do studies of long-term engagement of users contributing without accounts—like ours. Evidence-based management of peer production communities faces multiple critical challenges: countering underproduction, enabling diverse participation, expanding community safety, building long-term commitment, and minimizing power concentration. Some of these goals may force difficult tradeoffs, as interventions intended to address one area may make others more difficult to address.
7 Conclusion
Although previous research suggests that the underproduction of peer produced goods is widespread, understanding how underproduction happens and how it can be addressed is still emerging. Our work advances a theory that experience contributes to changes in production. We test this theory and find some support: experienced contributors tend to select underproduced goods, whether they contribute with or without an account. However, contrary to our expectations, contributors not using an account are not different than those contributing using an account. In addition to the explanations of social rewards explored in previous work, we suggest that task selection changes in contributors without an account may be explained by our proposed notion of ‘technosocial learning.’ Our modern communication environment relies on peer production for both content and infrastructure. Understanding how to better counter underproduction is a necessary part of protecting and continuing to expand the benefit of public goods developed through peer production.
References
- Ackermann, (2023) Ackermann, R. (2023). The future of open source is still very much in flux. MIT Technology Review.
- Adams and Brückner, (2015) Adams, J. and Brückner, H. (2015). Wikipedia, sociology, and the promise and pitfalls of Big Data. Big Data & Society, 2(2):2053951715614332.
- Avieson, (2022) Avieson, B. (2022). Editors, sources and the ‘go back’ button: Wikipedia’s framework for beating misinformation. First Monday.
- Belenzon and Schankerman, (2015) Belenzon, S. and Schankerman, M. (2015). Motivation and sorting of human capital in open innovation. Strategic Management Journal, 36(6):795–820.
- Benkler, (2006) Benkler, Y. (2006). The Wealth of Networks: How Social Production Transforms Markets and Freedom. Yale University Press, New Haven, CT.
- Benkler et al., (2015) Benkler, Y., Shaw, A., and Hill, B. M. (2015). Peer production: A form of collective intelligence. In Malone, T. W. and Bernstein, M. S., editors, Handbook of Collective Intelligence, pages 175–204. MIT Press, Cambridge, MA.
- Bryant et al., (2005) Bryant, S. L., Forte, A., and Bruckman, A. (2005). Becoming Wikipedian: transformation of participation in a collaborative online encyclopedia. In Proceedings of the 2005 International ACM SIGGROUP Conference on Supporting Group Work, pages 1–10, New York, NY. ACM.
- Budhathoki and Haythornthwaite, (2013) Budhathoki, N. R. and Haythornthwaite, C. (2013). Motivation for open collaboration: Crowd and community models and the case of OpenStreetMap. American Behavioral Scientist, 57(5):548–575.
- Champion and Hill, (2021) Champion, K. and Hill, B. M. (2021). Underproduction: An approach for measuring risk in open source software. In 2021 IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER), pages 388–399, Honolulu, HI, USA. IEEE.
- Chatterjee et al., (2021) Chatterjee, A., Guizani, M., Stevens, C., Emard, J., May, M. E., Burnett, M., and Ahmed, I. (2021). AID: An automated detector for gender-inclusivity bugs in OSS project pages. In 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE), pages 1423–1435, Madrid, ES. IEEE.
- Coelho et al., (2018) Coelho, J., Valente, M. T., Silva, L. L., and Hora, A. (2018). Why we engage in FLOSS: Answers from core developers. In Proceedings of the 11th International Workshop on Cooperative and Human Aspects of Software Engineering, Gothenburg, Sweden. ACM.
- Eghbal, (2016) Eghbal, N. (2016). Roads and Bridges: The Unseen Labor Behind Our Digital Infrastructure. Ford Foundation.
- Evans et al., (2015) Evans, S., Mabey, J., and Mandiberg, M. (2015). Editing for equality: The outcomes of the Art+Feminism Wikipedia Edit-a-thons. Art Documentation: Journal of the Art Libraries Society of North America, 34(2):194–203.
- Farič and Potts, (2014) Farič, N. and Potts, H. W. (2014). Motivations for contributing to health-related articles on Wikipedia: An interview study. Journal of Medical Internet Research, 16(12):e260.
- Ford et al., (2018) Ford, H., Pensa, I., Devouard, F., Pucciarelli, M., and Botturi, L. (2018). Beyond notification: Filling gaps in peer production projects. New Media & Society, 20(10):3799–3817.
- Forte et al., (2017) Forte, A., Andalibi, N., and Greenstadt, R. (2017). Privacy, anonymity, and perceived risk in open collaboration: a study of Tor users and Wikipedians. In Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing, pages 1800–1811, New York, NY. ACM.
- Gallagher et al., (2019) Gallagher, K., Fioravanti, M., and Kozaitis, S. (2019). Teaching software maintenance. In 2019 IEEE International Conference on Software Maintenance and Evolution (ICSME), pages 353–362, Cleveland, OH, USA. IEEE.
- Germonprez et al., (2019) Germonprez, M., Lipps, J., and Goggins, S. (2019). The rising tide: Open source’s steady transformation. First Monday, 24(8).
- Halfaker, (2017) Halfaker, A. (2017). Interpolating quality dynamics in Wikipedia and demonstrating the Keilana effect. In Proceedings of the 13th International Symposium on Open Collaboration, pages 1–9. ACM Press.
- Halfaker et al., (2016) Halfaker, A., Morgan, J., Sarabadani, A., and Wight, A. (2016). ORES: Facilitating re-mediation of Wikipedia’s socio-technical problems. Working Paper, Wikimedia Research.
- Herfort et al., (2023) Herfort, B., Lautenbach, S., Porto de Albuquerque, J., Anderson, J., and Zipf, A. (2023). A spatio-temporal analysis investigating completeness and inequalities of global urban building data in OpenStreetMap. Nature Communications, 14(1):3985.
- Hilderbrand et al., (2020) Hilderbrand, C., Perdriau, C., Letaw, L., Emard, J., Steine-Hanson, Z., Burnett, M., and Sarma, A. (2020). Engineering gender-inclusivity into software: ten teams’ tales from the trenches. In Proceedings of the ACM/IEEE 42nd International Conference on Software Engineering, pages 433–444, Seoul South Korea. ACM.
- Hill and Shaw, (2013) Hill, B. M. and Shaw, A. (2013). The Wikipedia gender gap revisited: characterizing survey response bias with propensity score estimation. PLoS ONE, 8(6):e65782.
- Hill and Shaw, (2021) Hill, B. M. and Shaw, A. (2021). The hidden costs of requiring accounts: Quasi-experimental evidence from peer production. Communication Research, 48(6):771–795.
- Houtti et al., (2022) Houtti, M., Johnson, I., Cepeda, J., Khandelwal, S., Bhatnagar, A., and Terveen, L. (2022). "We need a woman in music": Exploring Wikipedia’s values on article priority. Proceedings of the ACM on Human-Computer Interaction, 6(CSCW2):1–28.
- Houtti et al., (2023) Houtti, M., Johnson, I., and Terveen, L. (2023). Leveraging recommender systems to reduce content gaps on peer production platforms. arXiv:2307.08669 [cs].
- Karczewska, (2023) Karczewska, A. (2023). GLAM WikiProjects as a form of organization of cooperation between Wikipedia and cultural institutions. European Conference on Knowledge Management, 24(1):619–627.
- Konieczny, (2012) Konieczny, P. (2012). Wikis and Wikipedia as a teaching tool: Five years later. First Monday.
- Kraut et al., (2012) Kraut, R. E., Resnick, P., and Kiesler, S. (2012). Building Successful Online Communities: Evidence-based Social Design. MIT Press, Cambridge, MA.
- Krishnamurthy et al., (2014) Krishnamurthy, S., Ou, S., and Tripathi, A. K. (2014). Acceptance of monetary rewards in open source software development. Research Policy, 43(4):632–644.
- Kuznetsov, (2006) Kuznetsov, S. (2006). Motivations of contributors to Wikipedia. ACM SIGCAS Computers and Society, 36.
- Langrock and González-Bailón, (2022) Langrock, I. and González-Bailón, S. (2022). The gender divide in Wikipedia: Quantifying and assessing the impact of two feminist interventions. Journal of Communication.
- Lave and Wenger, (1991) Lave, J. and Wenger, E. (1991). Situated Learning: Legitimate Peripheral Participation. Cambridge University Press, Cambridge, UK.
- March and Dasgupta, (2020) March, L. and Dasgupta, S. (2020). Wikipedia edit-a-thons as sites of public pedagogy. Proceedings of the ACM on Human-Computer Interaction, 100:1–100:26(CSCW2). In press.
- McDowell and Vetter, (2022) McDowell, Z. J. and Vetter, M. A. (2022). Wikipedia as open educational practice: Experiential learning, critical information literacy, and social justice. Social Media + Society, 8(1).
- McMahon et al., (2017) McMahon, C., Johnson, I. L., and Hecht, B. J. (2017). The substantial interdependence of Wikipedia and Google: A case study on the relationship between peer production communities and information technologies. In International AAAI Conference on Web and Social Media (ICWSM 2017), pages 142–151, Palo Alto, California. AAAI.
- Meng and Wu, (2013) Meng, B. and Wu, F. (2013). COMMONS/COMMODITY: Peer production caught in the Web of the commercial market. Information, Communication & Society, 16(1):125–145.
- Morgan and Halfaker, (2018) Morgan, J. T. and Halfaker, A. (2018). Evaluating the impact of the Wikipedia Teahouse on newcomer socialization and retention. In Proceedings of the 14th International Symposium on Open Collaboration, OpenSym ’18, pages 20:1–20:7, New York, NY. ACM.
- Oreg and Nov, (2008) Oreg, S. and Nov, O. (2008). Exploring motivations for contributing to open source initiatives: The roles of contribution context and personal values. Computers in Human Behavior, 24(5):2055–2073.
- Panciera et al., (2009) Panciera, K., Halfaker, A., and Terveen, L. (2009). Wikipedians are born, not made: a study of power editors on Wikipedia. In Proceedings of the ACM 2009 International conference on Supporting Group Work, pages 51–60, New York, NY. ACM.
- Preece and Shneiderman, (2009) Preece, J. and Shneiderman, B. (2009). The reader-to-leader framework: motivating technology-mediated social participation. AIS Transactions on Human-Computer Interaction, 1(1):13–32.
- Qiu et al., (2023) Qiu, H. S., Zhao, Z. H., Yu, T. K., Wang, J., Ma, A., Fang, H., Dabbish, L., and Vasilescu, B. (2023). Gender representation among contributors to open-source infrastructure : An analysis of 20 package manager ecosystems. In 2023 IEEE/ACM 45th International Conference on Software Engineering: Software Engineering in Society (ICSE-SEIS), pages 180–187, Melbourne, Australia. IEEE.
- Rafaeli and Ariel, (2008) Rafaeli, S. and Ariel, Y. (2008). Online motivational factors: incentives for participation and contribution in Wikipedia. In Barak, A., editor, Psychological Aspects of Cyberspace: Theory, Research, Applications. Cambridge University Press.
- Restivo and van de Rijt, (2014) Restivo, M. and van de Rijt, A. (2014). No praise without effort: experimental evidence on how rewards affect Wikipedia’s contributor community. Information, Communication & Society, 17(4):1–12.
- Sarkar et al., (2023) Sarkar, S., Waldman-Brown, A., and Clegg, S. (2023). A digital ecosystem as an institutional field: curated peer production as a response to institutional voids revealed by COVID -19. R&D Management, 53(4):695–708.
- Schmahl et al., (2020) Schmahl, K. G., Viering, T. J., Makrodimitris, S., Naseri Jahfari, A., Tax, D., and Loog, M. (2020). Is Wikipedia succeeding in reducing gender bias? Assessing changes in gender bias in Wikipedia using word embeddings. In Proceedings of the Fourth Workshop on Natural Language Processing and Computational Social Science, pages 94–103, Online. Association for Computational Linguistics.
- Shah, (2006) Shah, S. K. (2006). Motivation, governance, and the viability of hybrid forms in open source software development. Management Science, 52(7):1000–1014.
- Shaw and Hill, (2014) Shaw, A. and Hill, B. M. (2014). Laboratories of oligarchy? How the iron law extends to peer production. Journal of Communication, 64(2):215–238.
- Silva et al., (2020) Silva, J. O., Wiese, I., German, D. M., Treude, C., Gerosa, M. A., and Steinmacher, I. (2020). Google summer of code: Student motivations and contributions. Journal of Systems and Software, 162:110487.
- Tan et al., (2020) Tan, X., Zhou, M., and Sun, Z. (2020). A first look at good first issues on GitHub. In Proceedings of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, pages 398–409, Virtual Event USA. ACM.
- Thebault-Spieker et al., (2018) Thebault-Spieker, J., Hecht, B., and Terveen, L. (2018). Geographic biases are ‘born, not made’: Exploring contributors’ spatiotemporal behavior in OpenStreetMap. In Proceedings of the 2018 ACM Conference on Supporting Groupwork, pages 71–82. ACM Press.
- Tracy and Carkin, (2014) Tracy, P. E. and Carkin, D. M. (2014). Adjusting for design effects in disproportionate stratified sampling designs through weighting. Crime & Delinquency, 60(2):306–325.
- Tripodi, (2021) Tripodi, F. (2021). Ms. Categorized: Gender, notability, and inequality on Wikipedia. New Media & Society.
- Wang et al., (2021) Wang, L., Li, R., Zhu, J., Bai, G., and Wang, H. (2021). A large-scale empirical study of COVID-19 themed GitHub repositories. In 2021 IEEE 45th Annual Computers, Software, and Applications Conference (COMPSAC), pages 914–923, Madrid, Spain. IEEE.
- Warncke-Wang et al., (2015) Warncke-Wang, M., Ranjan, V., Terveen, L., and Hecht, B. (2015). Misalignment between supply and demand of quality content in peer production communities. In Proceedings of the Ninth International AAAI Conference on Web and Social Media, pages 493–502.
- Wilfahrt and Michelitch, (2022) Wilfahrt, M. and Michelitch, K. (2022). Improving open-source information on African politics, one student at a time. PS: Political Science & Politics, 55(2):450–455. Publisher: Cambridge University Press.
- Willer, (2009) Willer, R. (2009). A status theory of collective action. In Advances in Group Processes, volume 26, pages 133–163.
- Wu et al., (2007) Wu, C.-G., Gerlach, J. H., and Young, C. E. (2007). An empirical analysis of open source software developers’ motivations and continuance intentions. Information & Management, 44(3):253–262.
- Xing and Vetter, (2020) Xing, J. and Vetter, M. (2020). Editing for equity: Understanding instructor motivations for integrating cross-disciplinary Wikipedia assignments. First Monday.
- Xu and Li, (2015) Xu, B. and Li, D. (2015). An empirical study of the motivations for content contribution and community participation in Wikipedia. Information & management, 52(3):275–286.
- Yang and Lai, (2010) Yang, H.-L. and Lai, C.-Y. (2010). Motivations of Wikipedia content contributors. Computers in Human Behavior, 26(6):1377–1383.
Acknowledgements
Morten Warncke-Wang of the WMF provided both valuable advice on multiple occasions and access to his dataset. Aaron Halfaker provided advice in the effective use of the ORES machine learning system for quality analysis. A pilot version of this project was completed as part of the Advanced Regression course taught by Jeffrey Arnold in the Center for Statistics in the Social Sciences at University of Washington. Members of the Community Data Science Collective provided advice on early drafts of this project, including Jeremy Foote, Floor Fiers, Wm Salt Hale, Sohyeon Hwang, Sejal Khatri, Charles Kiene, Sneha Narayan, Aaron Shaw, and Nathan TeBlunthuis. The creation of our dataset was aided by the use of advanced computational, storage, and networking infrastructure provided by the Hyak supercomputer system at the University of Washington. The authors gratefully acknowledge support from the National Science Foundation, awards CNS-1703736 and CNS-1703049, and the Sloan Foundation through the Ford/Sloan Digital Infrastructure Initiative, Sloan Award 2018-11356.