Âé¶¹ÒùÔº

March 16, 2018

Analyzing why Wikipedia often overlooks stories of women in history

Credit: JJ Jordan from Pexels
× close
Credit: JJ Jordan from Pexels

Movements like #MeToo are drawing increased attention to the systemic discrimination facing women in a range of professional fields, from Hollywood and journalism to banking and government.

Discrimination is also a problem on user-driven sites like Wikipedia. Wikipedia is the . In January, the English-language version of the online encyclopedia had , more than 2000 percent higher than other online reference sites such as IMDb or Dictionary.com.

The volume of traffic on Wikipedia's site – coupled with its integration into and digital assistants like Alexa and Siri – makes Wikipedia the predominant source of information on the web. YouTube even recently announced that it would start including on highly contested topics. But studies show that Wikipedia underrepresents content on .

At the Rochester Institute of Technology, we're taking steps to empower our students and our global community to address issues of on Wikipedia.

Signs of bias

Driven by , Wikipedia's content can change in almost real time. That makes it a prime resource for current events, popular culture, sports and other evolving topics.

Get free science updates with Science X Daily and Weekly Newsletters — to customize your preferences!

But relying on volunteers leads to systemic biases – both in content creation and improvement. A 2013 study estimated that women only accounted for . Wikipedia co-founder Jimmy Wales believes much since then, despite several organized efforts.

If women don't actively edit Wikipedia at the same rate as men, topics of interest to women are at risk of receiving disproportionately low coverage. One study found that Wikipedia's coverage of women was , but entries on women still constituted less than 30 percent of biographical coverage. Entries on women also and are more likely to include information on romantic relationships and family roles.

What's more, Wikipedia's policies state that all content must be Since women throughout history have been less represented in published literature than men, it can be challenging to find reliable published sources on women.

An obituary in a paper of record is often a criterion for inclusion as a biographical entry in Wikipedia. So it should be no surprise that women are underrepresented as subjects in this vast . As the New York Times itself noted, its obituaries since 1851 – an oversight the paper now hopes to address through its "Overlooked" series.

Categorization can also be an issue. In 2013, a revealed that some editors had moved women's entries from gender-neutral categories (e.g., "American novelists") to gender-focused subcategories (e.g., "American women novelists").

Wikipedia is not the only online resource that suffers from such biases. The user-contributed online mapping service . On GitHub, an online development platform, women's contributions have a higher acceptance rate than men, but a study showed that the rate drops noticeably when the contributor .

Gender bias is also an ongoing issue in content development and search algorithms. Google Translate has been shown to and, for a time, LinkedIn in search results when users searched for a woman.

What can be done?

The solution to systemic biases that plague the web remains unclear. But libraries, museums, individual editors and the Wikimedia Foundation itself continue to make efforts to improve gender representation on sites such as Wikipedia.

Organized edit-a-thons can create a community around editing and developing underrepresented content. Edit-a-thons aim to increase the number of active female editors on Wikipedia, while empowering participants to edit entries on women during the event and into the future.

Later this month, our university library will host its second annual Women on Wikipedia Edit-a-thon in celebration of Women's History Month. The goal is to improve the content on at least 100 women in one afternoon.

For the past four years, students in our school's American Women's and Gender History course have worked to create new or substantially edit existing Wikipedia entries about women. One student created an entry on deaf-blind pioneer , while another added roughly 1,500 words to jazz artist entry.

This class was supported by , which encourages educators and students to contribute to Wikipedia in academic settings.

Through this assignment, students can immediately see how their efforts contribute to the larger conversation around women's history topics. One student said that it was .

Other efforts to address gender bias on Wikipedia include ; organized editing communities such as and ; and .

Wikipedia's dependence on volunteer editors has resulted in several systemic issues, but it also offers an opportunity for self-correction. Organized efforts help to give voice to women previously ignored by other resources.

Provided by The Conversation

Load comments (0)

This article has been reviewed according to Science X's and . have highlighted the following attributes while ensuring the content's credibility:

Get Instant Summarized Text (GIST)

This summary was automatically generated using LLM.