Designing ethically with AI: How Wikimedia can harness machine learning in a responsible and human-centered way – Wikimedia Foundation
Skip to content
Designing ethically with AI: How Wikimedia can harness machine learning in a responsible and human-centered way
By
Jonathan T. Morgan
18 July 2019
Bluesky
Email
Link copied
The past few years have seen an explosion of journalism, scholarship, and advocacy around the topic of ethical AI. This attention reflects a growing recognition that technology companies often fail to put the needs of the people who use
machine learning
(or “AI”) technology, and of society as a whole, ahead of their business goals.
Much of the public conversation on the topic of ethical AI has revolved around
general
principles
like
fairness
transparency
, and
accountability
. Articulating the principles that underlie ethical AI is an important step. But technology companies also need practical guidance on how to apply those principles when they develop products based on AI, so that they can identify major risks and make informed decisions.
What would a
minimum viable process
(MVP) for ethical AI product development look like at Wikimedia, given our strengths, weaknesses, mission, and values? How do we use AI to support knowledge equity, ensure the knowledge integrity, and help our movement thrive without undermining our values?
Towards a MVP for ethical AI
The Wikimedia Foundation’s Research team has begun to tackle these questions in a new white paper.
Ethical & Human centered AI at Wikimedia
takes the
2030 strategic direction
as a starting point, building from
the observation
that
“Developing and harnessing technology in socially equitable and constructive ways—and preventing unintended negative consequences—requires thoughtful leadership and technical vigilance.”
The white paper was developed through an extensive literature review and consultation with subject matter experts, and builds off of other
recent work
by the Foundation’s Research and Audiences teams.
The white paper has two main components. First, it presents a set of
risk scenarios
—short vignettes that describe the release of a hypothetical AI-powered product, and some plausible consequences of that release on Wikimedia’s content, contributors, or readers. Second, it proposes a set of improvements we can make to the
process
we follow when we develop AI-powered products, and to the
design
of the products themselves, that will help us avoid the negative consequences described in the scenarios.
Could algorithmically-generated section recommendations inadvertently increase gender bias in biographies of women? (risk scenario A: Reinforcing existing bias.) Text from the English Wikipedia article about
"Alice Frey"
CC BY-SA 3.0
Image
by unknown, used in the article under fair use. Mockup by Wikimedia Foundation,
CC BY-SA 3.0
Identifying and addressing risks
The
risk scenarios
are intended to spur discussion among AI product stakeholders—product teams, research scientists, organizational decision-makers, and volunteer communities. Scenarios like these can be used in discussions around product planning, development, and evaluation to raise important questions. They can help us uncover assumptions that might otherwise be left unstated, and highlight tensions between immediate goals and foundational values. The goal is to help people grapple with these trade-offs and identify alternative approaches that minimize the risk of unintended consequences.
Each of the six risk scenarios address a complex ethical issue that AI products can make worse—like the risk of reinforcing systemic bias, discouraging diversity, and creating inequity in access to information. They also help uncover subtler issues—like the risk of disrupting community workflows or subverting editorial judgement when we automate processes that are currently performed by people. While the negative outcomes described in the risk scenarios are hypothetical, each one is based on a realistic Wikimedia-specific AI product use case.
The ORES quality prediction algorithm provides detailed information on how it decides what quality category an article belongs in, increasing transparency and accountability (Proposal 5: Build interpretable models.) Text from the English Wikipedia article about
Aaron Halfaker
CC BY-SA 3.0
Image
by Myleen Hollero/Wikimedia Foundation,
CC BY-SA 3.0
. Screenshot by Wikimedia Foundation,
CC BY-SA 3.0
The eight
process improvement proposals
described in the white paper lay out courses of action that Wikimedia can take when developing AI products. Following these recommendations can help researchers and product teams identify risks and prevent negative impacts, and ensure that we continue to get better building AI products over time.
Some of the proposals focus on improving our
software development process
for AI products. They describe steps we should take when we develop machine learning algorithms, assess potential product applications for those algorithms, deploy those products on Wikimedia websites, and evaluate success and failure.
Other proposals focus on the design of the AI technologies themselves, and the tools and user interfaces we build around them. They describe ethical
design patterns
intended to allow the readers and contributors who use our AI products to understand how the algorithms work, provide feedback, and take control of their experience.
Looking forward
The technological, social, and regulatory landscape around AI is changing rapidly. The technology industry as a whole has only recently begun to acknowledge that the ethos of “move fast and break things” is neither an effective nor an ethical way to build complex and powerful products capable of having unexpected, disruptive, and often devastating impacts on individuals, communities, and social institutions. As a non-profit, mission-driven organization with a global reach, the Wikimedia Foundation must hold itself to a higher standard. We can’t afford to build first and ask ethical questions later.
In many ways, the Wikimedia movement is ahead of the game here. We are a geographically and culturally diverse group of people united by a common cause. We already practice the kind of transparency, values-driven design, and consensus-based decision-making that are necessary to leverage the opportunities presented by AI technology while avoiding the damage it can cause. Because the Wikimedia Foundation serves as a steward for the code, content, and communities within the Movement, it is important that we consider the kinds of risks outlined in this white paper, and adopting solutions to anticipate and address them.
You can read more about this project on
Meta-Wiki
and view the white paper on
Wikimedia Commons
Jonathan T. Morgan, Senior Design Researcher, Wikimedia Foundation
Bluesky
Email
Link copied
Read more:
AI
Artificial intelligence
Ethical AI
Machine learning
Research
Wikimedia Foundation
Related
Read further in the pursuit of knowledge
Wikipedia’s value in the age of generative AI
Technology
If there was a generative artificial intelligence system that could, on its own, write all the information contained in Wikipedia, would it be the same as Wikipedia today?
12 July 2023
By
Selena Deckelmann
First grants announced from the Wikimedia Endowment to support technical innovation across Wikipedia and Wikimedia projects
Technology
The Wikimedia Endowment, the long-term fund established in 2016 to support the future of Wikimedia sites, has announced its first recipients of grant funding. The initiatives that will receive grant funding include Abstract Wikipedia, Kiwix, Machine Learning, and Wikidata.
13 April 2023
By
Wikimedia Foundation
Thanks for the thanks! Examining the efficacy of Wikipedia’s thanks feature
Technology
Much like Asgard, Wikipedia is not a place but a people. Specifically, an extensive community of volunteers from around the world who donate their time each day to building, curating, and watching over the largest collection of knowledge ever assembled. Unlike Asgard, the people who contribute to Wikipedia often only rarely meet outside of the….
25 September 2019
By
Swati Goel
Ashton Anderson
and
Leila Zia
Help us unlock the world’s knowledge.
As a nonprofit, Wikipedia and our related free knowledge projects are powered primarily through donations.
Donate now
Questions about the Wikimedia Foundation or our projects? Get in touch with our team.
Contact
Follow
Bluesky
Photo credits
Bothrops bilineatus PN Carrasco
File provided by Wikimedia Commons
Edunavia1
CC BY-SA 4.0
Open the Knowledge Journalism Awards - Exteded
ICFJ
Wikipedia 25 - Birthday blog header
Wikimedia Foundation
CC BY-SA 4.0
Vitoria - Graffiti & Murals 0838
File provided by Wikimedia Commons
Unknown; Photo by Zarateman
CC0
AI Blog header
Wikimedia Foundation
CC BY-SA 4.0
A teacher trying Kiwix
File provided by Wikimedia Commons
Kunokuno
CC BY-SA 4.0
Thank you (Unsplash)
File provided by Wikimedia Commons
Aaron Burden
Unsplash prior to 5 June 2017/CC0
Art+Feminism Wikipedia edit-a-thon
File provided by Wikimedia Commons
Jens Mohr
CC BY-SA 3.0
US