Difference between revisions of "2019/Grants/Explicit credibility signal data on wikipedia"

From WikiConference North America
< 2019‎ | Grants
Jump to navigation Jump to search
m
(Moved to https://wikiconference.org/wiki/2019/Grants/Wikipedia_deployment_of_credibility_signals_app)
 
(2 intermediate revisions by the same user not shown)
Line 1: Line 1:
  +
Moved to https://wikiconference.org/wiki/2019/Grants/Wikipedia_deployment_of_credibility_signals_app
{{WCNA 2019 Grant Submission
 
|name=Sandro Hawke
 
|username=Sandro_Hawke
 
|email=sandro{{@}}w3.org
 
|resume=* Larger project website: https://credweb.org
 
* My resume: https://hawke.org/resume-2020/
 
|geography=global
 
|type=Technology
 
|idea=See https://docs.google.com/document/d/1kdwuzWqnh3-As3Uyiyoo2Uk7AnzmXmVTZARBOwpY4gY/edit
 
 
|importance=For Wikipedia, this idea promises to help in the fight against misinformation, making it easier for wikipedians to collaborate among themselves and with the broader world in identifying credible and non-credible sources.
 
 
For the world at large, the stakes are much higher, as this approach has the potential to turn the tide against misinformation across all technology platforms.
 
|inprogress=We are developing the relevant concepts and tools (as seen at https://credweb.org) but have not begun deployment in the wikipedia community or tooling to work with wikipedia data feeds.
 
|relevance=There are many connections between this Credibility Signals work and Wikipedia:
 
 
* Wikipedia has always needed to be able to separate fact from fiction. While it does this very well, these tools might make the task easier. Specifically, this can rapidly highlight which sources have unacceptably low credibility and help with sorting out why particularly sources are viewed as credible or not credible.
 
* Wikipedia has always needed to reduce harm done by careless and malicious users. It does this very well, but again, these tools might make the task easier, assisting in tracking and management of the reputation of users, which can be used in modifying their privileges.
 
* Because of its great expertise in these fields, the Wikipedia community is an excellent proving ground for these technologies. Flaws in the technologies that might eventually lead to failure in the broader media ecosystem are likely to be spotted very quickly by wikipedians, giving time to improve the designs before wider deployment.
 
|impact=If successful, this project will show a clear way that people can collaborate online in protecting themselves and their communities from misinformation. This method can be adopted by communities and platforms around the world to greatly reduce misinformation and other online harms.
 
|scalability=Yes, this plan is phenomenally scalable. If it becomes fully established as a decentralized ecosystem, as designed, it will operate and grow with zero effort or support from us or Wikimedia.
 
 
It is based on existing social practices, where each individual manages their own credibility assessment process (deciding what to believe), using what they can glean from their surroundings, including their social network. This process scales linearly with the number of individuals, with each individual deciding how much of their own resources to devote to each assessment they make. Adding computers and networking to this existing human process should greatly improve the efficiency and accuracy of this process, without altering this scaling behavior.
 
 
In its approach to decentralization, this design avoids any central bottleneck. Every individual and organization is free to deploy as much human and computing resources as they choose, without needing approval or support from us or anyone else. This allows the kind of scaling to billions of users that we see in the web and email, which are similarly decentralized. If the system provides sufficient value to users, as we expect, this approach might grow to global scale in a matter of months.
 
|people=This funding request is to help support my time in leading and organizing this project and doing elements of the work for which I am unable to find volunteers or other funding. I bring experience and expertise in all the necessary challenge areas, including credibility signals, community development, web application development, decentralized systems, and consensus process.
 
|inclusiveness=This project has no direct connection to diversity or inclusiveness. We are committed to working to addressing any indirect impacts which might arise.
 
|challenges=In general, we are reducing risk in this ambitious project by minimizing complexity and using a progression of small prototypes and experiments.
 
 
Challenges include:
 
 
* '''Getting people to look at credibility data'''. Approach: make it salient and visually appealing. For example, see credibility network demo at https://credweb.org/viewer/ which has elements that are compelling and fun; it becomes salient when we let people add in the sources they care about and get to see how others judge those sources. We can bootstrap with existing wikipedia data feeds of likes and reverts as an initial proxy for credibility between wikipedians and draw on existing source credibility work for data on external sources.
 
* '''Getting people to author credibility data'''. Once people are engaged in the data as a consumer, we hypothesize they will be motivated to engage as a producer to "correct" the data, to express what they believe or know. Additionally, a culture of contributing data to help the world, already common among wikipedians, should help. There are a range of ways to simplify or even gamify the contribution step, if necessary.
 
* '''Harmful participants'''. Since we propose to primarily and initially use credibility data hosted on wikipedia user pages, to some degree the existing community safety mechanisms will still apply. We would like to demonstrate, however, that such mechanisms can be largely replaced by credibility data itself. In theory, people observed to do harm can be identified and have their actions demoted like non-credible content.
 
* '''Getting people to trust the system'''. Approach: transparency and feedback. Make it clear which individuals are the source of each bit of data, with clear provenance and change tracking. Have the interface promote a virtuous cycle of improving the data and improving one's own credibility. This is similar to wikipedia's own mechanisms for being trustworthy (to people who know how it works).
 
 
|cost=10k USD for the Wikipedia aspects (outlined here) of the Credibility Signals work
 
|expenses=To support my time on this work
 
|time=Up to 12 months, in three phases:
 
* Phase 1 - up to four months - refine deployment plan, identify partners, settle issues within credweb CG
 
* Phase 2 - about 2 months - active development of tools; release
 
* Phase 3 - up to six months - revise and improve, based on user experienc
 
|previous=Yes, my work has been primarily grant funded for many years. Some highlights with web pages maintained by others:
 
* 2018 Google (see "W3C") https://www.blog.google/outreach-initiatives/google-news-initiative/elevating-quality-journalism/
 
* 2013 Knight Foundation https://knightfoundation.org/articles/introducing-crosscloud-project-get-your-data-out-silos/
 
* 2012 NSF https://www.nsf.gov/awardsearch/showAward?AWD_ID=1313789
 
* 2005 DARPA http://xml.coverpages.org/ni2005-02-21-a.html
 
}}
 

Latest revision as of 05:09, 5 April 2020