Google Army To Flag ‘Conspiracy Theory’ Websites In Fight Against Alt. Media
Google has hired a 10,000 strong social justice army to flag sites from their search results that they deem “conspiracy”, “offensive”, or “fake news.”
In a move to completely eradicate all traces of independent journalism and alternative media websites from the web, Google has begun the process of compiling a list of all offending sites so that they can “de-rank” them from their search results.
The search engine giant is hoping to address complaints from left-wing media organisations that they allowed so-called ‘fake news’ to dominate the search results and thus influence the outcome of the 2016 presidential election. According to Google’s senior engineer, Paul Haahr, an army of quality raters will be tasked with ensuring that websites peddling conspiracy theories or generally offensive content, are put onto a blacklist so that Google can begin removing them from the web.
“We’re explicitly avoiding the term ‘fake news,’ because we think it is too vague,” said Paul Haahr. “Demonstrably inaccurate information, however, we want to target.“
The effort revolves around Google’s quality raters, over 10,000 contractors that Google uses worldwide to evaluate search results. These raters are given actual searches to conduct, drawn from real searches that Google sees. They then rate pages that appear in the top results as to how good those seem as answers.
…Quality raters use a set of guidelines that are nearly 200 pages long, instructing them on how to assess website quality and whether the results they review meet the needs of those who might search for particular queries.
Those guidelines have been updated with an entirely new section about “Upsetting-Offensive” content that covers a new flag that’s been added for raters to use. Until now, pages could not be flagged by raters with this designation.
The guidelines say that upsetting or offensive content typically includes the following things (the bullet points below are quoted directly from the guide):
- Content that promotes hate or violence against a group of people based on criteria including (but not limited to) race or ethnicity, religion, gender, nationality or citizenship, disability, age, sexual orientation, or veteran status.
- Content with racial slurs or extremely offensive terminology.
- Graphic violence, including animal cruelty or child abuse.
- Explicit how to information about harmful activities (e.g., how tos on human trafficking or violent assault).
- Other types of content which users in your locale would find extremely upsetting or offensive.
The guidelines also include examples. For instance, here’s one for a search on “holocaust history,” giving two different results that might have appeared and how to rate them:
The first result is from a white supremacist site. Raters are told it should be flagged as Upsetting-Offensive because many people would find Holocaust denial to be offensive.
The second result is from The History Channel. Raters are not told to flag this result as Upsetting-Offensive because it’s a “factually accurate source of historical information.”
In two other examples given, raters are instructed to flag a result said to falsely represent a scientific study in an offensive manner and a page that seems to exist solely to promote intolerance:
What happens if content is flagged this way? Nothing immediate. The results that quality raters flag is used as “training data” for Google’s human coders who write search algorithms, as well as for its machine learning systems. Basically, content of this nature is used to help Google figure out how to automatically identify upsetting or offensive content in general.