If you Google “Was the
Holocaust real?” right now, seven out of the top 10 results will be
Holocaust denial sites. If you Google “Was Hitler bad?,” one of the top
results is an article titled, “10 Reasons Why Hitler Was One Of The Good
Guys.”
In December, responding to weeks of criticism, Google said
that it tweaked it algorithm to push down Holocaust denial and
anti-Semitic sites. But now, just a month later, their fix clearly
hasn’t worked.
In addition to hateful
search results, Google has had a similar problem with its
“autocompletes” — when Google anticipates the rest of a query from its
first word or two. Google autocompletes have often embodied racist and sexist stereotypes. Google image search has also generated biased results, absurdly tagging some photos of black people as “gorillas.”
The result of these horrific
search results can be deadly. Google search results reportedly helped
shape the racism of Dylann Roof, who murdered nine people in a
historically black South Carolina church in 2015. Roof said
that when he Googled “black on white crime, the first website I came to
was the Council of Conservative Citizens,” which is a white supremacist
organization. “I have never been the same since that day,” he said. And
of course, in December, a Facebook-fueled fake news story about Hillary
Clinton prompted a man to shoot up a pizza parlor in Washington D.C.
The fake story reportedly originated in a white supremacist’s tweet.
These terrifying acts of
violence and hate are likely to continue if action isn’t taken. Without a
transparent curation process, the public has a hard time judging the
legitimacy of online sources. In response, a growing movement
of academics, journalists and technologists is calling for more
algorithmic accountability from Silicon Valley giants. As algorithms
take on more importance in all walks of life, they are increasingly a
concern of lawmakers. Here are some steps Silicon Valley companies and legislators should take to move toward more transparency and accountability:
1. Obscure content that’s damaging and not of public interest.
When it comes to search
results about an individual person’s name, many countries have
aggressively forced Google to be more careful in how it provides
information. Thanks to the Court of Justice of the European Union,
Europeans can now request
the removal of certain search results revealing information that is
“inadequate, irrelevant, no longer relevant or excessive,” unless there
is a greater public interest in being able to find the information via a
search on the name of the data subject.
Such removals are a middle
ground between information anarchy and censorship. They neither
disappear information from the internet (it can be found at the original
source) nor allow it to dominate the impression of the aggrieved
individual. They are a kind of obscurity that lets ordinary individuals
avoid having a single incident indefinitely dominate search results on
his or her name. For example, a woman in Spain whose husband was
murdered 20 years ago successfully forced Google Spain to take news of the murder off search results on her name.
Such removals are a middle ground between information anarchy and censorship.
2. Label, monitor and explain hate-driven search results.
In 2004, anti-Semites
boosted a Holocaust-denial site called “Jewwatch” into the top 10
results for the query “Jew.” Ironically, some of those horrified by the
site may have helped by linking to it in order to criticize it. The more
a site is linked to, the more prominence Google’s algorithm gives it in
search results.
Google responded to
complaints by adding a headline at the top of the page entitled “An
explanation of our search results.” A web page linked to the headline
explained why the offensive site appeared so high in the relevant
rankings, thereby distancing Google from the results. The label,
however, no longer appears. In Europe and many other countries,
lawmakers should consider requiring such labeling in the case of obvious
hate speech. To avoid mainstreaming extremism, labels may link to
accounts of the history and purpose of groups with innocuous names like “Council of Conservative Citizens.”
In the U.S., this type of
regulation may be considered a form of “compelled speech,” barred by the
First Amendment. Nevertheless, better labeling practices for food and
drugs have escaped First Amendment scrutiny in the U.S., and why should
information itself be different? As law professor Mark Patterson has demonstrated,
many of our most important sites of commerce are markets for
information: search engines are not offering products and services
themselves but information about products and services, which
may well be decisive in determining which firms and groups fail and
which succeed. If they go unregulated, easily manipulated by whoever can
afford the best search engine optimization, people may be left at the
mercy of unreliable and biased sources.
Better labeling practices for food and drugs have escaped First Amendment scrutiny in the U.S. Why should information itself be different?
3. Audit logs of the data fed into algorithmic systems.
We also need to get to the
bottom of how some racist or anti-Semitic groups and individuals are
manipulating search. We should require immutable audit logs of the data
fed into algorithmic systems. Machine-learning, predictive analytics or
algorithms may be too complex for a person to understand, but the data
records are not.
A relatively simple set of
reforms could vastly increase the ability of entities outside Google and
Facebook to determine whether and how the firms’ results and news feeds
are being manipulated. There is rarely adequate profit motive for firms
themselves to do this — but motivated non-governmental organizations
can help them be better guardians of the public sphere.
4. Possibly ban certain content.
In cases where computational
reasoning behind search results really is too complex to be understood
in conventional narratives or equations intelligible to humans, there is
another regulatory approach available: to limit the types of
information that can be provided.
Though such an approach
would raise constitutional objections in the U.S., nations like France
and Germany have outright banned certain Nazi sites and memorabilia.
Policymakers should also closely study laws regarding “incitement to genocide”
to develop guidelines for censoring hate speech with a clear and
present danger of causing systematic slaughter or violence against
vulnerable groups.
It’s a small price to pay for a public sphere less warped by hatred.
5. Permit limited outside annotations to defamatory posts and hire more humans to judge complaints.
In the U.S. and elsewhere, limited annotations ― “rights of reply”
― could be permitted in certain instances of defamation of individuals
or groups. Google continues to maintain that it doesn’t want human
judgment blurring the autonomy of its algorithms. But even spelling
suggestions depend on human judgment, and in fact, Google developed that
feature not only by means of algorithms but also through a painstaking,
iterative interplay between computer science experts and human beta
testers who report on their satisfaction with various results
configurations.
It’s true that the policy
for alternative spellings can be applied generally and automatically
once the testing is over, while racist and anti-Semitic sites might
require fresh and independent judgment after each complaint. But that is
a small price to pay for a public sphere less warped by hatred.
We should commit to
educating users about the nature of search and other automated content
curation and creation. Search engine users need media literacy to
understand just how unreliable Google can be. But we also need vigilant
regulators to protect the vulnerable and police the worst abuses. Truly
accountable algorithms will only result from a team effort by theorists
and practitioners, lawyers, social scientists, journalists and others.
This is an urgent, global cause with committed and mobilized experts
ready to help. Let’s hope that both digital behemoths and their
regulators are listening.
EDITOR’S NOTE: The WorldPost reached out to Google for comment and received the following from a Google spokesperson.
Search ranking:
Google was built on providing people with high-quality and authoritative results for their search queries. We strive to give users a breadth of content from a variety of sources, and we’re committed to the principle of a free and open web. Understanding which pages on the web best answer a query is a challenging problem, and we don’t always get it right.
When non-authoritative information ranks too high in our search results, we develop scalable, automated approaches to fix the problems, rather than manually removing these one-by-one. We are working on improvements to our algorithm that will help surface more high quality, credible content on the web, and we’ll continue to improve our algorithms over time in order to tackle these challenges.
Autocomplete:
We’ve received a lot of questions about Autocomplete, and we want to help people understand how it works: Autocomplete predictions are algorithmically generated based on users’ search activity and interests. Users search for a wide range of material on the web ― 15 percent of searches we see every day are new. Because of this, terms that appear in Autocomplete may be unexpected or unpleasant. We do our best to prevent offensive terms, like porn and hate speech, from appearing, but we don’t always get it right. Autocomplete isn’t an exact science, and we’re always working to improve our algorithms.
Image search:
Our image search results are a reflection of content from across the web, including the frequency with which types of images appear and the way they’re described online. This means that sometimes unpleasant portrayals of subject matter online can affect what image search results appear for a given query. These results don’t reflect Google’s own opinions or beliefs.
Source:- http://www.huffingtonpost.com/entry/holocaust-google-algorithm_us_587e8628e4b0c147f0bb9893
0 comments:
Post a Comment