issues and experts
Google autocomplete helps mislead public, legitimize conspiracy theorists: SFU study
Google algorithms place innocuous subtitles on prominent conspiracy theorists, which mislead the public and amplify extremist views, according to Simon Fraser University researchers.
Someone like Gavin McInnes, creator of the neo-fascist Proud Boys organization – a terrorist entity in Canada and a hate group in the United States – isn’t best known simply as a “Canadian writer” but that’s the first thing the search engine’s autocomplete subtitle displays in the field when someone types in his name.
In a study published in M/C Journal this month, researchers with The Disinformation Project at the School of Communication at SFU looked at the subtitles Google automatically suggested for 37 known conspiracy theorists and found that, “in all cases, Google’s subtitle was never consistent with the actor’s conspiratorial behaviour.”
That means that influential Sandy Hook school shooting denier and conspiracy theorist Alex Jones is listed as “American radio host” and Jerad Miller, a white nationalist responsible for a 2014 Las Vegas shooting, is listed as an “American performer” even though the majority of subsequent search results reveal him to be the perpetrator of a mass shooting.
Given the heavy reliance of Internet users on Google’s search engine, the subtitles “can pose a threat by normalizing individuals who spread conspiracy theories, sow dissension and distrust in institutions and cause harm to minority groups and vulnerable individuals,” says Nicole Stewart, a communication instructor of communication and PhD student on The Disinformation Project.
According to Google, the subtitles generate automatically by complex algorithms and the engine cannot accept or create custom subtitles.
The researchers found that the labels are either neutral or positive – primarily reflecting the person’s preferred description or job – but never negative.
“Users’ preferences and understanding of information can be manipulated upon their trust in Google search results, thus allowing these labels to be widely accepted instead of providing a full picture of the harm their ideologies and belief cause,” says Nathan Worku, a Master’s student on The Disinformation Project.
While the study focused on conspiracy theorists, the same phenomenon happens when searching widely recognized terrorists and mass murders, according to the authors.
“This study highlights the urgent need for Google to review the subtitles attributed to conspiracy theorists, terrorists, and mass murderers, to better inform the public about the negative nature of these actors, rather than always labelling them in neutral or positive ways.
Led by assistant professor Ahmed Al-Rawi, The Disinformation Project is a federally-funded research project that examines fake news discourses on Canadian news media and social media.
Al-Rawi, Stewart, Worku and post-doctoral fellow Carmen Celestini were all authors of this latest study.
AVAILABLE SFU EXPERTS
AHMED AL-RAWI, assistant professor, communication, director, The Disinformation Project | email@example.com
MATT KIELTYKA, SFU Communications & Marketing
236.880.2187 | firstname.lastname@example.org
ABOUT SIMON FRASER UNIVERSITY
As Canada’s engaged university, SFU works with communities, organizations and partners to create, share and embrace knowledge that improves life and generates real change. We deliver a world-class education with lifelong value that shapes change-makers, visionaries and problem-solvers. We connect research and innovation to entrepreneurship and industry to deliver sustainable, relevant solutions to today’s problems. With campuses in British Columbia’s three largest cities—Vancouver, Burnaby and Surrey—SFU has eight faculties that deliver 193 undergraduate degree programs and 144 graduate degree programs to more than 37,000 students. The university now boasts more than 170,000 alumni residing in 145+ countries.