A Pew report suggests corporations and governments could misuse algorithms.
Some tech experts fear the mathematical equations underlying much of today's data analytics technology could ultimately be dangerous for individuals.
In a poll of about 1,300 people who closely observe technology development, about 37 percent concluded algorithms could have more negative societal impacts than positive ones, according to a new Pew report.
"[S]ometimes, the application of algorithms created with good intentions leads to unintended consequences," said the report, which canvassed futurists including Defense One's Patrick Tucker.
» Get the best federal technology news and ideas delivered right to your inbox. Sign up here.
About 38 percent thought positive impacts would outweigh the negatives, and a quarter thought the impact would be evenly split between positive and negative.
The potential misuse of algorithms has garnered increased attention recently. In May, the White House published a report warning about the downsides of big data. For instance, the algorithm that suggests a credit score for an applicant with no credit history might draw on that person's phone bills or social media connections; but it could also create greater disparities between communities by providing a lower credit score to applicants who share attributes with other low-credit customers, such as shopping at the same grocery store.
Especially when algorithms are making decisions that could have major social impacts, "humans are considered to be an 'input' to the process and they are not seen as real, thinking, feeling, changing beings," the report said. Experts agreed this system creates a "flawed, logic-driven society and that as the process evolves—that is, as algorithms begin to write the algorithms—humans may get left out of the loop, letting the robots decide.'"
The Pew report pointed to Microsoft's recent Twitter experiment—in which a bot named Tay was instructed to communicate with other users and to learn from their behavior, but quickly began spouting racist and sexist messages—as another example of algorithms going awry.
A lack of diversity in the tech workforce could be reinforcing biases coded into algorithms, at least one expert thought.
Across industries, "algorithms will be primarily designed by white and Asian men—with data selected by these same privileged actors—for the benefit of consumers like themselves," wrote Justin Reich, executive director at the MIT Teaching Systems Lab. "The harms of new technology will be most experienced by those already disadvantaged in society, where advertising algorithms offer bailbondsman ads that assume readers are criminals, loan applications that penalize people for proxies so correlated with race that they effectively penalize people based on race and similar issues."
One expert suggested countering these challenges by creating a "Food and Drug Administration for Algorithms" that could involve a "pre-market review to prevent—without stifling innovation—the introduction of unacceptably dangerous algorithms into the market," researcher Andrew Tutt said in that report.
Pew conducted the research with Elon University's Imagining the Internet Center.