Big Data

This Is What a Facial-Detection Algorithm Looks Like in 3D

art4all/Shutterstock.com

Once the heady stuff of films like Minority Report, facial recognition algorithms are now a part of so many technologies that it’s hard to keep track of them. iPhones, iPhoto, and Facebook all try to detect bodies and faces with biometric software. Just last week, Facebook announced that its proprietary algorithm correctly identified faces 97.25 percent of the time—meaning it works almost as well as humans do.  

But when these mathematical engines hunt for a face in a video, just what do they look for? How does software see a face?

The artist and writer Zach Blas has tried to show just what programs see with his project Face Cages. He’s constructed, literally, face cages—metallic sculptures that fit, painfully, onto a user’s face—that represent the shapes and polygons that algorithms use to hunt for faces.

He’s made the invisible world of algorithms, in other words, a little more IRL.

A whole slew of artists are up to something similar. The New York-based Adam Harvey has made anti-algorithmic make-up that thwarts the facial patterns algorithms look for, just as dazzle camouflage on Navy ships thwarted artillery in World War I. Simone Niquille, in Amsterdam, has extended this idea,printing t-shirts so full of faces that algorithms don’t know which one to focus on.

And the London-based artist James Bridle has made a“surveillance spaulder”: A shoulder pad that alerts wearers when they’re under the gaze of a surveillance camera.

All these are little projects of disrupting or reminding—tools to frustrate the software, or to alert people to the its existence. Part of Blas’s goal here is maybe even more political: He would like to remind viewers that faces vary wildly, and that algorithms can’t capture their full diversity.

“Asian women’s hands fail to be legible to fingerprint devices; eyes with cataracts hinder iris scans,” he argues. To Blas, besides all their other possible faults, facial-detection algorithms exacerbate all the differences that people are already punished for. The perfectible accounting of the algorithm—97.25 percent efficiency!—when forced onto a fleshy face recalls, for Blas, “handcuffs, prison bars, and torture devices used during slavery.”

No wonder, then, his face cages are part of a much larger—and still-in-progress—project, the Facial Weaponization Suite.

(Image via art4all/Shutterstock.com)

Threatwatch Alert

Spearphishing / Stolen credentials / User accounts compromised

Hackers fund flights to London for Romanian pickpockets through iThing scam

See threatwatch report

JOIN THE DISCUSSION

Close [ x ] More from Nextgov
// July 23