New York Just Set a ‘Dangerous Precedent’ on Algorithms, Experts Warn

Ryan DeBerardinis/Shutterstock.com

NYC’s task force on algorithms was supposed to be a beacon of transparent government. It couldn’t even gain access to basic information.

It was supposed to be groundbreaking. When New York City’s task force to develop policy on algorithm technologies was introduced two years ago, it was praised as a beacon of transparent and equitable government. It was supposed to inform other policymakers grappling with how to address their own use of automated technologies that make decisions in place of humans.

But for all its good intentions, the effort was bogged down in a bureaucratic morass. The task force failed at even completing a first necessary step in its work: getting access to basic information about automated systems already in use, according to task force members and observers.

“The fact they were unable to even get information about what tools the city was using is very problematic,” said Deirdre Mulligan, associate professor at the UC Berkeley School of Information. Algorithmic tools “in and of themselves embed really significant policy choices.” A spokesman for New York City did not respond to an inquiry about why the task force did not provide a catalog of automated systems.

shadow report published last week in response to the initiative has issued a stark warning cry about “the limitations of existing bureaucratic procedures” to tackle automated decisions.

Cities use algorithmic and automated technologies to make all sorts of decisions, from determining where kids are sent to school, to deciding whether someone should be released on bail before a criminal trial.

But thus far, there have been few protections to ensure these technologies are making fair decisions. And early experiences with AI tools have shown that they can be categorically unfair, baking into automated tools the biases that already existed in the humans and data that trained them. For example, algorithms used by law enforcement to assess and score criminal or child safety risks have been found to produce results that can negatively and disproportionately affect African Americans.  

In other words, automated systems make decisions that affect people’s lives in very real ways right now, said Meredith Whittaker, co-founder of the AI Now Institute at NYU and a member of New York City’s Automated Decision System task force. “This is not a theoretical concern,” she told CityLab.

It was this sort of concern that led New York City to create a task force as a first step toward setting policy around the use of algorithmic tools. The task force did convene and make some structural recommendations. Among the most significant: It called for creation of an “organizational structure” within city government to serve as a centralized hub for future policy guidance and automated systems management. After the report was unveiled, Mayor Bill de Blasio responded to this recommendation by announcing a new Algorithms Management and Policy Officer position within the Mayor’s Office of Operations.

“This Officer will be carrying the important work of the Task Force forward so that city agency use of algorithms will be fair, equitable, and transparent,” the New York City Hall spokesperson told CityLab.  In a press release touting the new position, members of the task force, some of whom serve in city government roles, praised this first step.

But Rashida Richardson, editor of the shadow report and director of policy research at the AI Now Institute at NYU, said the new algorithms officer role is “is just an effort for the Mayor and the city to seem like they did something.”

And one task force member who does not work for city government, Julia Stoyanovich, assistant professor of computer science and engineering at NYU, acknowledged in the release that automated decision system “management and oversight are difficult.” She continued, "I am particularly looking forward to seeing evidence of a strong commitment on the part of the City to a broad education effort.”

Part of the challenge for task force members was that city agencies, accustomed to keeping details of vendor technologies secret, were unwilling to provide even a basic list of automated systems already in use.

“There were delays, and obfuscations, and then, by the spring of 2019, outright denials. It wasn’t because the data didn’t exist,” wrote Albert Fox Cahn, executive director of civil rights and privacy group Surveillance Technology Oversight Project, who participated in some task force meetings, in FastCompany. “The City Council passed the task force into law to hold the administration accountable for its secretive use of algorithms, but that was the last thing the administration wanted.”

In response, a city spokesman said: “Mr. Cahn represented his former employer three times out of several dozen Task Force meetings, working sessions, public forums, and community roundtables. In his op-ed, Mr. Cahn has completely misrepresented the facts of the process, and his understanding of the Task Force's legal mandate is fundamentally flawed.”

If governments knew more about the systems they use, they could evaluate their potential impact through processes like algorithmic impact assessments. This is why civil rights and tech justice advocates want governments to reveal technical details about automated systems including source code, algorithms, and the data employed for training.  

But illuminating these details has been a struggle, especially as tech firms insist on protecting such information from disclosure. If this level of disclosure ever happens, it is very rare.

Indeed, the city’s own official task force report acknowledged the group “faced challenges” in balancing government transparency with considerations including “proprietary and other sensitive types of City information.”

The shadow report, endorsed by the ACLU, NAACP Legal Defense and Educational Fund, and more than 20 legal, human rights and tech justice organizations, argued that municipalities need to do a better job of ensuring control over contracts with tech providers. Its authors, technology experts and advocates for civil rights and algorithmic accountability, called for New York’s city council to pass a law requiring more transparent procurement of automated decision systems.

Richardson said deep-pocketed municipalities like New York could use their leverage when negotiating tech contracts to require more disclosure as a condition of doing business with the city. “New York City should use its purchasing power to only contract with firms willing to disclose the details of their technology, how they will be used, and who could potentially be harmed. Any firm not willing to be transparent about their technology should be left off the NYC payroll.”

Richardson, who was part of a coalition that sought more community engagement during the task force process, worried that the failure to make any meaningful progress in understanding how New York City is using algorithms will set a perverse example for other governments.

“The concern is around New York City setting dangerous precedent that the city doesn’t need to do anything,” she said.

In fact, New York City isn’t the only place that’s considered how to reform its use of algorithms. But even cities that have already mulled more aggressive regulation of algorithmic technologies haven’t disclosed much about their use of them. Portland, Oregon, is weighing a ban on algorithm-based facial recognition tech and recently passed a privacy resolution calling for evaluation of automated decision system impacts. But it has not conducted an inventory of automated decision systems, said Hector Dominguez, open data coordinator at the city’s data technology oversight body Smart City PDX.

“We are looking at what NYC, San Francisco, Oakland and Seattle and other sister cities are doing,” Dominguez wrote in an email.

San Francisco and Oakland have both banned facial recognition technology use by city government. Seattle has not, but the city updates a list of surveillance technologies every quarter and includes non-surveillance tech systems used by city agencies. The city and county of San Francisco unveiled a toolkit for assessing algorithmic impact in 2018.

A movement to update traditional government technology purchasing policies for these and other new technologies is already underway. Berkeley professors Mulligan and Kenneth Bamberger have called for “a move from a procurement mindset to policymaking mindset.”

In a paper expected to be published in the Berkeley Technology Law Journal next year, they write, “Through policy choices and design, we can build purposeful tools that are aligned with values chosen based on reason, expertise, transparency, and robust and ongoing deliberation and oversight.”

X
This website uses cookies to enhance user experience and to analyze performance and traffic on our website. We also share information about your use of our site with our social media, advertising and analytics partners. Learn More / Do Not Sell My Personal Information
Accept Cookies
X
Cookie Preferences Cookie List

Do Not Sell My Personal Information

When you visit our website, we store cookies on your browser to collect information. The information collected might relate to you, your preferences or your device, and is mostly used to make the site work as you expect it to and to provide a more personalized web experience. However, you can choose not to allow certain types of cookies, which may impact your experience of the site and the services we are able to offer. Click on the different category headings to find out more and change our default settings according to your preference. You cannot opt-out of our First Party Strictly Necessary Cookies as they are deployed in order to ensure the proper functioning of our website (such as prompting the cookie banner and remembering your settings, to log into your account, to redirect you when you log out, etc.). For more information about the First and Third Party Cookies used please follow this link.

Allow All Cookies

Manage Consent Preferences

Strictly Necessary Cookies - Always Active

We do not allow you to opt-out of our certain cookies, as they are necessary to ensure the proper functioning of our website (such as prompting our cookie banner and remembering your privacy choices) and/or to monitor site performance. These cookies are not used in a way that constitutes a “sale” of your data under the CCPA. You can set your browser to block or alert you about these cookies, but some parts of the site will not work as intended if you do so. You can usually find these settings in the Options or Preferences menu of your browser. Visit www.allaboutcookies.org to learn more.

Sale of Personal Data, Targeting & Social Media Cookies

Under the California Consumer Privacy Act, you have the right to opt-out of the sale of your personal information to third parties. These cookies collect information for analytics and to personalize your experience with targeted ads. You may exercise your right to opt out of the sale of personal information by using this toggle switch. If you opt out we will not be able to offer you personalised ads and will not hand over your personal information to any third parties. Additionally, you may contact our legal department for further clarification about your rights as a California consumer by using this Exercise My Rights link

If you have enabled privacy controls on your browser (such as a plugin), we have to take that as a valid request to opt-out. Therefore we would not be able to track your activity through the web. This may affect our ability to personalize ads according to your preferences.

Targeting cookies may be set through our site by our advertising partners. They may be used by those companies to build a profile of your interests and show you relevant adverts on other sites. They do not store directly personal information, but are based on uniquely identifying your browser and internet device. If you do not allow these cookies, you will experience less targeted advertising.

Social media cookies are set by a range of social media services that we have added to the site to enable you to share our content with your friends and networks. They are capable of tracking your browser across other sites and building up a profile of your interests. This may impact the content and messages you see on other websites you visit. If you do not allow these cookies you may not be able to use or see these sharing tools.

If you want to opt out of all of our lead reports and lists, please submit a privacy request at our Do Not Sell page.

Save Settings
Cookie Preferences Cookie List

Cookie List

A cookie is a small piece of data (text file) that a website – when visited by a user – asks your browser to store on your device in order to remember information about you, such as your language preference or login information. Those cookies are set by us and called first-party cookies. We also use third-party cookies – which are cookies from a domain different than the domain of the website you are visiting – for our advertising and marketing efforts. More specifically, we use cookies and other tracking technologies for the following purposes:

Strictly Necessary Cookies

We do not allow you to opt-out of our certain cookies, as they are necessary to ensure the proper functioning of our website (such as prompting our cookie banner and remembering your privacy choices) and/or to monitor site performance. These cookies are not used in a way that constitutes a “sale” of your data under the CCPA. You can set your browser to block or alert you about these cookies, but some parts of the site will not work as intended if you do so. You can usually find these settings in the Options or Preferences menu of your browser. Visit www.allaboutcookies.org to learn more.

Functional Cookies

We do not allow you to opt-out of our certain cookies, as they are necessary to ensure the proper functioning of our website (such as prompting our cookie banner and remembering your privacy choices) and/or to monitor site performance. These cookies are not used in a way that constitutes a “sale” of your data under the CCPA. You can set your browser to block or alert you about these cookies, but some parts of the site will not work as intended if you do so. You can usually find these settings in the Options or Preferences menu of your browser. Visit www.allaboutcookies.org to learn more.

Performance Cookies

We do not allow you to opt-out of our certain cookies, as they are necessary to ensure the proper functioning of our website (such as prompting our cookie banner and remembering your privacy choices) and/or to monitor site performance. These cookies are not used in a way that constitutes a “sale” of your data under the CCPA. You can set your browser to block or alert you about these cookies, but some parts of the site will not work as intended if you do so. You can usually find these settings in the Options or Preferences menu of your browser. Visit www.allaboutcookies.org to learn more.

Sale of Personal Data

We also use cookies to personalize your experience on our websites, including by determining the most relevant content and advertisements to show you, and to monitor site traffic and performance, so that we may improve our websites and your experience. You may opt out of our use of such cookies (and the associated “sale” of your Personal Information) by using this toggle switch. You will still see some advertising, regardless of your selection. Because we do not track you across different devices, browsers and GEMG properties, your selection will take effect only on this browser, this device and this website.

Social Media Cookies

We also use cookies to personalize your experience on our websites, including by determining the most relevant content and advertisements to show you, and to monitor site traffic and performance, so that we may improve our websites and your experience. You may opt out of our use of such cookies (and the associated “sale” of your Personal Information) by using this toggle switch. You will still see some advertising, regardless of your selection. Because we do not track you across different devices, browsers and GEMG properties, your selection will take effect only on this browser, this device and this website.

Targeting Cookies

We also use cookies to personalize your experience on our websites, including by determining the most relevant content and advertisements to show you, and to monitor site traffic and performance, so that we may improve our websites and your experience. You may opt out of our use of such cookies (and the associated “sale” of your Personal Information) by using this toggle switch. You will still see some advertising, regardless of your selection. Because we do not track you across different devices, browsers and GEMG properties, your selection will take effect only on this browser, this device and this website.