Data Violence and How Bad Engineering Choices Can Damage Society


Blog by Anna Lauren Hoffmann: “…In 2015, a black developer in New York discovered that Google’s algorithmic photo recognition software had tagged pictures of him and his friends as gorillas.

The same year, Facebook auto-suspended Native Americans for using their real names, and in 2016, facial recognition was found to struggle to read black faces.

Software in airport body scanners has flagged transgender bodies as threatsfor years. In 2017, Google Translate took gender-neutral pronouns in Turkish and converted them to gendered pronouns in English — with startlingly biased results.

“Violence” might seem like a dramatic way to talk about these accidents of engineering and the processes of gathering data and using algorithms to interpret it. Yet just like physical violence in the real world, this kind of “data violence” (a term inspired by Dean Spade’s concept of administrative violence) occurs as the result of choices that implicitly and explicitly lead to harmful or even fatal outcomes.

Those choices are built on assumptions and prejudices about people, intimately weaving them into processes and results that reinforce biases and, worse, make them seem natural or given.

Take the experience of being a woman and having to constantly push back against rigid stereotypes and aggressive objectification.

Writer and novelist Kate Zambreno describes these biases as “ghosts,” a violent haunting of our true reality. “A return to these old roles that we play, that we didn’t even originate. All the ghosts of the past. Ghosts that aren’t even our ghosts.”

Structural bias is reinforced by the stereotypes fed to us in novels, films, and a pervasive cultural narrative that shapes the lives of real women every day, Zambreno describes. This extends to data and automated systems that now mediate our lives as well. Our viewing and shopping habits, our health and fitness tracking, our financial information all conspire to create a “data double” of ourselves, produced about us by third parties and standing in for us on data-driven systems and platforms.

These fabrications don’t emerge de novo, disconnected from history or social context. Rather, they often pick up and unwittingly spit out a tangled mess of historical conditions and current realities.

Search engines are a prime example of how data and algorithms can conspire to amplify racist and sexist biases. The academic Safiya Umoja Noble threw these messy entanglements into sharp relief in her book Algorithms of OppressionGoogle Search, she explains, has a history of offering up pages of porn for women from particular racial or ethnic groups, and especially black women. Google have also served up ads for criminal background checksalongside search results for African American–sounding names, as former Federal Trade Commission CTO Latanya Sweeney discovered.

“These search engine results for women whose identities are already maligned in the media, such as Black women and girls, only further debase and erode efforts for social, political, and economic recognition and justice,” Noble says.

These kinds of cultural harms go well beyond search results. Sociologist Rena Bivens has shown how the gender categories employed by platforms like Facebook can inflict symbolic violences against transgender and nonbinary users in ways that may never be made obvious to users….(More)”.

Data Stewards: Data Leadership to Address the Challenges of the 21st Century


Data Stewards_screenshot

The GovLab at the NYU Tandon School of Engineering is pleased to announce the launch of its Data Stewards website — a new portal for connecting organizations across sectors that seek to promote responsible data leadership that can address the challenges of the 21st century — developed with generous support from the William and Flora Hewlett Foundation.

Increasingly, the private sector is collaborating with the public sector and researchers on ways to use private-sector data and analytical expertise for public good. With these new practices of data collaborations come the need to reimagine roles and responsibilities to steer the process of using this data, and the insights it can generate, to address society’s biggest questions and challenges: Data Stewards.

Today, establishing and sustaining these new collaborative and accountable approaches requires significant and time-consuming effort and investment of resources for both data holders on the supply side, and institutions that represent the demand. By establishing Data Stewardship as a function — recognized within the private sector as a valued responsibility — the practice of Data Collaboratives can become more predictable, scaleable, sustainable and de-risked.

Together with BrightFront Group and Adapt, we are:

  • Exploring the needs and priorities of current private sector Data Stewards who act as change agents within their firms. Responsible for determining what, when, how and with whom to share private data for public good, these individuals are critical catalysts for ensuring insights are turned into action.
  • Identifying and connecting existing Data Stewards across sectors and regions to create an online and in-person community for exchanging knowledge and best practices.
  • Developing methodologies, tools and frameworks to use data more responsibly, systematically and efficiently to decrease the transaction cost, time and energy currently needed to establish Data Collaboratives.

To learn more about the Data Stewards Initiative, including new insights, ideas, tools and information about the Data Steward of the Year Award program, visit datastewards.net.

If you are a Data Steward, or would like to join a community of practice to learn from your peers, please contact [email protected] to join the Network of Data Stewards.

For more information about The GovLab, visit thegovlab.org.

Using Blockchain Technology to Create Positive Social Impact


Randall Minas in Healthcare Informatics: “…Healthcare is yet another area where blockchain can make a substantial impact. Blockchain technology could be used to enable the WHO and CDC to better monitor disease outbreaks over time by creating distributed “ledgers” that are both secure and updated hundreds of times per day. Issued in near real-time, these updates would alert healthcare professionals to spikes in local cases almost immediately. Additionally, using blockchain would allow accurate diagnosis and streamline the isolation of clusters of cases as quickly as possible. Providing blocks of real-time disease information—especially in urban areas—would be invaluable.

In the United States, disease updates are provided in a Morbidity and Mortality Weekly Report (MMWR) from the CDC. This weekly report provides tables of current disease trends for hospitals and public health officials. Another disease reporting mechanism is the National Outbreak Reporting System (NORS), launched in 2009. NORS’ web-based tool provides outbreak data through 2016 and is accessible to the general public. There are two current weaknesses in the NORS reporting system and both can be addressed by blockchain technology.

The first issue lies in the number of steps required to accurately report each outbreak. A health department reports an outbreak to the NORS system, the CDC checks it for accuracy, analyzes the data, then provides a summary via the MMRW. Instantiating blockchain as the technology through which the NORS data is reported, every health department in the country could have preliminary data on disease trends at their fingertips without having to wait for the next MMRW publication.

The second issue is the inherent cybersecurity vulnerabilities using a web-based platform to monitor disease reporting. As we have seen with cyberattacks both domestic and abroad, cybersecurity vulnerabilities underlie most of our modern-day computing infrastructure. Blockchain was designed to be secure because it is decentralized across many computer networks and, since it was designed as a digital ledger, the previous data (or “blocks”) in the blockchain are difficult to alter.

While the NORS platform could be hacked with malware to gain access to our electricity and water infrastructure, instituting blockchain technology would limit the potential damage of the malware based on the inherent security of the technology. If this does not sound important, imagine the damage and ensuing panic that could be caused by a compromised NORS reporting a widespread Ebola outbreak.

The use of blockchain in monitoring epidemic outbreaks might not only apply to fast-spreading outbreaks like the flu, but also to epidemics that have lasted for decades. Since blockchain allows an unchangeable snapshot of data over time and can be anonymous, partner organizations could provide HIV test results to an individual’s “digital ledger” with a date of the test and the results.

Individuals could then exchange their HIV status securely, in an application, before engaging in high-risk behaviors. Since many municipalities provide free or low-cost, anonymous HIV testing, the use of blockchain would allow disease monitoring and exchange of status in a secure and trusted manner. The LGBTQ community and other high-risk communities could use an application to securely exchange HIV status with potential partners. With widespread adoption of this status-exchange system, an individual’s high-risk exposure could be limited, further reducing the spread of the epidemic.

While much of the creative application around blockchain has focused on supply chain-like models, including distribution of renewable energy and local sourcing of goods, it is important also to think innovatively about how blockchain can be used outside of supply chain and accounting.

In healthcare, blockchain has been discussed frequently in relation to electronic health records (EHRs), yet even that could be underappreciating the technology’s potential. Leaders in the blockchain arena should invest in application development for epidemic monitoring and disease control using blockchain technology. …(More)”.

Why Policymakers Should Care About “Big Data” in Healthcare


David W.Bates et al at Health Policy and Technology: “The term “big data” has gotten increasing popular attention, and there is growing focus on how such data can be used to measure and improve health and healthcare. Analytic techniques for extracting information from these data have grown vastly more powerful, and they are now broadly available. But for these approaches to be most useful, large amounts of data must be available, and barriers to use should be low. We discuss how “smart cities” are beginning to invest in this area to improve the health of their populations; provide examples around model approaches for making large quantities of data available to researchers and clinicians among other stakeholders; discuss the current state of big data approaches to improve clinical care including specific examples, and then discuss some of the policy issues around and examples of successful regulatory approaches, including deidentification and privacy protection….(More)”.

Superminds: The Surprising Power of People and Computers Thinking Together


Book by Thomas W. Malone: “If you’re like most people, you probably believe that humans are the most intelligent animals on our planet. But there’s another kind of entity that can be far smarter: groups of people. In this groundbreaking book, Thomas Malone, the founding director of the MIT Center for Collective Intelligence, shows how groups of people working together in superminds — like hierarchies, markets, democracies, and communities — have been responsible for almost all human achievements in business, government, science, and beyond. And these collectively intelligent human groups are about to get much smarter.

Using dozens of striking examples and case studies, Malone shows how computers can help create more intelligent superminds not just with artificial intelligence, but perhaps even more importantly with hyperconnectivity:  connecting humans to one another at massive scales and in rich new ways. Together, these changes will have far-reaching implications for everything from the way we buy groceries and plan business strategies to how we respond to climate change, and even for democracy itself. By understanding how these collectively intelligent groups work, we can learn how to harness their genius to achieve our human goals….(More)”.

Networked publics: multi-disciplinary perspectives on big policy issues


Special issue of Internet Policy Review edited by William Dutton: “…is the first to bring together the best policy-oriented papers presented at the annual conference of the Association of Internet Researchers (AoIR). This issue is anchored in the 2017 conference in Tartu, Estonia, which was organised around the theme of networked publics. The seven papers span issues concerning whether and how technology and policy are reshaping access to information, perspectives on privacy and security online, and social and legal perspectives on informed consent of internet users. As explained in the editorial to this issue, taken together, the contributions to this issue reflect the rise of new policy, regulatory and governance issues around the internet and social media, an ascendance of disciplinary perspectives in what is arguably an interdisciplinary field, and the value that theoretical perspectives from cultural studies, law and the social sciences can bring to internet policy research.

Editorial: Networked publics: multi-disciplinary perspectives on big policy issues
William H. Dutton, Michigan State University

Political topic-communities and their framing practices in the Dutch Twittersphere
Maranke Wieringa, Daniela van Geenen, Mirko Tobias Schäfer, & Ludo Gorzeman

Big crisis data: generality-singularity tensions
Karolin Eva Kappler

Cryptographic imaginaries and the networked public
Sarah Myers West

Not just one, but many ‘Rights to be Forgotten’
Geert Van Calster, Alejandro Gonzalez Arreaza, & Elsemiek Apers

What kind of cyber security? Theorising cyber security and mapping approaches
Laura Fichtner

Algorithmic governance and the need for consumer empowerment in data-driven markets
Stefan Larsson

Standard form contracts and a smart contract future
Kristin B. Cornelius

…(More)”.

International Data Flows and Privacy: The Conflict and its Resolution


World Bank Policy Research Working Paper by Aaditya Mattoo and Joshua P Meltzer: “The free flow of data across borders underpins today’s globalized economy. But the flow of personal dataoutside the jurisdiction of national regulators also raises concerns about the protection of privacy. Addressing these legitimate concerns without undermining international integration is a challenge. This paper describes and assesses three types of responses to this challenge: unilateral development of national or regional regulation, such as the European Union’s Data Protection Directive and forthcoming General Data Protection Regulation; international negotiation of trade disciplines, most recently in the Comprehensive and Progressive Agreement for Trans-Pacific Partnership (CPTPP); and international cooperation involving regulators, most significantly in the EU-U.S. Privacy Shield Agreement.

The paper argues that unilateral restrictions on data flows are costly and can hurt exports, especially of data-processing and other data-based services; international trade rules that limit only the importers’ freedom to regulate cannot address the challenge posed by privacy; and regulatory cooperation that aims at harmonization and mutual recognition is not likely to succeed, given the desirable divergence in national privacy regulation. The way forward is to design trade rules (as the CPTPP seeks to do) that reflect the bargain central to successful international cooperation (as in the EU-US Privacy Shield): regulators in data destination countries would assume legal obligations to protect the privacy of foreign citizens in return for obligations on data source countries not to restrict the flow of data. Existing multilateral rules can help ensure that any such arrangements do not discriminate against and are open to participation by other countries….(More)”.

The Future of Fishing Is Big Data and Artificial Intelligence


Meg Wilcox at Civil Eats: “New England’s groundfish season is in full swing, as hundreds of dayboat fishermen from Rhode Island to Maine take to the water in search of the region’s iconic cod and haddock. But this year, several dozen of them are hauling in their catch under the watchful eye of video cameras as part of a new effort to use technology to better sustain the area’s fisheries and the communities that depend on them.

Video observation on fishing boats—electronic monitoring—is picking up steam in the Northeast and nationally as a cost-effective means to ensure that fishing vessels aren’t catching more fish than allowed while informing local fisheries management. While several issues remain to be solved before the technology can be widely deployed—such as the costs of reviewing and storing data—electronic monitoring is beginning to deliver on its potential to lower fishermen’s costs, provide scientists with better data, restore trust where it’s broken, and ultimately help consumers gain a greater understanding of where their seafood is coming from….

Muto’s vessel was outfitted with cameras, at a cost of about $8,000, through a collaborative venture between NOAA’s regional office and science centerThe Nature Conservancy (TNC), the Gulf of Maine Research Institute, and the Cape Cod Commercial Fishermen’s Alliance. Camera costs are currently subsidized by NOAA Fisheries and its partners.

The cameras run the entire time Muto and his crew are out on the water. They record how the fisherman handle their discards, the fish they’re not allowed to keep because of size or species type, but that count towards their quotas. The cost is lower than what he’d pay for an in-person monitor.The biggest cost of electronic monitoring, however, is the labor required to review the video. …

Another way to cut costs is to use computers to review the footage. McGuire says there’s been a lot of talk about automating the review, but the common refrain is that it’s still five years off.

To spur faster action, TNC last year spearheaded an online competition, offering a $50,000 prize to computer scientists who could crack the code—that is, teach a computer how to count fish, size them, and identify their species.

“We created an arms race,” says McGuire. “That’s why you do a competition. You’ll never get the top minds to do this because they don’t care about your fish. They all want to work for Google, and one way to get recognized by Google is to win a few of these competitions.”The contest exceeded McGuire’s expectations. “Winners got close to 100 percent in count and 75 percent accurate on identifying species,” he says. “We proved that automated review is now. Not in five years. And now all of the video-review companies are investing in machine leaning.” It’s only a matter of time before a commercial product is available, McGuire believes….(More).

Open Standards for Data


Guidebook by the Open Data Institute: “Standards for data are often seen as a technical topic that is only relevant to developers and other technologists.

Using this guidebook we hope to highlight that standards are an important tool that are worthy of wider attention.

Standards have an important role in helping us to consistently and repeatably share data. But they are also a tool to help implement policy, create and shape markets and drive social change.

The guidebook isn’t intended to be read from start to finish. Instead we’ve focused on curating a variety of guidance, tools and resources that will be relevant no matter your experience.

On top of providing useful background and case studies, we’ve also provided pointers to help you find existing standards.

Other parts of the guidebook will be most relevant when you’re engaged in the process of scoping and designing new standards….(More)”.

The promise and peril of military applications of artificial intelligence


Michael C. Horowitz at the Bulletin of the Atomic Scientists: “Artificial intelligence (AI) is having a moment in the national security space. While the public may still equate the notion of artificial intelligence in the military context with the humanoid robots of the Terminatorfranchise, there has been a significant growth in discussions about the national security consequences of artificial intelligence. These discussions span academia, business, and governments, from Oxford philosopher Nick Bostrom’s concern about the existential risk to humanity posed by artificial intelligence to Tesla founder Elon Musk’s concern that artificial intelligence could trigger World War III to Vladimir Putin’s statement that leadership in AI will be essential to global power in the 21st century.

What does this really mean, especially when you move beyond the rhetoric of revolutionary change and think about the real world consequences of potential applications of artificial intelligence to militaries? Artificial intelligence is not a weapon. Instead, artificial intelligence, from a military perspective, is an enabler, much like electricity and the combustion engine. Thus, the effect of artificial intelligence on military power and international conflict will depend on particular applications of AI for militaries and policymakers. What follows are key issues for thinking about the military consequences of artificial intelligence, including principles for evaluating what artificial intelligence “is” and how it compares to technological changes in the past, what militaries might use artificial intelligence for, potential limitations to the use of artificial intelligence, and then the impact of AI military applications for international politics.

The potential promise of AI—including its ability to improve the speed and accuracy of everything from logistics to battlefield planning and to help improve human decision-making—is driving militaries around the world to accelerate their research into and development of AI applications. For the US military, AI offers a new avenue to sustain its military superiority while potentially reducing costs and risk to US soldiers. For others, especially Russia and China, AI offers something potentially even more valuable—the ability to disrupt US military superiority. National competition in AI leadership is as much or more an issue of economic competition and leadership than anything else, but the potential military impact is also clear. There is significant uncertainty about the pace and trajectory of artificial intelligence research, which means it is always possible that the promise of AI will turn into more hype than reality. Moreover, safety and reliability concerns could limit the ways that militaries choose to employ AI…(More)”,