Filling Public Data Gaps


Report by Judah Axelrod, Karolina Ramos, and Rebecca Bullied: “Data are central to understanding the lived experiences of different people and communities and can serve as a powerful force for promoting racial equity. Although public data, including foundational sources for policymaking such as the US Census Bureau’s American Community Survey (ACS), offer accessible information on a range of topics, challenges of timeliness, granularity, representativeness, and degrees of disaggregation can limit those data’s utility for real-time analysis. Private data—data produced by private-sector organizations either through standard business or to market as an asset for purchase—can serve as a richer, more granular, and higher-frequency supplement or alternative to public data sources. This raises questions about how well private data assets can offer race-disaggregated insights that can inform policymaking.

In this report, we explore the current landscape of public-private data sharing partnerships that address topic areas where racial equity research faces data gaps: wealth and assets, financial well-being and income, and employment and job quality. We held 20 semistructured interviews with current producers and users of private-sector data and subject matter experts in the areas of data-sharing models and ethical data usage. Our findings are divided into five key themes:

  • Incentives and disincentives, benefits, and risks to public-private data sharing
    Agreements with prestigious public partners can bolster credibility for private firms and broaden their customer base, while public partners benefit from access to real-time, granular, rich data sources. But data sharing is often time and labor intensive, and firms can be concerned with conflicting business interests or diluting the value of proprietary data assets.
  • Availability of race-disaggregated data sources
    We found no examples in our interviews of race-disaggregated data sources related to our thematic focus areas that are available externally. However, there are promising methods for data imputation, linkage, and augmentation through internal surveys.
  • Data collaboratives in practice
    Most public-private data sharing agreements we learned about are between two parties and entail free or “freemium” access. However, we found promising examples of multilateral agreements that diversify the data-sharing landscape.
  • From data champions to data stewards
    We found many examples of informal data champions who bear responsibility for relationship-building and securing data partnerships. This role has yet to mature to an institutionalized data steward within private firms we interviewed, which can make data sharing a fickle process.
  • Considerations for ethical data usage
    Data privacy and transparency about how data are accessed and used are prominent concerns among prospective data users. Interviewees also stressed the importance of not privileging existing quantitative data above qualitative insights in cases where communities have offered long-standing feedback and narratives about their own experiences facing racial inequities, and that policymakers should not use a need to collect more data as an excuse for delaying policy action.

Our research yielded several recommendations for data producers and users that engage in data sharing, and for funders seeking to advance data-sharing efforts and promote racial equity…(More)”

Operationalizing Digital Self Determination


Paper by Stefaan G. Verhulst: “We live in an era of datafication, one in which life is increasingly quantified and transformed into intelligence for private or public benefit. When used responsibly, this offers new opportunities for public good. However, three key forms of asymmetry currently limit this potential, especially for already vulnerable and marginalized groups: data asymmetries, information asymmetries, and agency asymmetries. These asymmetries limit human potential, both in a practical and psychological sense, leading to feelings of disempowerment and eroding public trust in technology. Existing methods to limit asymmetries (e.g., consent) as well as some alternatives under consideration (data ownership, collective ownership, personal information management systems) have limitations to adequately address the challenges at hand. A new principle and practice of digital self-determination (DSD) is therefore required.
DSD is based on existing concepts of self-determination, as articulated in sources as varied as Kantian philosophy and the 1966 International Covenant on Economic, Social and Cultural Rights. Updated for the digital age, DSD contains several key characteristics, including the fact that it has both an individual and collective dimension; is designed to especially benefit vulnerable and marginalized groups; and is context-specific (yet also enforceable). Operationalizing DSD in this (and other) contexts so as to maximize the potential of data while limiting its harms requires a number of steps. In particular, a responsible operationalization of DSD would consider four key prongs or categories of action: processes, people and organizations, policies, and products and technologies…(More)”.

Digital rights and principles: a digital transformation for EU citizens


Press Release: “The Commission welcomes the agreement reached yesterday with the Parliament and the Council on the European declaration on digital rights and principles. The declaration, proposed in January, establishes a clear reference point about the kind of human-centred digital transformation that the EU promotes and defends, at home and abroad.

graphic showing a circle with text Your Digital Principles and different icons with a text below the circle At the heart of Europe's digital transformation

It builds on key EU values and freedoms and will benefit all individuals and businesses. The declaration will also provide a guide for policymakers and companies when dealing with new technologies. The declaration focuses on six key areas: putting people at the centre of the digital transformation; solidarity and inclusion; freedom of choice; participation in digital life; safety and security; and sustainability…(More)” See also: European Digital Rights and Principles

Vulnerable People and Data Protection Law


Book by Gianclaudio Malgieri: “Human vulnerability has traditionally been viewed through the lens of specific groups of people, such as ethnic minorities, children, the elderly, or people with disabilities. With the rise of digital media, our perceptions of vulnerable groups and individuals have been reshaped as new vulnerabilities and different vulnerable sub-groups of users, consumers, citizens, and data subjects emerge.

Vulnerable People and Data Protection Law not only depicts these problems but offers the reader a detailed investigation of the concept of data subjects and a reconceptualisation of the notion of vulnerability within the General Data Protection Regulation. The regulation offers a forward-facing set of tools that – though largely underexplored – are essential in rebalancing power asymmetries and mitigating induced vulnerabilities in the age of artificial intelligence.

This book proposes a layered approach to data subject definition. Considering the new potentialities of the digital market, the new awareness about cognitive weaknesses, and the new philosophical sensitivity about vulnerability conditions, the author looks for a more general definition of vulnerability that goes beyond traditional labels. In doing so, he seeks to promote a ‘vulnerability-aware’ interpretation of the GDPR.

A heuristic analysis that re-interprets the whole GDPR, this work is a must-read for both scholars of data protection law and for policymakers looking to strengthen regulations and protect the data of vulnerable individuals…(More)”.

Digitization, Surveillance, Colonialism


Essay by Carissa Veliz: “As I write these words, articles are mushrooming in newspapers and magazines about how privacy is more important than ever after the Supreme Court ruling that has overturned the constitutionality of the right to have an abortion in the United States. In anti-abortion states, browsing histories, text messages, location data, payment data, and information from period-tracking apps can all be used to prosecute both women seeking an abortion and anyone aiding them. The National Right to Life Committee recently published policy recommendations for anti-abortion states that include criminal penalties for people who provide information about self-managed abortions, whether over the phone or online. Women considering an abortion are often in distress, and now they cannot even reach out to friends or family without endangering themselves and others. 

So far, Texas, Oklahoma, and Idaho have passed citizen-enforced abortion bans, according to which anyone can file a civil lawsuit to report an abortion and have the chance of winning at least ten thousand dollars. This is an incredible incentive to use personal data towards for-profit witch-hunting. Anyone can buy personal data from data brokers and fish for suspicious behavior. The surveillance machinery that we have built in the past two decades can now be put to use by authorities and vigilantes to criminalize pregnant women and their doctors, nurses, pharmacists, friends, and family. How productive.

It is not true, however, that the overturning of Roe v. Wade has made privacy more important than ever. Rather, it has provided yet another illustration of why privacy has always been and always will be important. That it is happening in the United States is helpful, because human beings are prone to thinking that whatever happens “over there” — say, in China now, or in East Germany during the Cold War  to those “other people,” doesn’t happen to us — until it does. 

Privacy is important because it protects us from possible abuses of power. As long as human beings are human beings and organizations are organizations, abuses of power will be a constant temptation and threat. That is why it is supremely reckless to build a surveillance architecture. You never know when that data might be used against you — but you can be fairly confident that sooner or later it will be used against you. Collecting personal data might be convenient, but it is also a ticking bomb; it amounts to sensitive material waiting for the chance to turn into an instance of public shaming, extortion, persecution, discrimination, or identity theft. Do you think you have nothing to hide? So did many American women on June 24, only to realize that week that their period was late. You have plenty to hide — you just don’t know what it is yet and whom you should hide it from.

In the digital age, the challenge of protecting privacy is more formidable than most people imagine — but it is nowhere near impossible, and every bit worth putting up a fight for, if you care about democracy or freedom. The challenge is this: the dogma of our time is to turn analog into digital, and as things stand today, digitization is tantamount to surveillance…(More)”.

How Confucianism could put fear about Artificial Intelligence to bed


Article by Tom Cassauwers: “Western culture has had a long history of individualism, warlike use of technology, Christian apocalyptic thinking and a strong binary between body and soul. These elements might explain the West’s obsession with the technological apocalypse and its opposite: techno-utopianism. In Asia, it’s now common to explain China’s dramatic rise as a leader in AI and robotics as a consequence of state support from the world’s largest economy. But what if — in addition to the massive state investment — China and other Asian nations have another advantage, in the form of Eastern philosophies?

There’s a growing view among independent researchers and philosophers that Confucianism and Buddhism could offer healthy alternative perspectives on the future of technology. And with AI and robots rapidly increasing in importance across industries, it’s time for the West to turn to the East for answers…

So what would a non-Western way of thinking about tech look like? First, there might be a different interpretation of personhood. Both Confucianism and Buddhism potentially open up the way for nonhumans to reach the status of humans. In Confucianism, the state of reaching personhood “is not a given. You need to work to achieve it,” says Wong. The person’s attitude toward certain ethical virtues determines whether or not they reach the status of a human. That also means that “we can attribute personhood to nonhuman things like robots when they play ethically relevant roles and duties as humans,” Wong adds.

Buddhism offers a similar argument, where robots can hypothetically achieve a state of enlightenment, which is present everywhere, not only in humans — an argument made as early as the 1970s by Japanese roboticist Masahiro Mori. It may not be a coincidence that robots enjoy some of their highest social acceptance in Japan, with its Buddhist heritage. “Westerners are generally reluctant about the nature of robotics and AI, considering only humans as true beings, while Easterners more often consider devices as similar to humans,” says Jordi Vallverdú, a professor of philosophy at the Autonomous University of Barcelona….(More)”

A Philosophy for Future Generations


Book by Tiziana Andina: “If societies, like institutions, are built to endure, then the bond that exists between generations must be considered. Constructing a framework to establish a philosophy of future generations, Tiziana Andina explores the factors that make it possible for a society to reproduce over time.

Andina’s study of the diachronic structure of societies considers the never-ending passage of generations, as each new generation comes to form a part of the new social fabric and political model.

Her model draws on the anthropologies offered by classical political philosophies such as Hobbes and Machiavelli and the philosophies of power as discussed by Nietzsche. She confronts the ethics and function of this fundamental relationship, examines the role of transgenerationality in the formation and endurance of Western democracies and recognizes an often overlooked problem: each new generation must form part of social and political arrangements designed for them by the generations that came before…(More)”.

Can Privacy Nudges be Tailored to Individuals’ Decision Making and Personality Traits?


Paper by Logan Warberg, Alessandro Acquisti and Douglas Sicker: “While the effectiveness of nudges in influencing user behavior has been documented within the literature, most prior work in the privacy field has focused on ‘one-size-fits-all’ interventions. Recent behavioral research has identified the potential of tailoring nudges to users by leveraging individual differences in decision making and personality. We present the results of three online experiments aimed at investigating whether nudges tailored to various psychometric scales can influence participants’ disclosure choices. Each study adopted a difference-in-differences design, testing whether differences in disclosure rates for participants presented with a nudge were affected by differences along various psychometric variables. Study 1 used a hypothetical disclosure scenario to measure participants’ responses to a single nudge. Study 2 and its replication (Study 3) tested responses in real disclosure scenarios to two nudges. Across all studies, we failed to find significant effects robustly linking any of the measured psychometric variables to differences in disclosure rates. We describe our study design and results along with a discussion of the practicality of using decision making and personality traits to tailor privacy nudges…(More)”.

Measuring human rights: facing a necessary challenge


Essay by Eduardo Burkle: “Given the abundance of data available today, many assume the world already has enough accurate metrics on human rights performance. However, the political sensitivity of human rights has proven a significant barrier to access. Governments often avoid producing and sharing this type of information.

States’ compliance with their human rights obligations often receives a lot of attention. But there is still much discussion about how to measure it. At the same time, statistics and data increasingly drive political and bureaucratic decisions. This, in turn, brings some urgency to the task of ensuring the best possible data are available.

Establishing cross-national human rights measures is vital for research, advocacy, and policymaking. It can also have a direct effect on people’s enjoyment of human rights. Good data allow states and actors to evaluate how well their country is performing. It also lets them make comparisons that highlight which policies and institutions are truly effective in promoting human rights.

Good human rights data does more than simply evaluate how well a country is performing – it also identifies which policies and institutions are truly effective in promoting human rights

Such context makes it crucial to arm researchers, journalists, advocates, practitioners, investors, and companies with reliable information when raising human rights issues in their countries, and around the world…(More)”.

Protecting Children in Cyberconflicts


Paper by Eleonore Pauwels: “Just as digital technologies have transformed myriad aspects of daily life, they are now transforming war, politics and the social fabric.

This rapid analysis examines the ways in which cyberconflict adversely affects children and offers actions that could strengthen safeguards to protect them.

Cyberconflict can impact children directly or indirectly. Harms range from direct targeting for influence and recruitment into armed forces and armed groups, to personal data manipulation and theft, to cyber attacks on infrastructure across sectors critical to child well-being such as education and health facilities.

Many experts believe that the combination of existing international humanitarian law, international criminal law, human rights law, and child rights law is adequate to address the emerging issues posed by cyberconflict. Nevertheless, several key challenges persist. Attribution of cyber attacks to specific actors and ensuring accountability has proven challenging, particularly in the so-called grey zone between war and peace.

There is an urgent need to clarify how child rights apply in the digital space and for Member States to place these rights at the centre of regulatory frameworks and legislation on new technologies…(More)”.