Work and meaning in the age of AI


Report by Daniel Susskind: “It is often said that work is not only a source of income but also of meaning. In this paper, I explore the theoretical and empirical literature that addresses this relationship between work and meaning. I show that the relationship is far less clear than is commonly supposed: There is a great heterogeneity in its nature, both among today’s workers and workers over time. I explain why this relationship matters for policymakers and economists concerned about the impact of technology on work. In the short term, it is important for predicting labour market outcomes of interest. It also matters for understanding how artificial intelligence (AI) affects not only the quantity of work but its quality as well: These new technologies may erode the meaning that people get from their work. In the medium term, if jobs are lost, this relationship also matters for designing bold policy interventions like the ‘Universal Basic Income’ and ‘Job Guarantee Schemes’: Their design, and any choice between them, is heavily dependent on policymakers’—often tacit—assumptions about the nature of this underlying relationship between work and meaning. For instance, policymakers must decide whether to simply focus on replacing lost income alone (as with a Universal Basic Income) or, if they believe that work is an important and non-substitutable source of meaning, on protecting jobs for that additional role as well (as with a Job Guarantee Scheme). In closing, I explore the challenge that the age of AI presents for an important feature of liberal political theory: the idea of ‘neutrality.’..(More)”

Ready, set, share: Researchers brace for new data-sharing rules


Jocelyn Kaiser and Jeffrey Brainard in Science: “…By 2025, new U.S. requirements for data sharing will extend beyond biomedical research to encompass researchers across all scientific disciplines who receive federal research funding. Some funders in the European Union and China have also enacted data-sharing requirements. The new U.S. moves are feeding hopes that a worldwide movement toward increased sharing is in the offing. Supporters think it could speed the pace and reliability of science.

Some scientists may only need to make a few adjustments to comply with the policies. That’s because data sharing is already common in fields such as protein crystallography and astronomy. But in other fields the task could be weighty, because sharing is often an afterthought. For example, a study involving 7750 medical research papers found that just 9% of those published from 2015 to 2020 promised to make their data publicly available, and authors of just 3% actually shared, says lead author Daniel Hamilton of the University of Melbourne, who described the finding at the International Congress on Peer Review and Scientific Publication in September 2022. Even when authors promise to share their data, they often fail to follow through. Out of 21,000 journal articles that included data-sharing plans, a study published in PLOS ONE in 2020 found, fewer than 21% provided links to the repository storing the data.

Journals and funders, too, have a mixed record when it comes to supporting data sharing. Research presented at the September 2022 peer-review congress found only about half of the 110 largest public, corporate, and philanthropic funders of health research around the world recommend or require grantees to share data…

“Health research is the field where the ethical obligation to share data is the highest,” says Aidan Tan, a clinician-researcher at the University of Sydney who led the study. “People volunteer in clinical trials and put themselves at risk to advance medical research and ultimately improve human health.”

Across many fields of science, researchers’ support for sharing data has increased during the past decade, surveys show. But given the potential cost and complexity, many are apprehensive about the NIH policy, and other requirements to follow. “How we get there is pretty messy right now,” says Parker Antin, a developmental biologist and associate vice president for research at the University of Arizona. “I’m really not sure whether the total return will justify the cost. But I don’t know of any other way to find out than trying to do it.”

Science offers this guide as researchers prepare to plunge in….(More)”.

The State of Open Data Policy Repository


The State of Open Data Policy Repository is a collection of recent policy developments surrounding open data, data reuse, and data collaboration around the world. 

A refinement of compilation of policies launched at the Open Data Policy Summit last year, the State of Open Data Policy Online Repository is an interactive resource that looks at recent legislation, directives, and proposals that affect open data and data collaboration all around the world. It captures what kinds of data collaboration issues policymakers are currently focused on and where the momentum for data innovation is heading in countries around the world.

Users can filter policies according to region, country, focus, and type of data sharing. The review currently surfaced approximately 60 examples of recent legislative acts, proposals, directives, and other policy documents, from which the Open Data Policy Lab draws findings about the need to promote more innovative policy frameworks.

This collection shows that, despite increased interest in the third wave conception of open data, policy development remains nascent. It is primarily concerned with open data repositories at the expense of alternative forms of collaboration. Most policies listed focus on releasing government data and, elsewhere, most nations still don’t have open data rules or a method to put the policies in place. 

This work reveals a pressing need for institutions to create frameworks that can direct data professionals since there are worries that inaction may both allow for misuse of data and lead to missed chances to use data…(More)”.

Computational Social Science for the Public Good: Towards a Taxonomy of Governance and Policy Challenges


Chapter by Stefaan G. Verhulst: “Computational Social Science (CSS) has grown exponentially as the process of datafication and computation has increased. This expansion, however, is yet to translate into effective actions to strengthen public good in the form of policy insights and interventions. This chapter presents 20 limiting factors in how data is accessed and analysed in the field of CSS. The challenges are grouped into the following six categories based on their area of direct impact: Data Ecosystem, Data Governance, Research Design, Computational Structures and Processes, the Scientific Ecosystem, and Societal Impact. Through this chapter, we seek to construct a taxonomy of CSS governance and policy challenges. By first identifying the problems, we can then move to effectively address them through research, funding, and governance agendas that drive stronger outcomes…(More)”. Full Book: Handbook of Computational Social Science for Policy

Automating Immigration and Asylum: The Uses of New Technologies in Migration and Asylum Governance in Europe


Report by Derya Ozkul: “The EU’s Artificial Intelligence Act proposal categorises AI uses for immigration, asylum and border as high risk, but new technologies are already used in many aspects of migration and asylum ‘management’ beyond imagination. To be able to reflect on the AI Act proposal, we first need to understand what current uses are, but this information is not always publicly available.

The new report by the Algorithmic Fairness for Asylum Seekers and Refugees (AFAR) project shows the multitude of uses of new technologies across Europe at the national and the EU levels. In particular, the report explores in detail the use of forecasting tools, risk assessment and triaging systems, processing of short- and long-term residency and citizenship applications, document verification, speech and dialect recognition, distribution of welfare benefits, matching tools, mobile phone data extraction and electronic monitoring, across Europe. It highlights the need for transparency and thorough training of decision-makers, as well as the inclusion of migrants’ interests in the design, decision, and implementation stages…(More)”.

The Health of Democracies During the Pandemic: Results from a Randomized Survey Experiment


Paper by Marcella Alsan et al: “Concerns have been raised about the “demise of democracy”, possibly accelerated by pandemic-related restrictions. Using a survey experiment involving 8,206 respondents from five Western democracies, we find that subjects randomly exposed to information regarding civil liberties infringements undertaken by China and South Korea to contain COVID-19 became less willing to sacrifice rights and more worried about their long-term-erosion. However, our treatment did not increase support for democratic procedures more generally, despite our prior evidence that pandemic-related health risks diminished such support. These results suggest that the start of the COVID-19 crisis was a particularly vulnerable time for democracies…(More)”.

Global Renewables Watch


About: “The Global Renewables Watch is a first-of-its-kind living atlas intended to map and measure all utility-scale solar and wind installations on Earth using artificial intelligence (AI) and satellite imagery, allowing users to evaluate clean energy transition progress and track trends over time. It also provides unique spatial data on land use trends to help achieve the dual aims of the environmental protection and increasing renewable energy capacity….(More)”

The Smartness Mandate


Book by Orit Halpern and Robert Mitchell: “Smart phones. Smart cars. Smart homes. Smart cities. The imperative to make our world ever smarter in the face of increasingly complex challenges raises several questions: What is this “smartness mandate”? How has it emerged, and what does it say about our evolving way of understanding—and managing—reality? How have we come to see the planet and its denizens first and foremost as data-collecting instruments?

In The Smartness Mandate, Orit Halpern and Robert Mitchell radically suggest that “smartness” is not primarily a technology, but rather an epistemology. Through this lens, they offer a critical exploration of the practices, technologies, and subjects that such an understanding relies upon—above all, artificial intelligence and machine learning. The authors approach these not simply as techniques for solving problems of calculations, but rather as modes of managing life (human and other) in terms of neo-Darwinian evolution, distributed intelligences, and “resilience,” all of which have serious implications for society, politics, and the environment.

The smartness mandate constitutes a new form of planetary governance, and Halpern and Mitchell aim to map the logic of this seemingly inexorable and now naturalized demand to compute, illuminate the genealogy of how we arrived here, and point to alternative imaginaries of the possibilities and potentials of smart technologies and infrastructures…(More)”.

Data Is What Data Does: Regulating Use, Harm, and Risk Instead of Sensitive Data


Paper by Daniel J. Solove: “Heightened protection for sensitive data is becoming quite trendy in privacy laws around the world. Originating in European Union (EU) data protection law and included in the EU’s General Data Protection Regulation (GDPR), sensitive data singles out certain categories of personal data for extra protection. Commonly recognized special categories of sensitive data include racial or ethnic origin, political opinions, religious or philosophical beliefs, trade union membership, health, sexual orientation and sex life, biometric data, and genetic data.

Although heightened protection for sensitive data appropriately recognizes that not all situations involving personal data should be protected uniformly, the sensitive data approach is a dead end. The sensitive data categories are arbitrary and lack any coherent theory for identifying them. The borderlines of many categories are so blurry that they are useless. Moreover, it is easy to use non-sensitive data as a proxy for certain types of sensitive data.

Personal data is akin to a grand tapestry, with different types of data interwoven to a degree that makes it impossible to separate out the strands. With Big Data and powerful machine learning algorithms, most non-sensitive data can give rise to inferences about sensitive data. In many privacy laws, data that can give rise to inferences about sensitive data is also protected as sensitive data. Arguably, then, nearly all personal data can be sensitive, and the sensitive data categories can swallow up everything. As a result, most organizations are currently processing a vast amount of data in violation of the laws.

This Article argues that the problems with the sensitive data approach make it unworkable and counterproductive — as well as expose a deeper flaw at the root of many privacy laws. These laws make a fundamental conceptual mistake — they embrace the idea that the nature of personal data is a sufficiently useful focal point for the law. But nothing meaningful for regulation can be determined solely by looking at the data itself. Data is what data does. Personal data is harmful when its use causes harm or creates a risk of harm. It is not harmful if it is not used in a way to cause harm or risk of harm.

To be effective, privacy law must focus on use, harm, and risk rather than on the nature of personal data. The implications of this point extend far beyond sensitive data provisions. In many elements of privacy laws, protections should be based on the use of personal data and proportionate to the harm and risk involved with those uses…(More)”.

The Doctor Who Wasn’t There: Technology, History, and the Limits of Telehealth


Book by Jeremy A. Greene: “The Doctor Who Wasn’t There traces the long arc of enthusiasm for—and skepticism of—electronic media in health and medicine. Over the past century, a series of new technologies promised to democratize access to healthcare. From the humble telephone to the connected smartphone, from FM radio to wireless wearables, from cable television to the “electronic brains” of networked mainframe computers: each new platform has promised a radical reformation of the healthcare landscape. With equal attention to the history of technology, the history of medicine, and the politics and economies of American healthcare, physician and historian Jeremy A. Greene explores the role that electronic media play, for better and for worse, in the past, present, and future of our health.

Today’s telehealth devices are far more sophisticated than the hook-and-ringer telephones of the 1920s, the radios that broadcasted health data in the 1940s, the closed-circuit televisions that enabled telemedicine in the 1950s, or the online systems that created electronic medical records in the 1960s. But the ethical, economic, and logistical concerns they raise are prefigured in the past, as are the gaps between what was promised and what was delivered. Each of these platforms also produced subtle transformations in health and healthcare that we have learned to forget, displaced by promises of ever newer forms of communication that took their place. 

Illuminating the social and technical contexts in which electronic medicine has been conceived and put into practice, Greene’s history shows the urgent stakes, then and now, for those who would seek in new media the means to build a more equitable future for American healthcare….(More)”.