Data-driven decisions: the case for randomised policy trials


Speech by Andrew Leigh: “…In 1747, 31-year-old Scottish naval surgeon James Lind set about determining the most effective treatment for scurvy, a disease that was killing thousands of sailors around the world. Selecting 12 sailors suffering from scurvy, Lind divided them into six pairs. Each pair received a different treatment: cider; sulfuric acid; vinegar; seawater; a concoction of nutmeg, garlic and mustard; and two oranges and a lemon. In less than a week, the pair who had received oranges and lemons were back on active duty, while the others languished. Given that sulphuric acid was the British Navy’s main treatment for scurvy, this was a crucial finding.

The trial provided robust evidence for the powers of citrus because it created a credible counterfactual. The sailors didn’t choose their treatments, nor were they assigned based on the severity of their ailment. Instead, they were randomly allocated, making it likely that difference in their recovery were due to the treatment rather than other characteristics.

Lind’s randomised trial, one of the first in history, has attained legendary status. Yet because 1747 was so long ago, it is easy to imagine that the methods he used are no longer applicable. After all, Lind’s research was conducted at a time before electricity, cars and trains, an era when slavery was rampant and education was reserved for the elite. Surely, some argue, ideas from such an age have been superseded today.

In place of randomised trials, some put their faith in ‘big data’. Between large-scale surveys and extensive administrative datasets, the world is awash in data as never before. Each day, hundreds of exabytes of data are produced. Big data has improved the accuracy of weather forecasts, permitted researchers to study social interactions across racial and ethnic lines, enabled the analysis of income mobility at a fine geographic scale and much more…(More)”

‘Evidence banks’ can drive better decisions in public life


Article by Anjana Ahuja: “Modern life is full of urgent questions to which governments should be seeking answers. Does working from home — WFH — damage productivity? Do LTNs (low-traffic neighbourhoods) cut air pollution? Are policy ideas that can be summed up in three-letter acronyms more palatable to the public than those, say Ulez, requiring four? 

That last one is tongue-in-cheek — but the point stands. While clinical trials can tell us reasonably confidently whether a drug or treatment works, a similar culture of evaluation is generally lacking for other types of intervention, such as crime prevention. Now research funders are stepping into the gap to build “evidence banks” or evidence syntheses: globally accessible one-stop shops for assessing the weight of evidence on a particular topic.

Last month, the Economic and Social Research Council, together with the Wellcome Trust, pledged a total of around £54mn to develop a database and tools that can collate and make sense of evidence in complex areas like climate change and healthy ageing. The announcement, Nature reports, was timed to coincide with the UN Summit of the Future, a conference in New York geared to improving the world for future generations.

Go-to repositories of good quality information that can feed the policy machine are essential. They normalise the role of robust evidence in public life. This matters: the policy pipeline has too often lacked due diligence. That can mean public money being squandered on ineffective wheezes, or worse.

Take Scared Straight, a crime prevention scheme originating in the US around 40 years ago and adopted in the UK. It was designed to keep teens on the straight and narrow by introducing them to prisoners. Those dalliances with delinquents were counterproductive. A review showed that children taking part were more likely to end up committing crimes than those who did not participate in the scheme…(More)”.

The Unaccountability Machine — why do big systems make bad decisions?


FT Review of book by Dan Davies: “The starting point of Davies’ entertaining, insightful book is that the uncontrolled proliferation of accountability sinks is one of the central drivers of what historian Adam Tooze calls the “polycrisis” of the 21st century. Their influence reaches far beyond frustrated customers endlessly on hold to “computer says no” service departments. In finance, banking crises regularly recur — yet few individual bankers are found at fault. If politicians’ promises flop, they complain they have no power; the Deep State is somehow to blame.

The origin of the problem, Davies argues, is the managerial revolution that began after the second world war, abetted by the advent of cheap computing power and the diffusion of algorithmic decision-making into every sphere of life. These systems have ended up “acting like a car’s crumple-zone to shield any individual manager from a disastrous decision”, he writes. While attractive from the individual’s perspective, they scramble the feedback on which society as a whole depends.

Yet the story, Davies continues, is not so simple. Seen from another perspective, accountability sinks are entirely reasonable responses to the ever-increasing complexity of modern economies. Standardisation and explicit policies and procedures offer the only feasible route to meritocratic recruitment, consistent service and efficient work. Relying on the personal discretion of middle managers would simply result in a different kind of mess…(More)”.

Scientists around the world call to protect research on one of humanity’s greatest short-term threats – Disinformation


Forum on Democracy and Information: “At a critical time for understanding digital communications’ impact on societies, research on disinformation is endangered. 

In August, researchers around the world bid farewell to CrowdTangle – the Meta-owned social media monitoring tool. The decision by Meta to close the number one platform used to track mis- and disinformation, in what is a major election year, only to present its alternative tool Meta Content Library and API, has been met with a barrage of criticism.

If, as suggested by the World Economic Forum’s 2024 global risk report, disinformation is one of the biggest short-term threats to humanity, our collective ability to understand how it spreads and impacts our society is crucial. Just as we would not impede scientific research into the spread of viruses and disease, nor into natural ecosystems or other historical and social sciences, disinformation research must be permitted to be carried out unimpeded and with access to information needed to understand its complexity. Understanding the political economy of disinformation as well as its technological dimensions is also a matter of public health, democratic resilience, and national security.

By directly affecting the research community’s ability to open social media black boxes, this radical decision will also, in turn, hamper public understanding of how technology affects democracy. Public interest scrutiny is also essential for the next era of technology, notably for the world’s largest AI systems, which are similarly proprietary and opaque. The research community is already calling on AI companies to learn from the mistakes of social media and guarantee protections for good faith research. The solution falls on multiple shoulders and the global scientific community, civil society, public institutions and philanthropies must come together to meaningfully foster and protect public interest research on information and democracy…(More)”.

Leveraging AI for Democracy: Civic Innovation on the New Digital Playing Field


Report by Beth Kerley, Carl Miller, and Fernanda Campagnucci: “Like social media before them, new AI tools promise to change the game when it comes to civic engagement. These technologies offer bold new possibilities for investigative journalists, anticorruption advocates, and others working with limited resources to advance democratic norms.

Yet the transformation wrought by AI advances is far from guaranteed to work in democracy’s favor. Potential threats to democracy from AI have drawn wide attention. To better the odds for prodemocratic actors in a fluid technological environment, systematic thinking about how to make AI work for democracy is needed.

The essays in this report outline possible paths toward a prodemocratic vision for AI. An overview essay by Beth Kerley based on insights from an International Forum for Democratic Studies expert workshop reflects on the critical questions that confront organizations seeking to deploy AI tools. Fernanda Campagnucci, spotlighting the work of Open Knowledge Brasil to open up government data, explores how AI advances are creating new opportunities for citizens to scrutinize public information. Finally, Demos’s Carl Miller sheds light on how AI technologies that enable new forms of civic deliberation might change the way we think about democratic participation itself…(More)“.

How to rebuild democracy to truly harness the power of the people


Article by Kyle Ellingson: “Many of us entered this so-called super-election year with a sense of foreboding. So far, not much has happened to allay those fears. Russia’s war on Ukraine is exacerbating a perception that democracy is threatened in Europe and beyond. In the US, Donald Trump, a presidential candidate with self-professed autocratic tendencies, has faced two assassination attempts. And more broadly, people seem to be losing faith in politics. “Most people from a diverse array of countries around the world lack confidence in the performance of their political institutions,” says a 2024 report by the International Institute for Democracy and Electoral Assistance.

On many objective measures, too, democracy isn’t functioning as it should. The systems we call democracies tend to favour the rich. Political violence is growing, as is legislative gridlock, and worldwide, elections are becoming less free and fair. Some 30 years after commentators crowed about the triumph of Western liberal democracy, their prediction seems further than ever from being realised. What happened?

According to Lex Paulson at the University Mohammed VI Polytechnic in Rabat, Morocco, we have lost sight of what democracy is. “We have made a terrible confusion between the system known as a republic – which relies on elections, parties and a permanent governing class – and the system known as a democracy, in which citizens directly participate in decisions and rotate power.” …(More)”.

China: Autocracy 2.0


Paper by David Y. Yang: “Autocracy 2.0, exemplified by modern China, is economically robust, technologically advanced, globally engaged, and controlled through subtle and sophisticated methods. What defines China’s political economy, and what drives Autocracy 2.0? What is its future direction? I start by discussing two key challenges autocracies face: incentives and information. I then describe Autocracy 1.0’s reliance on fear and repression to address these issues. It makes no credible promises, using coercion for compliance, resulting in a low-information environment. Next, I introduce Autocracy 2.0, highlighting its significant shift in handling commitment and information challenges. China uses economic incentives to align interests with regime survival, fostering support. It employs advanced bureaucratic structures and technology to manage incentives and information, enabling success in a high-information environment. Finally, I explore Autocracy 3.0’s potential. In China, forces might revert to Autocracy 1.0, using technology for state control as growth slows but aspirations stay high. Globally, modern autocracies, led by China, are becoming major geopolitical forces, challenging the liberal democratic order…(More)”.

Rethinking ‘Checks and Balances’ for the A.I. Age


Article by Steve Lohr: “A new project, orchestrated by Stanford University and published on Tuesday, is inspired by the Federalist Papers and contends that today is a broadly similar historical moment of economic and political upheaval that calls for a rethinking of society’s institutional arrangements.

In an introduction to its collection of 12 essays, called the Digitalist Papers, the editors overseeing the project, including Erik Brynjolfsson, director of the Stanford Digital Economy Lab, and Condoleezza Rice, secretary of state in the George W. Bush administration and director of the Hoover Institution, identify their overarching concern.

“A powerful new technology, artificial intelligence,” they write, “explodes onto the scene and threatens to transform, for better or worse, all legacy social institutions.”

The most common theme in the diverse collection of essays: Citizens need to be more involved in determining how to regulate and incorporate A.I. into their lives. “To build A.I. for the people, with the people,” as one essay summed it up.

The project is being published as the technology is racing ahead. A.I. enthusiasts see a future of higher economic growth, increased prosperity and a faster pace of scientific discovery. But the technology is also raising fears of a dystopian alternative — A.I. chatbots and automated software not only replacing millions of workers, but also generating limitless misinformation and worsening political polarization. How to govern and guide A.I. in the public interest remains an open question…(More)”.

China’s Hinterland Becomes A Critical Datascape


Article by Gary Zhexi Zhang: “In 2014, the southwestern province of Guizhou, a historically poor and mountainous area, beat out rival regions to become China’s first “Big Data Comprehensive Pilot Zone,” as part of a national directive to develop the region — which is otherwise best known as an exporter of tobacco, spirits and coal — into the infrastructural backbone of the country’s data industry. Since then, vast investment has poured into the province. Thousands of miles of highway and high-speed rail tunnel through the mountains. Driving through the province can feel vertiginous: Of the hundred highest bridges in the world, almost half are in Guizhou, and almost all were built in the last 15 years.

In 2015, Xi Jinping visited Gui’an New Area to inaugurate the province’s transformation into China’s “Big Data Valley,” exemplifying the central government’s goal to establish “high quality social and economic development,” ubiquitously advertised through socialist-style slogans plastered on highways and city streets…(More)”.

Artificial intelligence (AI) in action: A preliminary review of AI use for democracy support


Policy paper by Grahm Tuohy-Gaydos: “…provides a working definition of AI for Westminster Foundation for Democracy (WFD) and the broader democracy support sector. It then provides a preliminary review of how AI is being used to enhance democratic practices worldwide, focusing on several themes including: accountability and transparency, elections, environmental democracy, inclusion, openness and participation, and women’s political leadership. The paper also highlights potential risks and areas of development in the future. Finally, the paper shares five recommendations for WFD and democracy support organisations to consider advancing their ‘digital democracy’ agenda. This policy paper also offers additional information regarding AI classification and other resources for identifying good practice and innovative solutions. Its findings may be relevant to WFD staff members, international development practitioners, civil society organisations, and persons interested in using emerging technologies within governmental settings…(More)”.