The Legal Singularity


Book by Abdi Aidid and Benjamin Alarie: “…argue that the proliferation of artificial intelligence–enabled technology – and specifically the advent of legal prediction – is on the verge of radically reconfiguring the law, our institutions, and our society for the better.

Revealing the ways in which our legal institutions underperform and are expensive to administer, the book highlights the negative social consequences associated with our legal status quo. Given the infirmities of the current state of the law and our legal institutions, the silver lining is that there is ample room for improvement. With concerted action, technology can help us to ameliorate the problems of the law and improve our legal institutions. Inspired in part by the concept of the “technological singularity,” The Legal Singularity presents a future state in which technology facilitates the functional “completeness” of law, where the law is at once extraordinarily more complex in its specification than it is today, and yet operationally, the law is vastly more knowable, fairer, and clearer for its subjects. Aidid and Alarie describe the changes that will culminate in the legal singularity and explore the implications for the law and its institutions…(More)”.

The Early History of Counting


Essay by Keith Houston: “Figuring out when humans began to count systematically, with purpose, is not easy. Our first real clues are a handful of curious, carved bones dating from the final few millennia of the three-​million-​year expanse of the Old Stone Age, or Paleolithic era. Those bones are humanity’s first pocket calculators: For the prehistoric humans who carved them, they were mathematical notebooks and counting aids rolled into one. For the anthropologists who unearthed them thousands of years later, they were proof that our ability to count had manifested itself no later than 40,000 years ago.

In 1973, while excavating a cave in the Lebombo Mountains, near South Africa’s border with Swaziland, Peter Beaumont found a small, broken bone with twenty-​nine notches carved across it. The so-​called Border Cave had been known to archaeologists since 1934, but the discovery during World War II of skeletal remains dating to the Middle Stone Age heralded a site of rare importance. It was not until Beaumont’s dig in the 1970s, however, that the cave gave up its most significant treasure: the earliest known tally stick, in the form of a notched, three-​inch long baboon fibula.

On the face of it, the numerical instrument known as the tally stick is exceedingly mundane. Used since before recorded history—​still used, in fact, by some cultures—​to mark the passing days, or to account for goods or monies given or received, most tally sticks are no more than wooden rods incised with notches along their length. They help their users to count, to remember, and to transfer ownership. All of which is reminiscent of writing, except that writing did not arrive until a scant 5,000 years ago—​and so, when the Lebombo bone was determined to be some 42,000 years old, it instantly became one of the most intriguing archaeological artifacts ever found. Not only does it put a date on when Homo sapiens started counting, it also marks the point at which we began to delegate our memories to external devices, thereby unburdening our minds so that they might be used for something else instead. Writing in 1776, the German historian Justus Möser knew nothing of the Lebombo bone, but his musings on tally sticks in general are strikingly apposite:

The notched tally stick itself testifies to the intelligence of our ancestors. No invention is simpler and yet more significant than this…(More)”.

An AI Model Tested In The Ukraine War Is Helping Assess Damage From The Hawaii Wildfires


Article by Irene Benedicto: “On August 7, 2023, the day before the Maui wildfires started in Hawaii, a constellation of earth-observing satellites took multiple pictures of the island at noon, local time. Everything was quiet, still. The next day, at the same, the same satellites captured images of fires consuming the island. Planet, a San Francisco-based company that owns the largest fleet of satellites taking pictures of the Earth daily, provided this raw imagery to Microsoft engineers, who used it to train an AI model designed to analyze the impact of disasters. Comparing before and after the fire photographs, the AI model created maps that highlighted the most devastated areas of the island.

With this information, the Red Cross rearranged its work on the field that same day to respond to the most urgent priorities first, helping evacuate thousands of people who’ve been affected by one of the deadliest fires in over a century. The Hawaii wildfires have already killed over a hundred people, a hundred more remain missing and at least 11,000 people have been displaced. The relief efforts are ongoing 10 days after the start of the fire, which burned over 3,200 acres. Hawaii Governor Josh Green estimated the recovery efforts could cost $6 billion.

Planet and Microsoft AI were able to pull and analyze the satellite imagery so quickly because they’d struggled to do so the last time they deployed their system: during the Ukraine war. The successful response in Maui is the result of a year and a half of building a new AI tool that corrected fundamental flaws in the previous system, which didn’t accurately recognize collapsed buildings in a background of concrete.

“When Ukraine happened, all the AI models failed miserably,” Juan Lavista, chief scientist at Microsoft AI, told Forbes.

The problem was that the company’s previous AI models were mainly trained with natural disasters in the U.S. and Africa. But devastation doesn’t look the same when it is caused by war and in an Eastern European city. “We learned that having one single model that would adapt to every single place on earth was likely impossible,” Lavista said…(More)”.

Tools of digital innovation in public affairs management: A practice-oriented analysis


Paper by Alberto Bitonti: “While the literature on digital transformation is growing in several fields, research on the effects of digital innovation in the practice of public affairs is still scattered and unsystematic, mostly focusing on interest groups’ social media strategies. However, digital innovation has begun to change the practice of public affairs management in many areas, especially in the form of datafication, AI analytics, and cloud-based knowledge management platforms. Growing possibilities in the use of data science and evidence-informed strategic decision-making have arisen in domains traditionally shaped by intuitions and non-codified professional experience. Based on desk research of case studies and hands-on analyses of three increasingly popular public affairs management software platforms (FiscalNote, Quorum, KMIND), this article develops a practice-oriented analysis of various digital tools and functionalities available to public affairs practitioners today, tackling a gap in the literature on how digital innovation can impact the management of several activities along the different phases of a public affairs campaign (monitoring and analysis, strategy design, action, assessment). The article thus highlights how digital innovation goes way beyond the sheer use of social media in communication activities, impacting the practice of public affairs on a deeper and more strategic level…(More)”.

What is the value of data? A review of empirical methods


Paper by Diane Coyle and Annabel Manley: “With the growing use of digital technologies, data have become core to many organizations’ decisions, with its value widely acknowledged across public and private sectors. Yet few comprehensive empirical approaches to establishing the value of data exist, and there is no consensus about which methods should be applied to specific data types or purposes. This paper examines a range of data valuation methodologies proposed in the existing literature. We propose a typology linking methods to different data types and purposes…(More)”.

Driving Excellence in Official Statistics: Unleashing the Potential of Comprehensive Digital Data Governance


Paper by Hossein Hassani and Steve McFeely: “With the ubiquitous use of digital technologies and the consequent data deluge, official statistics faces new challenges and opportunities. In this context, strengthening official statistics through effective data governance will be crucial to ensure reliability, quality, and access to data. This paper presents a comprehensive framework for digital data governance for official statistics, addressing key components, such as data collection and management, processing and analysis, data sharing and dissemination, as well as privacy and ethical considerations. The framework integrates principles of data governance into digital statistical processes, enabling statistical organizations to navigate the complexities of the digital environment. Drawing on case studies and best practices, the paper highlights successful implementations of digital data governance in official statistics. The paper concludes by discussing future trends and directions, including emerging technologies and opportunities for advancing digital data governance…(More)”.

The Urgent Need to Reimagine Data Consent


Article by Stefaan G. Verhulst, Laura Sandor & Julia Stamm: “Recognizing the significant benefits that can arise from the use and reuse of data to tackle contemporary challenges such as migration, it is worth exploring new approaches to collect and utilize data that empower individuals and communities, granting them the ability to determine how their data can be utilized for various personal, community, and societal causes. This need is not specific to migrants alone. It applies to various regions, populations, and fields, ranging from public health and education to urban mobility. There is a pressing demand to involve communities, often already vulnerable, to establish responsible access to their data that aligns with their expectations, while simultaneously serving the greater public good.

We believe the answer lies through a reimagination of the concept of consent. Traditionally, consent has been the tool of choice to secure agency and individual rights, but that concept, we would suggest, is no longer sufficient to today’s era of datafication. Instead, we should strive to establish a new standard of social license. Here, we’ll define what we mean by a social license and outline some of the limitations of consent (as it is typically defined and practiced today). Then we’ll describe one possible means of securing social license—through participatory decision -making…(More)”.

How data-savvy cities can tackle growing ethical considerations


Bloomberg Cities Network: “Technology for collecting, combining, and analyzing data is moving quickly, putting cities in a good position to use data to innovate in how they solve problems. However, it also places a responsibility on them to do so in a manner that does not undermine public trust. 

To help local governments deal with these issues, the London Office of Technology and Innovation, or LOTI, has a set of recommendations for data ethics capabilities in local government. One of those recommendations—for cities that are mature in their work in this area—is to hire a dedicated data ethicist.

LOTI exists to support dozens of local boroughs across London in their collective efforts to tackle big challenges. As part of that mission, LOTI hired Sam Nutt to serve as a data ethicist that local leaders can call on. The move reflected the reality that most local councils don’t have the capacity to have their own data ethicist on staff and it put LOTI in a position to experiment, learn, and share out lessons learned from the approach.

Nutt’s role provides a potential framework other cities looking to hire data ethicists can build on. His position is based on job specifications for data ethicists published by the UK government. He says his work falls into three general areas. First, he helps local councils work through ethical questions surrounding individual data projects. Second, he helps them develop more high-level policies, such as the Borough of Camden’s Data Charter. And third, he provides guidance on how to engage staff, residents, and stakeholders around the implications of using technology, including research on what’s new in the field. 

As an example of the kinds of ethical issues that he consults on, Nutt cites repairs in publicly subsidized housing. Local leaders are interested in using algorithms to help them prioritize use of scarce maintenance resources. But doing so raises questions about what criteria should be used to bump one resident’s needs above another’s. 

“If you prioritize, for example, the likelihood of a resident making a complaint, you may be baking in an existing social inequality, because some communities do not feel as empowered to make complaints as others,” Nutt says. “So it’s thinking through what the ethical considerations might be in terms of choices of data and how you use it, and giving advice to prevent potential biases from creeping in.” 

Nutt acknowledges that most cities are too resource constrained to hire a staff data ethicist. What matters most, he says, is that local governments create mechanisms for ensuring that ethical considerations of their choices with data and technology are considered. “The solution will never be that everyone has to hire a data ethicist,” Nutt says. “The solution is really to build ethics into your default ways of working with data.”

Stefaan Verhulst agrees. “The question for government is: Is ethics a position? A function? Or an institutional responsibility?” says Verhulst, Co-Founder of The GovLab and Director of its Data Program. The key is “to figure out how we institutionalize this in a meaningful way so that we can always check the pulse and get rapid input with regard to the social license for doing certain kinds of things.”

As the data capabilities of local governments grow, it’s also important to empower all individuals working in government to understand ethical considerations within the work they’re doing, and to have clear guidelines and codes of conduct they can follow. LOTI’s data ethics recommendations note that hiring a data ethicist should not be an organization’s first step, in part because “it risks delegating ethics to a single individual when it should be in the domain of anyone using or managing data.”

Training staff is a big part of the equation. “It’s about making the culture of government sensitive to these issues,” Verhulst says, so “that people are aware.”..(More)”.

Should Computers Decide How Much Things Cost?


Article by Colin Horgan: “In the summer of 2012, the Wall Street Journal reported that the travel booking website Orbitz had, in some cases, been suggesting to Apple users hotel rooms that cost more per night than those it was showing to Windows users. The company found that people who used Mac computers spent as much as 30 percent more a night on hotels. It was one of the first high-profile instances where the predictive capabilities of algorithms were shown to impact consumer-facing prices.

Since then, the pool of data available to corporations about each of us (the information we’ve either volunteered or that can be inferred from our web browsing and buying histories) has expanded significantly, helping companies build ever more precise purchaser profiles. Personalized pricing is now widespread, even if many consumers are only just realizing what it is. Recently, other algorithm-driven pricing models, like Uber’s surge or Ticketmaster’s dynamic pricing for concerts, have surprised users and fans. In the past few months, dynamic pricing—which is based on factors such as quantity—has pushed up prices of some concert tickets even before they hit the resale market, including for artists like Drake and Taylor Swift. And while personalized pricing is slightly different, these examples of computer-driven pricing have spawned headlines and social media posts that reflect a growing frustration with data’s role in how prices are dictated.

The marketplace is said to be a realm of assumed fairness, dictated by the rules of competition, an objective environment where one consumer is the same as any other. But this idea is being undermined by the same opaque and confusing programmatic data profiling that’s slowly encroaching on other parts of our lives—the algorithms. The Canadian government is currently considering new consumer-protection regulations, including what to do to control algorithm-based pricing. While strict market regulation is considered by some to be a political risk, another solution may exist—not at the point of sale but at the point where our data is gathered in the first place.

In theory, pricing algorithms aren’t necessarily bad…(More)”.

A Web of Our Own Making: The Nature of Digital Formation


Book by Antón Barba-Kay: “There no longer seems any point to criticizing the internet. We indulge in the latest doom-mongering about the evils of social media-on social media. We scroll through routine complaints about the deterioration of our attention spans. We resign ourselves to hating the internet even as we spend much of our waking lives with it. Yet our unthinking surrender to its effects-to the ways it recasts our aims and desires-is itself digital technology’s most powerful achievement. A Web of Our Own Making examines how online practices are reshaping our lives outside our notice. Barba-Kay argues that digital technology is a ‘natural technology’-a technology so intuitive as to conceal the extent to which it transforms our attention. He shows how and why this technology is reconfiguring knowledge, culture, politics, aesthetics, and theology. The digital revolution is primarily taking place not in Silicon Valley but within each of us…(More)”.