We Need To Rewild The Internet


Article by Maria Farrell and Robin Berjon: “In the late 18th century, officials in Prussia and Saxony began to rearrange their complex, diverse forests into straight rows of single-species trees. Forests had been sources of food, grazing, shelter, medicine, bedding and more for the people who lived in and around them, but to the early modern state, they were simply a source of timber.

So-called “scientific forestry” was that century’s growth hacking. It made timber yields easier to count, predict and harvest, and meant owners no longer relied on skilled local foresters to manage forests. They were replaced with lower-skilled laborers following basic algorithmic instructions to keep the monocrop tidy, the understory bare.

Information and decision-making power now flowed straight to the top. Decades later when the first crop was felled, vast fortunes were made, tree by standardized tree. The clear-felled forests were replanted, with hopes of extending the boom. Readers of the American political anthropologist of anarchy and order, James C. Scott, know what happened next.

It was a disaster so bad that a new word, Waldsterben, or “forest death,” was minted to describe the result. All the same species and age, the trees were flattened in storms, ravaged by insects and disease — even the survivors were spindly and weak. Forests were now so tidy and bare, they were all but dead. The first magnificent bounty had not been the beginning of endless riches, but a one-off harvesting of millennia of soil wealth built up by biodiversity and symbiosis. Complexity was the goose that laid golden eggs, and she had been slaughtered…(More)”.

On the Manipulation of Information by Governments


Paper by Ariel Karlinsky and Moses Shayo: “Governmental information manipulation has been hard to measure and study systematically. We hand-collect data from official and unofficial sources in 134 countries to estimate misreporting of Covid mortality during 2020-21. We find that between 45%–55% of governments misreported the number of deaths. The lion’s share of misreporting cannot be attributed to a country’s capacity to accurately diagnose and report deaths. Contrary to some theoretical expectations, there is little evidence of governments exaggerating the severity of the pandemic. Misreporting is higher where governments face few social and institutional constraints, in countries holding elections, and in countries with a communist legacy…(More)”

Crowdsourcing for collaborative crisis communication: a systematic review


Paper by Maria Clara Pestana, Ailton Ribeiro and Vaninha Vieira: “Efficient crisis response and support during emergency scenarios rely on collaborative communication channels. Effective communication between operational centers, civilian responders, and public institutions is vital. Crowdsourcing fosters communication and collaboration among a diverse public. The primary objective is to explore the state-of-the-art in crowdsourcing for collaborative crisis communication guided by a systematic literature review. The study selected 20 relevant papers published in the last decade. The findings highlight solutions to facilitate rapid emergency responses, promote seamless coordination between stakeholders and the general public, and ensure data credibility through a rigorous validation process…(More)”.

The Formalization of Social Precarities


Anthology edited by Murali Shanmugavelan and Aiha Nguyen: “…explores platformization from the point of view of precarious gig workers in the Majority World. In countries like Bangladesh, Brazil, and India — which reinforce social hierarchies via gender, race, and caste — precarious workers are often the most marginalized members of society. Labor platforms made familiar promises to workers in these countries: work would be democratized, and people would have the opportunity to be their own boss. Yet even as platforms have upended the legal relationship between worker and employer, they have leaned into social structures to keep workers precarious — and in fact formalized those social precarities through surveillance and data collection…(More)”.

A Brief History of Automations That Were Actually People


Article by Brian Contreras: “If you’ve ever asked a chatbot a question and received nonsensical gibberish in reply, you already know that “artificial intelligence” isn’t always very intelligent.

And sometimes it isn’t all that artificial either. That’s one of the lessons from Amazon’s recent decision to dial back its much-ballyhooed “Just Walk Out” shopping technology, a seemingly science-fiction-esque software that actually functioned, in no small part, thanks to behind-the-scenes human labor.

This phenomenon is nicknamed “fauxtomation” because it “hides the human work and also falsely inflates the value of the ‘automated’ solution,” says Irina Raicu, director of the Internet Ethics program at Santa Clara University’s Markkula Center for Applied Ethics.

Take Just Walk Out: It promises a seamless retail experience in which customers at Amazon Fresh groceries or third-party stores can grab items from the shelf, get billed automatically and leave without ever needing to check out. But Amazon at one point had more than 1,000 workers in India who trained the Just Walk Out AI model—and manually reviewed some of its sales—according to an article published last year on the Information, a technology business website.

An anonymous source who’d worked on the Just Walk Out technology told the outlet that as many as 700 human reviews were needed for every 1,000 customer transactions. Amazon has disputed the Information’s characterization of its process. A company representative told Scientific American that while Amazon “can’t disclose numbers,” Just Walk Out has “far fewer” workers annotating shopping data than has been reported. In an April 17 blog post, Dilip Kumar, vice president of Amazon Web Services applications, wrote that “this is no different than any other AI system that places a high value on accuracy, where human reviewers are common.”…(More)”

The End of the Policy Analyst? Testing the Capability of Artificial Intelligence to Generate Plausible, Persuasive, and Useful Policy Analysis


Article by Mehrdad Safaei and Justin Longo: “Policy advising in government centers on the analysis of public problems and the developing of recommendations for dealing with them. In carrying out this work, policy analysts consult a variety of sources and work to synthesize that body of evidence into useful decision support documents commonly called briefing notes. Advances in natural language processing (NLP) have led to the continuing development of tools that can undertake a similar task. Given a brief prompt, a large language model (LLM) can synthesize information in content databases. This article documents the findings from an experiment that tested whether contemporary NLP technology is capable of producing public policy relevant briefing notes that expert evaluators judge to be useful. The research involved two stages. First, briefing notes were created using three models: NLP generated; human generated; and NLP generated/human edited. Next, two panels of retired senior public servants (with only one panel informed of the use of NLP in the experiment) were asked to judge the briefing notes using a heuristic evaluation rubric. The findings indicate that contemporary NLP tools were not able to, on their own, generate useful policy briefings. However, the feedback from the expert evaluators indicates that automatically generated briefing notes might serve as a useful supplement to the work of human policy analysts. And the speed with which the capabilities of NLP tools are developing, supplemented with access to a larger corpus of previously prepared policy briefings and other policy-relevant material, suggests that the quality of automatically generated briefings may improve significantly in the coming years. The article concludes with reflections on what such improvements might mean for the future practice of policy analysis…(More)”.

Unleashing collective intelligence for public decision-making: the Data for Policy community


Paper by Zeynep Engin, Emily Gardner, Andrew Hyde, Stefaan Verhulst and Jon Crowcroft: “Since its establishment in 2014, Data for Policy (https://dataforpolicy.org) has emerged as a prominent global community promoting interdisciplinary research and cross-sector collaborations in the realm of data-driven innovation for governance and policymaking. This report presents an overview of the community’s evolution from 2014 to 2023 and introduces its six-area framework, which provides a comprehensive mapping of the data for policy research landscape. The framework is based on extensive consultations with key stakeholders involved in the international committees of the annual Data for Policy conference series and the open-access journal Data & Policy published by Cambridge University Press. By presenting this inclusive framework, along with the guiding principles and future outlook for the community, this report serves as a vital foundation for continued research and innovation in the field of data for policy...(More)”.oeoMMrMrM..Andrew Hyde,Stefaan Verhulst[Opens in a new window] and

The AI That Could Heal a Divided Internet


Article by Billy Perrigo: “In the 1990s and early 2000s, technologists made the world a grand promise: new communications technologies would strengthen democracy, undermine authoritarianism, and lead to a new era of human flourishing. But today, few people would agree that the internet has lived up to that lofty goal. 

Today, on social media platforms, content tends to be ranked by how much engagement it receives. Over the last two decades politics, the media, and culture have all been reshaped to meet a single, overriding incentive: posts that provoke an emotional response often rise to the top.

Efforts to improve the health of online spaces have long focused on content moderation, the practice of detecting and removing bad content. Tech companies hired workers and built AI to identify hate speech, incitement to violence, and harassment. That worked imperfectly, but it stopped the worst toxicity from flooding our feeds. 

There was one problem: while these AIs helped remove the bad, they didn’t elevate the good. “Do you see an internet that is working, where we are having conversations that are healthy or productive?” asks Yasmin Green, the CEO of Google’s Jigsaw unit, which was founded in 2010 with a remit to address threats to open societies. “No. You see an internet that is driving us further and further apart.”

What if there were another way? 

Jigsaw believes it has found one. On Monday, the Google subsidiary revealed a new set of AI tools, or classifiers, that can score posts based on the likelihood that they contain good content: Is a post nuanced? Does it contain evidence-based reasoning? Does it share a personal story, or foster human compassion? By returning a numerical score (from 0 to 1) representing the likelihood of a post containing each of those virtues and others, these new AI tools could allow the designers of online spaces to rank posts in a new way. Instead of posts that receive the most likes or comments rising to the top, platforms could—in an effort to foster a better community—choose to put the most nuanced comments, or the most compassionate ones, first…(More)”.

Mass Data Sharing in Smart Cities


Report by Berenika Drazewska and Mark Findlay: “There are at least two ways of understanding the importance of this Report and its implications. The essential research purpose was to examine the nature of mass data sharing between private and public agencies in the commerce and administration of certain smart cities. With this knowledge the research speculated on and selectively exposed the governance challenges posed by this sharing for stakeholders, citizen/residents in particular, in various data relationships and arrangements. Predicting that good data governance policy and practices can address these challenges, the Report proposes a model strategy that grows from commitments where stakeholders will employ trusted data spaces to create respectful and responsible data relationships, where the benefits of data sharing can also be achieved without compromising any stakeholder interests…(More)”.

Millions of gamers advance biomedical research


Article by McGill: “…4.5 million gamers around the world have advanced medical science by helping to reconstruct microbial evolutionary histories using a minigame included inside the critically and commercially successful video game, Borderlands 3. Their playing has led to a significantly refined estimate of the relationships of microbes in the human gut. The results of this collaboration will both substantially advance our knowledge of the microbiome and improve on the AI programs that will be used to carry out this work in future.

By playing Borderlands Science, a mini-game within the looter-shooter video game Borderlands 3, these players have helped trace the evolutionary relationships of more than a million different kinds of bacteria that live in the human gut, some of which play a crucial role in our health. This information represents an exponential increase in what we have discovered about the microbiome up till now. By aligning rows of tiles which represent the genetic building blocks of different microbes, humans have been able to take on tasks that even the best existing computer algorithms have been unable to solve yet…(More) (and More)”.