AI and Big Data: Disruptive Regulation


Book by Mark Findlay, Josephine Seah, and Willow Wong: “This provocative and timely book identifies and disrupts the conventional regulation and governance discourses concerning AI and big data. It suggests that, instead of being used as tools for exclusionist commercial markets, AI and big data can be employed in governing digital transformation for social good.

Analysing the ways in which global technology companies have colonised data access, the book reveals how trust, ethics, and digital self-determination can be reconsidered and engaged to promote the interests of marginalised stakeholders in data arrangement. Chapters examine the regulation of labour engagement in digital economies, the landscape of AI ethics, and a multitude of questions regarding participation, costs, and sustainability. Presenting several informative case studies, the book challenges some of the accepted qualifiers of frontier tech and data use and proposes innovative ways of actioning the more conventional regulatory components of big data.

Scholars and students in information and media law, regulation and governance, and law and politics will find this book to be critical reading. It will also be of interest to policymakers and the AI and data science community…(More)”.

We, the Data


Book by Wendy H. Wong: “Our data-intensive world is here to stay, but does that come at the cost of our humanity in terms of autonomy, community, dignity, and equality? In We, the Data, Wendy H. Wong argues that we cannot allow that to happen. Exploring the pervasiveness of data collection and tracking, Wong reminds us that we are all stakeholders in this digital world, who are currently being left out of the most pressing conversations around technology, ethics, and policy. This book clarifies the nature of datafication and calls for an extension of human rights to recognize how data complicate what it means to safeguard and encourage human potential.

As we go about our lives, we are co-creating data through what we do. We must embrace that these data are a part of who we are, Wong explains, even as current policies do not yet reflect the extent to which human experiences have changed. This means we are more than mere “subjects” or “sources” of data “by-products” that can be harvested and used by technology companies and governments. By exploring data rights, facial recognition technology, our posthumous rights, and our need for a right to data literacy, Wong has crafted a compelling case for engaging as stakeholders to hold data collectors accountable. Just as the Universal Declaration of Human Rights laid the global groundwork for human rights, We, the Data gives us a foundation upon which we claim human rights in the age of data…(More)”.

Extremely Online: The Untold Story of Fame, Influence, and Power on the Internet


Book by Taylor Lorenz: “For over a decade, Taylor Lorenz has been the authority on internet culture, documenting its far-reaching effects on all corners of our lives. Her reporting is serious yet entertaining and illuminates deep truths about ourselves and the lives we create online. In her debut book, Extremely Online, she reveals how online influence came to upend the world, demolishing traditional barriers and creating whole new sectors of the economy. Lorenz shows this phenomenon to be one of the most disruptive changes in modern capitalism.

By tracing how the internet has changed what we want and how we go about getting it, Lorenz unearths how social platforms’ power users radically altered our expectations of content, connection, purchasing, and power. Lorenz documents how moms who started blogging were among the first to monetize their personal brands online, how bored teens who began posting selfie videos reinvented fame as we know it, and how young creators on TikTok are leveraging opportunities to opt out of the traditional career pipeline. It’s the real social history of the internet.

Emerging seemingly out of nowhere, these shifts in how we use the internet seem easy to dismiss as fads. However, these social and economic transformations have resulted in a digital dynamic so unappreciated and insurgent that it ultimately created new approaches to work, entertainment, fame, and ambition in the 21st century…(More)”.

Five types of urban digital twins


Blog by Darrel Ronald: “The definition for urban digital twins is too vague — so it is important to create a clearer picture of the types of urban digital twins that are available. Not all digital twins are the same and each one comes with features and capabilities, strengths and weakness, as well as appropriate and inappropriate use cases….

Darrel Ronald
Urban Twin taxonomy, Source: Darrel Ronald, Spatiomatics

As shown in my proposed Urban Digital Twin Taxonomy above, I propose that we classify these products first based on their Main Functionality (the Use Case), then based on their Technology Platform. I highlight some of main products within the different categories and their product scope. Next, I detail the different types of twins and offer some brief strengths and weaknesses for each type. This taxonomy could apply to other industries such as architecture or manufacturing, but it is specifically applied to cities and urban development projects.

The main functionalities can be grouped by:

  • Modelling Twin
  • Computational Twin
  • Scenario Twin
  • Operational Twin
  • Experiential Twin

The technology platforms can be grouped by:

  • Computer Aided Design (CAD)
  • Web GIS
  • Geographic Information System (GIS)
  • Gaming…(More)”.

Data Repurposing through Compatibility: A Computational Perspective


Paper by Asia Biega: “Reuse of data in new contexts beyond the purposes for which it was originally collected has contributed to technological innovation and reducing the consent burden on data subjects. One of the legal mechanisms that makes such reuse possible is purpose compatibility assessment. In this paper, I offer an in-depth analysis of this mechanism through a computational lens. I moreover consider what should qualify as repurposing apart from using data for a completely new task, and argue that typical purpose formulations are an impediment to meaningful repurposing. Overall, the paper positions compatibility assessment as a constructive practice beyond an ineffective standard…(More)”

From Print to Pixels: The Changing Landscape of the Public Sphere in the Digital Age


Paper by Taha Yasseri: “This Mini Review explores the evolution of the public sphere in the digital age. The public sphere is a social space where individuals come together to exchange opinions, discuss public affairs, and engage in collective decision-making. It is considered a defining feature of modern democratic societies, allowing citizens to participate in public life and promoting transparency and accountability in the political process. This Mini Review discusses the changes and challenges faced by the public sphere in recent years, particularly with the advent of new communication technologies such as the Internet and social media. We highlight benefits such as a) increase in political participation, b) facilitation of collective action, c) real time spread of information, and d) democratization of information exchange; and harms such as a) increasing polarization of public discourse, b) the spread of misinformation, and c) the manipulation of public opinion by state and non-state actors. The discussion will conclude with an assessment of the digital age public sphere in established democracies like the US and the UK…(More)”.

Machine-assisted mixed methods: augmenting humanities and social sciences with artificial intelligence


Paper by Andres Karjus: “The increasing capacities of large language models (LLMs) present an unprecedented opportunity to scale up data analytics in the humanities and social sciences, augmenting and automating qualitative analytic tasks previously typically allocated to human labor. This contribution proposes a systematic mixed methods framework to harness qualitative analytic expertise, machine scalability, and rigorous quantification, with attention to transparency and replicability. 16 machine-assisted case studies are showcased as proof of concept. Tasks include linguistic and discourse analysis, lexical semantic change detection, interview analysis, historical event cause inference and text mining, detection of political stance, text and idea reuse, genre composition in literature and film; social network inference, automated lexicography, missing metadata augmentation, and multimodal visual cultural analytics. In contrast to the focus on English in the emerging LLM applicability literature, many examples here deal with scenarios involving smaller languages and historical texts prone to digitization distortions. In all but the most difficult tasks requiring expert knowledge, generative LLMs can demonstrably serve as viable research instruments. LLM (and human) annotations may contain errors and variation, but the agreement rate can and should be accounted for in subsequent statistical modeling; a bootstrapping approach is discussed. The replications among the case studies illustrate how tasks previously requiring potentially months of team effort and complex computational pipelines, can now be accomplished by an LLM-assisted scholar in a fraction of the time. Importantly, this approach is not intended to replace, but to augment researcher knowledge and skills. With these opportunities in sight, qualitative expertise and the ability to pose insightful questions have arguably never been more critical…(More)”.

Get a rabbit: Don’t trust the numbers · 


Article by John Lanchester: “At a dinner​ with the American ambassador in 2007, Li Keqiang, future premier of China, said that when he wanted to know what was happening to the country’s economy, he looked at the numbers for electricity use, rail cargo and bank lending. There was no point using the official GDP statistics, Li said, because they are ‘man-made’. That remark, which we know about thanks to WikiLeaks, is fascinating for two reasons. First, because it shows a sly, subtle, worldly humour – a rare glimpse of the sort of thing Chinese Communist Party leaders say in private. Second, because it’s true. A whole strand in contemporary thinking about the production of knowledge is summed up there: data and statistics, all of them, are man-made.

They are also central to modern politics and governance, and the ways we talk about them. That in itself represents a shift. Discussions that were once about values and beliefs – about what a society wants to see when it looks at itself in the mirror – have increasingly turned to arguments about numbers, data, statistics. It is a quirk of history that the politician who introduced this style of debate wasn’t Harold Wilson, the only prime minister to have had extensive training in statistics, but Margaret Thatcher, who thought in terms of values but argued in terms of numbers. Even debates that are ultimately about national identity, such as the referendums about Scottish independence and EU membership, now turn on numbers.

Given the ubiquity of this style of argument, we are nowhere near as attentive to its misuses as we should be. As the House of Commons Treasury Committee said dryly in a 2016 report on the economic debate about EU membership, ‘many of these claims sound factual because they use numbers.’ The best short book about the use and misuse of statistics is Darrell Huff’s How to Lie with Statistics, first published in 1954, a devil’s-advocate guide to the multiple ways in which numbers are misused in advertising, commerce and politics. (Single best tip: ‘up to’ is always a fib. It means somebody did a range of tests and has artfully chosen the most flattering number.) For all its virtues, though, even Huff’s book doesn’t encompass the full range of possibilities for statistical deception. In politics, the numbers in question aren’t just man-made but are often contentious, tendentious or outright fake.

Two fake numbers have been decisively influential in British politics over the baleful last thirteen years. The first was an outright lie: Vote Leave’s assertion that £350 million a week extra ‘for the NHS’ would be available if we left the EU. The real number for the UK’s net contribution to the EU was £110 million, but that didn’t matter, since the crucial thing for the Leave campaign was to make the number the focus of debate. The Treasury Committee said the number was fake, and so did the UK Statistics Authority. This had no, or perhaps even a negative, effect. In politics it doesn’t really matter what the numbers are, so much as whose they are. If people are arguing about your numbers, you’re winning…(More)“.

On the culture of open access: the Sci-hub paradox


Paper by Abdelghani Maddi and David Sapinho: “Shadow libraries, also known as ”pirate libraries”, are online collections of copyrighted publications that have been made available for free without the permission of the copyright holders. They have gradually become key players of scientific knowledge dissemination, despite their illegality in most countries of the world. Many publishers and scientist-editors decry such libraries for their copyright infringement and loss of publication usage information, while some scholars and institutions support them, sometimes in a roundabout way, for their role in reducing inequalities of access to knowledge, particularly in low-income countries. Although there is a wealth of literature on shadow libraries, none of this have focused on its potential role in knowledge dissemination, through the open access movement. Here we analyze how shadow libraries can affect researchers’ citation practices, highlighting some counter-intuitive findings about their impact on the Open Access Citation Advantage (OACA). Based on a large randomized sample, this study first shows that OA publications, including those in fully OA journals, receive more citations than their subscription-based counterparts do. However, the OACA has slightly decreased over the seven last years. The introduction of a distinction between those accessible or not via the Scihub platform among subscription-based suggest that the generalization of its use cancels the positive effect of OA publishing. The results show that publications in fully OA journals are victims of the success of Sci-hub. Thus, paradoxically, although Sci-hub may seem to facilitate access to scientific knowledge, it negatively affects the OA movement as a whole, by reducing the comparative advantage of OA publications in terms of visibility for researchers. The democratization of the use of Sci-hub may therefore lead to a vicious cycle, hindering efforts to develop full OA strategies without proposing a credible and sustainable alternative model for the dissemination of scientific knowledge…(More)”.

The Forgotten “Emerging” Technology


Report by Michael Garcia: “The widespread deployment of 5G devices in the United States will spur widespread use of augmented reality, virtual reality, and mixed reality applications—collectively known as extended reality. The over-commercialization of the term “metaverse” has impeded honest conversations about the implications of an insecure metaverse and the technologies associated with it. While these applications and devices will bring significant benefits, they will be accompanied by numerous cybersecurity challenges. As a result, U.S. policymakers run afoul of repeating past mistakes: failing to secure technology before it ushers in a new era of national security concerns. The U.S. government must work closely with industry, academia, nonprofits, and international partners to begin thinking about these consequential issues…(More)”.