Citizen science is booming during the pandemic


Sigal Samuel at Vox: “…The pandemic has driven a huge increase in participation in citizen science, where people without specialized training collect data out in the world or perform simple analyses of data online to help out scientists.

Stuck at home with time on their hands, millions of amateurs arouennd the world are gathering information on everything from birds to plants to Covid-19 at the request of institutional researchers. And while quarantine is mostly a nightmare for us, it’s been a great accelerant for science.

Early in the pandemic, a firehose of data started gushing forth on citizen science platforms like Zooniverse and SciStarter, where scientists ask the public to analyze their data online.It’s a form of crowdsourcing that has the added bonus of giving volunteers a real sense of community; each project has a discussion forum where participants can pose questions to each other (and often to the scientists behind the projects) and forge friendly connections.

“There’s a wonderful project called Rainfall Rescue that’s transcribing historical weather records. It’s a climate change project to understand how weather has changed over the past few centuries,” Laura Trouille, vice president of citizen science at the Adler Planetarium in Chicago and co-lead of Zooniverse, told me. “They uploaded a dataset of 10,000 weather logs that needed transcribing — and that was completed in one day!”

Some Zooniverse projects, like Snapshot Safari, ask participants to classify animals in images from wildlife cameras. That project saw daily classifications go from 25,000 to 200,000 per day in the initial days of lockdown. And across all its projects, Zooniverse reported that 200,000 participants contributed more than 5 million classifications of images in one week alone — the equivalent of 48 years of research. Although participation has slowed a bit since the spring, it’s still four times what it was pre-pandemic.

Many people are particularly eager to help tackle Covid-19, and scientists have harnessed their energy. Carnegie Mellon University’s Roni Rosenfeld set up a platform where volunteers can help artificial intelligence predict the spread of the coronavirus, even if they know nothing about AI. Researchers at the University of Washington invited people to contribute to Covid-19 drug discovery using a computer game called Foldit; they experimented with designing proteins that could attach to the virus that causes Covid-19 and prevent it from entering cells….(More)”.

Towards intellectual freedom in an AI Ethics Global Community


Paper by Christoph Ebell et al: “The recent incidents involving Dr. Timnit Gebru, Dr. Margaret Mitchell, and Google have triggered an important discussion emblematic of issues arising from the practice of AI Ethics research. We offer this paper and its bibliography as a resource to the global community of AI Ethics Researchers who argue for the protection and freedom of this research community. Corporate, as well as academic research settings, involve responsibility, duties, dissent, and conflicts of interest. This article is meant to provide a reference point at the beginning of this decade regarding matters of consensus and disagreement on how to enact AI Ethics for the good of our institutions, society, and individuals. We have herein identified issues that arise at the intersection of information technology, socially encoded behaviors, and biases, and individual researchers’ work and responsibilities. We revisit some of the most pressing problems with AI decision-making and examine the difficult relationships between corporate interests and the early years of AI Ethics research. We propose several possible actions we can take collectively to support researchers throughout the field of AI Ethics, especially those from marginalized groups who may experience even more barriers in speaking out and having their research amplified. We promote the global community of AI Ethics researchers and the evolution of standards accepted in our profession guiding a technological future that makes life better for all….(More)”.

Leave No Migrant Behind: The 2030 Agenda and Data Disaggregation


Guide by the International Organization for Migration (IOM): “To date, disaggregation of global development data by migratory status remains low. Migrants are largely invisible in official SDG data. As the global community approaches 2030, very little is known about the impact of the 2030 Agenda on migrants. Despite a growing focus worldwide on data disaggregation, namely the breaking down of data into smaller sub-categories, there is a lack of practical guidance on the topic that can be tailored to address individual needs and capacities of countries.

Developed by IOM’s Global Migration Data Analysis Centre (GMDAC), the guide titled ‘Leave No Migrant Behind: The 2030 Agenda and Data Disaggregation‘ centres on nine SDGs focusing on hunger, education, and gender equality among others. The document is the first of its kind, in that it seeks to address a range of different categorization interests and needs related to international migrants and suggests practical steps that practitioners can tailor to best fit their context…The guide also highlights the key role disaggregation plays in understanding the many positive links between migration and the SDGs, highlighting migrants’ contributions to the 2030 Agenda.

The guide outlines key steps for actors to plan and implement initiatives by looking at sex, gender, age and disability, in addition to migratory status. These steps include undertaking awareness raising, identifying priority indicators, conducting data mapping, and more….Read more about the importance of data disaggregation for SDG indicators here….(More)”

What Is Mobility Data? Where Is It Used?


Brief by Andrew J. Zahuranec, Stefaan Verhulst, Andrew Young, Aditi Ramesh, and Brennan Lake: “Mobility data is data about the geographic location of a device passively produced through normal activity. Throughout the pandemic, public health experts and public officials have used mobility data to understand patterns of COVID-19’s spread and the impact of disease control measures. However, privacy advocates and others have questioned the need for this data and raised concerns about the capacity of such data-driven tools to facilitate surveillance, improper data use, and other exploitative practices.

In April, The GovLab, Cuebiq, and the Open Data Institute released The Use of Mobility Data for Responding to the COVID-19 Pandemic, which relied on several case studies to look at the opportunities, risks, and challenges associated with mobility data. Today, we hope to supplement that report with a new resource: a brief on what mobility data is and the different types of data it can include. The piece is a one-pager to allow decision-makers to easily read it. It provides real-world examples from the report to illustrate how different data types can be used in a responsible way…..(More)”.

Socially Responsible Data Labeling


Blog By Hamed Alemohammad at Radiant Earth Foundation: “Labeling satellite imagery is the process of applying tags to scenes to provide context or confirm information. These labeled training datasets form the basis for machine learning (ML) algorithms. The labeling undertaking (in many cases) requires humans to meticulously and manually assign captions to the data, allowing the model to learn patterns and estimate them for other observations.

For a wide range of Earth observation applications, training data labels can be generated by annotating satellite imagery. Images can be classified to identify the entire image as a class (e.g., water body) or for specific objects within the satellite image. However, annotation tasks can only identify features observable in the imagery. For example, with Sentinel-2 imagery at the 10-meter spatial resolution, one cannot detect the more detailed features of interest, such as crop types but would be able to distinguish large croplands from other land cover classes.

Human error in labeling is inevitable and results in uncertainties and errors in the final label. As a result, it’s best practice to examine images multiple times and then assign a majority or consensus label. In general, significant human resources and financial investment is needed to annotate imagery at large scales.

In 2018, we identified the need for a geographically diverse land cover classification training dataset that required human annotation and validation of labels. We proposed to Schmidt Futures a project to generate such a dataset to advance land cover classification globally. In this blog post, we discuss what we’ve learned developing LandCoverNet, including the keys to generating good quality labels in a socially responsible manner….(More)”.

How we mapped billions of trees in West Africa using satellites, supercomputers and AI


Martin Brandt and Kjeld Rasmussen in The Conversation: “The possibility that vegetation cover in semi-arid and arid areas was retreating has long been an issue of international concern. In the 1930s it was first theorized that the Sahara was expanding and woody vegetation was on the retreat. In the 1970s, spurred by the “Sahel drought”, focus was on the threat of “desertification”, caused by human overuse and/or climate change. In recent decades, the potential impact of climate change on the vegetation has been the main concern, along with the feedback of vegetation on the climate, associated with the role of the vegetation in the global carbon cycle.

Using high-resolution satellite data and machine-learning techniques at supercomputing facilities, we have now been able to map billions of individual trees and shrubs in West Africa. The goal is to better understand the real state of vegetation coverage and evolution in arid and semi-arid areas.

Finding a shrub in the desert – from space

Since the 1970s, satellite data have been used extensively to map and monitor vegetation in semi-arid areas worldwide. Images are available in “high” spatial resolution (with NASA’s satellites Landsat MSS and TM, and ESA’s satellites Spot and Sentinel) and “medium or low” spatial resolution (NOAA AVHRR and MODIS).

To accurately analyse vegetation cover at continental or global scale, it is necessary to use the highest-resolution images available – with a resolution of 1 metre or less – and up until now the costs of acquiring and analysing the data have been prohibitive. Consequently, most studies have relied on moderate- to low-resolution data. This has not allowed for the identification of individual trees, and therefore these studies only yield aggregate estimates of vegetation cover and productivity, mixing herbaceous and woody vegetation.

In a new study covering a large part of the semi-arid Sahara-Sahel-Sudanian zone of West Africa, published in Nature in October 2020, an international group of researchers was able to overcome these limitations. By combining an immense amount of high-resolution satellite data, advanced computing capacities, machine-learning techniques and extensive field data gathered over decades, we were able to identify individual trees and shrubs with a crown area of more than 3 m2 with great accuracy. The result is a database of 1.8 billion trees in the region studied, available to all interested….(More)”

Supercomputing, machine learning, satellite data and field assessments allow to map billions of individual trees in West Africa. Martin Brandt, Author provided

Regulating Personal Data : Data Models and Digital Services Trade


Report by Martina Francesca Ferracane and Erik van der Marel: “While regulations on personal data diverge widely between countries, it is nonetheless possible to identify three main models based on their distinctive features: one model based on open transfers and processing of data, a second model based on conditional transfers and processing, and third a model based on limited transfers and processing. These three data models have become a reference for many other countries when defining their rules on the cross-border transfer and domestic processing of personal data.

The study reviews their main characteristics and systematically identifies for 116 countries worldwide to which model they adhere for the two components of data regulation (i.e. cross-border transfers and domestic processing of data). In a second step, using gravity analysis, the study estimates whether countries sharing the same data model exhibit higher or lower digital services trade compared to countries with different regulatory data models. The results show that sharing the open data model for cross-border data transfers is positively associated with trade in digital services, while sharing the conditional model for domestic data processing is also positively correlated with trade in digital services. Country-pairs sharing the limited model, instead, exhibit a double whammy: they show negative trade correlations throughout the two components of data regulation. Robustness checks control for restrictions in digital services, the quality of digital infrastructure, as well as for the use of alternative data sources….(More)”.

The Locus Charter


Press Release: “A coalition of location data practitioners has developed an ethics charter to promote responsible use of location technology. The Locus Charter, facilitated by The Benchmark Initiative and EthicalGEO, is a proposed set of common international principles that can guide responsible practice when using location data, including through safeguarding privacy, protecting the vulnerable, and addressing any harmful impacts of bias in data.

The Benchmark Initiative and EthicalGEO are inviting individuals, businesses, and government agencies from around the world to join The Locus Charter community and help to shape equitable and sustainable practice around the use of location data. Member organisations include the American Geographical Society and Britain’s mapping agency, Ordnance Survey.

Location data is currently at the heart of the debate around digital privacy. Tech giants Apple and Facebook are in conflict over how much apps should be able to track users. Recent research shows personal information can be inferred from location data collected from smartphones, and that anonymisation can often be reversed to reveal people’s identities. The New York Times has unveiled a largely hidden trade in location data about individual people, collected from smartphones. As phones and other devices generate more detailed location data, these challenges grow…

The Locus Charter aims to restore public trust in location technology, in order to enable its transformative power to, improve public health, enhance our response to the Covid-19 pandemic, fight climate change, protect the environment and more….(More)”.

Averting Catastrophe


Book by Cass Sunstein on “Decision Theory for COVID-19, Climate Change, and Potential Disasters of All Kinds…The world is increasingly confronted with new challenges related to climate change, globalization, disease, and technology. Governments are faced with having to decide how much risk is worth taking, how much destruction and death can be tolerated, and how much money should be invested in the hopes of avoiding catastrophe. Lacking full information, should decision-makers focus on avoiding the most catastrophic outcomes? When should extreme measures be taken to prevent as much destruction as possible?

Averting Catastrophe explores how governments ought to make decisions in times of imminent disaster. Cass R. Sunstein argues that using the “maximin rule,” which calls for choosing the approach that eliminates the worst of the worst-case scenarios, may be necessary when public officials lack important information, and when the worst-case scenario is too disastrous to contemplate. He underscores this argument by emphasizing the reality of “Knightian uncertainty,” found in circumstances in which it is not possible to assign probabilities to various outcomes. Sunstein brings foundational issues in decision theory in close contact with real problems in regulation, law, and daily life, and considers other potential future risks. At once an approachable introduction to decision-theory and a provocative argument for how governments ought to handle risk, Averting Catastrophe offers a definitive path forward in a world rife with uncertainty….(More)”.

Democratic institutions and prosperity: The benefits of an open society


Paper by the European Parliamentary Research Service: “The ongoing structural transformation and the rapid spread of the technologies of the fourth industrial revolution are challenging current democratic institutions and their established forms of governance and regulation.At the same time, these changes offer vast opportunities to enhance, strengthen and expand the existing democratic framework to reflect a more complex and interdependent world. This process has already begun in many democratic societies but further progress is needed.
Examining these issues involves looking at the impact of ongoing complex and simultaneous changes on the theoretical framework underpinning beneficial democratic regulation. More specifically, combining economic, legal and political perspectives, it is necessary to explore how some adaptations to existing democratic institutions could further improve the functioning of democracies while also delivering additional economic benefits to citizens and society as whole. The introduction of a series of promising new tools could offer a potential way to support democratic decision-makers in regulating complexity and tackling ongoing and future challenges. The first of these tools is to use strategic foresight to anticipate and control future events; the second is collective intelligence, following the idea that citizens are collectively capable of providing better solutions to regulatory problems than are public administrations; the third and fourth are concerned with design-thinking and algorithmic regulation respectively. Design-based approaches are credited with opening up innovative options for policy-makers, while algorithms hold the promise of enabling decision-making to handle complex issues while remaining participatory….(More)”.