Principles and Policies for “Data Free Flow With Trust”


Paper by Nigel Cory, Robert D. Atkinson, and Daniel Castro: “Just as there was a set of institutions, agreements, and principles that emerged out of Bretton Woods in the aftermath of World War II to manage global economic issues, the countries that value the role of an open, competitive, and rules-based global digital economy need to come together to enact new global rules and norms to manage a key driver of today’s global economy: data. Japanese Prime Minister Abe’s new initiative for “data free flow with trust,” combined with Japan’s hosting of the G20 and leading role in e-commerce negotiations at the World Trade Organization (WTO), provides a valuable opportunity for many of the world’s leading digital economies (Australia, the United States, and European Union, among others) to rectify the gradual drift toward a fragmented and less-productive global digital economy. Prime Minister Abe is right in proclaiming, “We have yet to catch up with the new reality, in which data drives everything, where the D.F.F.T., the Data Free Flow with Trust, should top the agenda in our new economy,” and right in his call “to rebuild trust toward the system for international trade. That should be a system that is fair, transparent, and effective in protecting IP and also in such areas as e-commerce.”

The central premise of this effort should be a recognition that data and data-driven innovation are a force for good. Across society, data innovation—the use of data to create value—is creating more productive and innovative economies, transparent and responsive governments, better social outcomes (improved health care, safer and smarter cities, etc.).3But to maximize the innovative and productivity benefits of data, countries that support an open, rules-based global trading system need to agree on core principles and enact common rules. The benefits of a rules-based and competitive global digital economy are at risk as a diverse range of countries in various stages of political and economic development have policy regimes that undermine core processes, especially the flow of data and its associated legal responsibilities; the use of encryption to protect data and digital activities and technologies; and the blocking of data constituting illegal, pirated content….(More)”.

Citizen, Science, and Citizen Science


Introduction by Shun-Ling and Chen Fa-ti Fan to special issue on citizen science: “The term citizen science has become very popular among scholars as well as the general public, and, given its growing presence in East Asia, it is perhaps not a moment too soon to have a special issue of EASTS on the topic. However, the quick expansion of citizen science, as a notion and a practice, has also spawned a mass of blurred meanings. The term is ill-defined and has been used in diverse ways. To avoid confusion, it is necessary to categorize the various and often ambiguous usages of the term and clarify their meanings.

As in any taxonomy, there are as many typologies as the particular perspectives, parameters, and criteria adopted for classification. There have been helpful attempts at classifying different modes of citizen science (Cooper and Lewenstein 2016Wiggins and Crowston 2012Haklay 2012). However, they focused primarily on the different approaches or methods in citizen science. Ottinger’s two categories of citizen science—“scientific authority driven” and “social movement based”—foreground the criteria of action and justification, but they unnecessarily juxtapose science and society; in any case, they may be too general and leaving out too much at the same time.1

In contrast, our classification will emphasize the different conceptions of citizen and citizenship in how we think about citizen science. We believe that this move can help us contextualize the ideas and practices of citizen science in the diverse socio-political conditions found in East Asia and beyond (Leach, Scoones, and Wynne 2005). To explain that point, we’ll begin with a few observations. First, the current discourse on citizen science tends to glide over such concepts as state, citizen, and the public and to assume that the reader will understand what they mean. This confidence originates in part from the fact that the default political framework of the discourse is usually Western (particularly Anglo-American). As a result, one often easily accepts a commonsense notion of participatory liberal democracy as the reference framework. However, one cannot assume that that is the de facto political framework for discussion of citizen science….(More)”.

A Symphony, Not a Solo: How Collective Management Organisations Can Embrace Innovation and Drive Data Sharing in the Music Industry


Paper by David Osimo, Laia Pujol Priego, Turo Pekari and Ano Sirppiniemi: “…data is becoming a fundamental source of competitive advantage in music, just as in other sectors, and streaming services in particular are generating large volume of new data offering unique insight around customer taste and behavior. (As Financial Times recently put it, the music
industry is having its “moneyball” moment) But how are the different players getting ready for this change?

This policy brief aims to look at the question from the perspective of CMOs, the organisations charged with redistributing royalties from music users to music rightsholders (such as musical authors and publishers).

The paper is divided in three sections. Part I will look at the current positioning of CMOs in this new data-intensive ecosystem. Part II will discuss how greater data sharing and reuse can maximize innovation, comparing the music industries with other industries. Part III will make policy and business-model reform recommendations for CMOs to stimulate data-driven innovation, internally and in the industry as a whole….(More)”

Data Stewardship on the map: A study of tasks and roles in Dutch research institutes


Report by Verheul, Ingeborg et al: “Good research requires good data stewardship. Data stewardship encompasses all the different tasks and responsibilities that relate to caring for data during the various phases of the whole research life cycle. The basic assumption is that the researcher himself/herself is primarily responsible for all data.

However, the researcher does need professional support to achieve this. To that end, diverse supportive data stewardship roles and functions have evolved in recent years. Often they have developed over the course of time.

Their functional implementation depends largely on their place in the organization. This comes as no surprise when one considers that data stewardship consists of many facets that are traditionally assigned to different departments. Researchers regularly take on data stewardship tasks as well, not only for themselves but also in a wider context for a research group. This data stewardship work often remains unnoticed….(More)”.

Data to the rescue


Podcast by Kenneth Cukier: “Access to the right data can be as valuable in humanitarian crises as water or medical care, but it can also be dangerous. Misused or in the wrong hands, the same information can put already vulnerable people at further risk. Kenneth Cukier hosts this special edition of Babbage examining how humanitarian organisations use data and what they can learn from the profit-making tech industry. This episode was recorded live from Wilton Park, in collaboration with the United Nations OCHA Centre for Humanitarian Data…(More)”.

Facebook releases a trio of maps to aid with fighting disease outbreaks


Sarah Perez at Techcrunch: “Facebook… announced a new initiative focused on using its data and technologies to help nonprofit organizations and universities working in public health better map the spread of infectious diseases around the world. Specifically, the company is introducing three new maps: population density maps with demographic estimates, movement maps and network coverage maps. These, says Facebook, will help the health partners to understand where people live, how they’re moving and if they have connectivity — all factors that can aid in determining how to respond to outbreaks, and where supplies should be delivered.

As Facebook explained, health organizations rely on information like this when planning public health campaigns. But much of the information they rely on is outdated, like older census data. In addition, information from more remote communities can be scarce.

By combining the new maps with other public health data, Facebook believes organizations will be better equipped to address epidemics.

The new high-resolution population density maps will estimate the number of people living within 30-meter grid tiles, and provide insights on demographics, including the number of children under five, the number of women of reproductive age, as well as young and elderly populations. These maps aren’t built using Facebook data, but are instead built by using Facebook’s AI capabilities with satellite imagery and census information.

Movement maps, meanwhile, track aggregate data about Facebook users’ movements via their mobile phones (when location services are enabled). At scale, health partners can combine this with other data to predict where other outbreaks may occur next….(More)”.

Data Protection and Digital Agency for Refugees


Paper by Dragana Kaurin: “For the millions of refugees fleeing conflict and persecution every year, access to information about their rights and control over their personal data are crucial for their ability to assess risk and navigate the asylum process. While asylum seekers are required to provide significant amounts of personal information on their journey to safety, they are rarely fully informed of their data rights by UN agencies or local border control and law enforcement staff tasked with obtaining and processing their personal information. Despite recent improvements in data protection mechanisms in the European Union, refugees’ informed consent for the collection and use of their personal data is rarely sought. Using examples drawn from interviews with refugees who have arrived in Europe since 2013, and an analysis of the impacts of the 2016 EU-Turkey deal on migration, this paper analyzes how the vast amount of data collected from refugees is gathered, stored and shared today, and considers the additional risks this collection process poses to an already vulnerable population navigating a perilous information-decision gap….(More)”.

Humans and Big Data: New Hope? Harnessing the Power of Person-Centred Data Analytics


Paper by Carmel Martin, Keith Stockman and Joachim P. Sturmberg: “Big data provide the hope of major health innovation and improvement. However, there is a risk of precision medicine based on predictive biometrics and service metrics overwhelming anticipatory human centered sense-making, in the fuzzy emergence of personalized (big data) medicine. This is a pressing issue, given the paucity of individual sense-making data approaches. A human-centric model is described to address the gap in personal particulars and experiences in individual health journeys. The Patient Journey Record System (PaJR) was developed to improve human-centric healthcare by harnessing the power of person-centred data analytics using complexity theory, iterative health services and information systems applications over a 10 year period. PaJR is a web-based service supporting usually bi-weekly telephone calls by care guides to individuals at risk of readmissions.

This chapter describes a case study of the timing and context of readmissions using human (biopsychosocial) particular data which is based on individual experiences and perceptions with differing patterns of instability. This Australian study, called MonashWatch, is a service pilot using the PaJR system in the Dandenong Hospital urban catchment area of the Monash Health network. State public hospital big data – the Victorian HealthLinks Chronic Care algorithm provides case finding for high risk of readmission based on disease and service metrics. Monash Watch was actively monitoring 272 of 376 intervention patients, with 195 controls over 22 months (ongoing) at the time of the study.

Three randomly selected intervention cases describe a dynamic interplay of self-reported change in health and health care, medication, drug and alcohol use, social support structure. While the three cases were at similar predicted risk initially, their cases represented different statistically different time series configurations and admission patterns. Fluctuations in admission were associated with (mal)alignment of bodily health with psychosocial and environmental influences. However human interpretation was required to make sense of the patterns as presented by the multiple levels of data.

A human-centric model and framework for health journey monitoring illustrates the potential for ‘small’ personal experience data to inform clinical care in the era of big data predominantly based on biometrics and medical industrial process. ….(More)”.

Open data could have helped us learn from another mining dam disaster


Paulo A. de Souza Jr. at Nature: “The recent Brumadinho dam disaster in Brazil is an example of infrastructure failure with catastrophic consequences. Over 300 people were reported dead or missing, and nearly 400 more were rescued alive. The environmental impact is massive and difficult to quantify. The frequency of these disasters demonstrates that the current assets for monitoring integrity and generating alerting managers, authorities and the public to ongoing change in tailings are, in many cases, not working as they should. There is also the need for adequate prevention procedures. Monitoring can be perfect, but without timely and appropriate action, it will be useless. Good management therefore requires quality data. Undisputedly, management practices of industrial sites, including audit procedures, must improve, and data and metadata available from preceding accidents should be better used. There is a rich literature available about design, construction, operation, maintenance and decommissioning of tailing facilities. These include guidelines, standards, case studies, technical reports, consultancy and audit practices, and scientific papers. Regulation varies from country to country and in some cases, like Australia and Canada, it is controlled by individual state agencies. There are, however, few datasets available that are shared with the technical and scientific community more globally; particularly for prior incidents. Conspicuously lacking are comprehensive data related to monitoring of large infrastructures such as mining dams.

Today, Scientific Data published a Data Descriptor presenting a dataset obtained from 54 laboratory experiments on the breaching of fluvial dikes because of flow overtopping. (Re)use of such data can help improve our understanding of fundamental processes underpinning industrial infrastructure collapse (e.g., fluvial dike breaching, mining dam failure), and assess the accuracy of numerical models for the prediction of such incidents. This is absolutely essential for better management of floods, mitigation of dam collapses, and similar accidents. The authors propose a framework that could exemplify how data involving similar infrastructure can be stored, shared, published, and reused…(More)”.

Data Collaboration for the Common Good: Enabling Trust and Innovation Through Public-Private Partnerships


World Economic Forum Report: “As the digital technologies of the Fourth Industrial Revolution continue to drive change throughout all sectors of the global economy, a unique moment exists to create a more inclusive, innovative and resilient society. Central to this change is the use of data. It is abundantly available but if improperly used will be the source of dangerous and unwelcome results.

When data is shared, linked and combined across sectoral and institutional boundaries, a multiplier effect occurs. Connecting one bit with another unlocks new insights and understandings that often weren’t anticipated. Yet, due to commercial limits and liabilities, the full value of data is often unrealized. This is particularly true when it comes to using data for the common good. While public-private data collaborations represent an unprecedented opportunity to address some of the world’s most urgent and complex challenges, they have generally been small and limited in impact. An entangled set of legal, technical, social, ethical and commercial risks have created an environment where the incentives for innovation have stalled. Additionally, the widening lack of trust among individuals and institutions creates even more uncertainty. After nearly a decade of anticipation on the promise of public-private data collaboration – with relatively few examples of success at global scale – a pivotal moment has arrived to encourage progress and move forward….(More)”

(See also http://datacollaboratives.org/).