Looking after and using data for public benefit


Heather Savory at the Office for National Statistics (UK): “Official Statistics are for the benefit of society and the economy and help Britain to make better decisions. They allow the formulation of better public policy and the effective measurement of those policies. They inform the direction of economic and commercial activities. They provide valuable information for analysts, researchers, public and voluntary bodies. They enable the public to hold organisations that spend public money to account, thus informing democratic debate.

The ability to harness the power of data is critical in enabling official statistics to support the most important decisions facing the country.

Under the new powers in the Digital Economy Act , ONS can now gain access to new and different sources of data including ‘administrative’ data from government departments and commercial data. Alongside the availability of these new data sources ONS is experiencing a strong demand for ad hoc insights alongside our traditional statistics.

We need to deliver more, faster, finer-grained insights into the economy and society. We need to deliver high quality, trustworthy information, on a faster timescale, to help decision-making. We will increasingly develop innovative data analysis methods, for example using images to gain insight from the work we’ve recently announced on Urban Forests….

I should explain here that our data is not held in one big linked database; we’re architecting our Data Access Platform so that data can be linked in different ways for different purposes. This is designed to preserve data confidentiality, so only the necessary subset of data is accessible by authorised people, for a certain purpose. To avoid compromising their effectiveness, we do not make public the specific details of the security measures we have in place, but our recently tightened security regime, which is independently assured by trusted external bodies, includes:

  • physical measures to restrict who can access places where data is stored;
  • protective measures for all data-related IT services;
  • measures to restrict who can access systems and data held by ONS;
  • controls to guard against staff or contractors misusing their legitimate access to data; including vetting to an appropriate level for the sensitivity of data to which they might have access.

One of the things I love about working in the public sector is that our work can be shared openly.

We live in a rapidly changing and developing digital world and we will continue to monitor and assess the data standards and security measures in place to ensure they remain strong and effective. So, as well as sharing this work openly to reassure all our data suppliers that we’re taking good care of their data, we’re also seeking feedback on our revised data policies.

The same data can provide different insights when viewed through different lenses or in different combinations. The more data is shared – with the appropriate safeguards of course – the more it has to give.

If you work with data, you’ll know that collaborating with others in this space is key and that we need to be able to share data more easily when it makes sense to do so. So, the second reason for sharing this work openly is that, if you’re in the technical space, we’d value your feedback on our approach and if you’re in the data space and would like to adopt the same approach, we’d love to support you with that – so that we can all share data more easily in the future….(More)

ONS’s revised policies on the use, management and security of data can befound here.

Inside the world’s ‘what works’ teams


Jen Gold at What Works Blog: “There’s a small but growing band of government teams around the world dedicated to making experiments happen. The Cabinet Office’s What Works Team, set up in 2013, was the first of its kind. But you’ll now find them in Canada, the US, Finland, Australia, Colombia, and the UAE.

All of these teams work across government to champion the testing and evaluation of new approaches to public service delivery. This blog takes a look at the many ways in which we’re striving to make experimentation the norm in our governments.

Unsurprisingly we’re all operating in very different contexts. Some teams were set up in response to central requirements for greater experimentation. Take Canada, for instance. In 2016 the Treasury Board directed departments and agencies to devote a fixed proportion of programme funds to “experimenting with new approaches” (building on Prime Minister Trudeau’s earlier instruction to Ministers). An Innovation and Experimentation Team was then set up in the Treasury Board to provide some central support.

Finland’s Experimentation Office, based in the Prime Minister’s Office, is in a similar position. The team supports the delivery of Prime Minister Juha Sipilä’s 2016 national action plan that calls for “a culture of experimentation” in public services and a series of flagship policy experiments.

Others, like the US Office of Evaluation Sciences (OES) and the Behavioural Economics Team of the Australian Government (BETA), grew out of political interest in using behavioural science experiments in public policy. But these teams now run experiments in a much broader set of areas.

What unites us is a focus on helping public servants generate and use new evidence in policy decisions and service delivery….(More)”.

Participation 2.0? Crowdsourcing Participatory Development @ DFID


Paper by Anke Schwittay, Paul Braund: “Through an empirical analysis of Amplify, a crowdsourcing platform funded by the UK’s Department for International Development (DFID), we examine the potential of ICTs to afford more participatory development. Especially interactive Web2.0 technologies are often assumed to enable the participation of marginalized groups in their development, through allowing them to modify content and generate their own communication. 

We use the concepts of platform politics and voice to show that while Amplify managers and designers invested time and resources to include the voices of Amplify beneficiaries on the platform and elicit their feedback on projects supported via the platform, no meaningful participation took place. Our analysis of the gaps between participatory rhetoric, policy and practice concludes with suggestions for how ICTs could be harnessed to contribute to meaningful participatory development that matters materially and politically.,,,(More)”

All of Us Research Program Expands Data Collection Efforts with Fitbit


NIH Press Release: “The All of Us Research Program has launched the Fitbit Bring-Your-Own-Device (BYOD) project. Now, in addition to providing health information through surveys, electronic health records, and biosamples, participants can choose to share data from their Fitbit accounts to help researchers make discoveries. The project is a key step for the program in integrating digital health technologies for data collection.

Digital health technologies, like mobile apps and wearable devices, can gather data outside of a hospital or clinic. This data includes information about physical activity, sleep, weight, heart rate, nutrition, and water intake, which can give researchers a more complete picture of participants’ health. The All of Us Research Program is now gathering this data in addition to surveys, electronic health record information, physical measurements, and blood and urine samples, working to make the All of Us resource one of the largest and most diverse data sets of its kind for health research.

“Collecting real-world, real-time data through digital technologies will become a fundamental part of the program,” said Eric Dishman, director of the All of Us Research Program. “This information, in combination with many other data types, will give us an unprecedented ability to better understand the impact of lifestyle and environment on health outcomes and, ultimately, develop better strategies for keeping people healthy in a very precise, individualized way.”…

All of Us is developing additional plans to incorporate digital health technologies. A second project with Fitbit is expected to launch later in the year. It will include providing devices to a limited number of All of Us participants who will be randomly invited to take part, to enable them to share wearable data with the program. And All of Us will add connections to other devices and apps in the future to further expand data collection efforts and engage participants in new ways….(More)”.

The Future of Civic Engagement


Report by Hollie Russon Gilman: “The 2018 mid-term voter turnout was the highest in 50 years. While vital, voting can’t sustain civic engagement in the long term. So, how do we channel near-term activism into long-term civic engagement?  In her essay, Gilman paints a picture of how new institutional structures, enabled by new technologies, could lead to a new “civic layer” in society that results in “a more responsive, participatory, collaborative, and adaptive future for civic engagement in governance decision making.”

Creating a New “Civic Layer.” The longer-term future presents an opportunity to set up institutionalized structures for engagement across local, state, and federal levels of government—creating a “civic layer.” Its precise form will evolve, but the basic concept is to establish a centralized interface within a com- munity to engage residents in governance decision making that interweaves digital and in-person engagement. People will earn “civic points” for engagement across a variety of activities—including every time they sign a petition, report a pot hole, or volunteer in their local community.

While creating a civic layer will require new institutional approaches, emerging technologies such as the Internet of Things (IoT), artificial intelligence (AI), and distributed ledger (e.g., blockchain) will also play a critical enabling role. These technologies will allow new institutional models to expand the concept of citizen coproduction of services in building a more responsive, connected, and engaged citizenry.

The following examples show different collaborative governance and technology components that will comprise the civic layer.  Each could be expanded and become interwoven into the fabric of civic life.

Use Collaborative Policymaking Models to Build a Civic Layer.  While we currently think of elections as a primary mode of citizen engagement with government, in the medium- to long-range future we could see collaborative policy models that become the de facto way people engage to supplement elections. Several of these engagement models are on the local level. However, with the formation of a civic layer these forms of engagement could become integrated into a federated structure enabling more scale, scope, and impact. Following are two promising models.

  • Participatory Budgeting can be broadly defined as the participation of citizens in the decision-making process of how to allocate their community’s budget among different priorities and in the monitoring of public spending. The process first came to the United States in 2009 through the work of the nonprofit Participatory Budgeting Project. Unlike traditional budget consultations held by some governments—which often amount to “selective listening” exercises—with participatory budgeting, citizens have an actual say in how a portion of a government’s investment budget is spent, with more money often allocated to poorer communities. Experts estimate that up to 2,500 local governments around the world have implemented participatory budgeting,
  • Citizens’Jury is another promising collaborative policymaking engagement model, pioneered in the 1980s and currently advocated by the nonprofit Jefferson Center in Minnesota. Three counties in rural Minnesota use this method as a foundation for Rural Climate Dialogues—regular gatherings where local residents hear from rural experts, work directly with their neighbors to design actionable community and policy recommendations, and share their feedback with public officials at a statewide meeting of rural Minnesota citizens, state agency representatives, and nonprofit organizations….(More)”.

The democratic potential of civic applications


Paper by Jäske, Maija and Ertiö, Titiana: “Recently, digital democratic applications have increased in presence and scope. This study clarifies how civic applications – bottom-up technologies that use open data to solve governance and policy challenges – can contribute to democratic governance. While civic applications claim to deepen democracy, systematic frameworks for assessing the democratic potential of civic apps are missing, because apps are often evaluated against technical criteria. This study introduces a framework for evaluating the democratic potential of civic apps, distinguishing six criteria: inclusiveness, deliberation, influence, publicity, mobilization, and knowledge production. The framework is applied to a case study of the Finnish DataDemo competition in 2014 by analyzing the institutional design features of six civic applications. It is argued that in terms of democratic governance, the greatest potential of civic apps lies in enhancing publicity and mobilization, while they should not be expected to increase inclusiveness or direct influence in decisions. Thus, our study contributes to understanding how civic applications can improve democracy in times of open data abundance….(More)”.

Democracy From Above? The Unfulfilled Promise of Nationally Mandated Participatory Reforms


Book by Stephanie L. McNulty: “People are increasingly unhappy with their governments in democracies around the world. In countries as diverse as India, Ecuador, and Uganda, governments are responding to frustrations by mandating greater citizen participation at the local and state level. Officials embrace participatory reforms, believing that citizen councils and committees lead to improved accountability and more informed communities. Yet there’s been little research on the efficacy of these efforts to improve democracy, despite an explosion in their popularity since the mid-1980s.Democracy from Above? tests the hypothesis that top-down reforms strengthen democracies and evaluates the conditions that affect their success.

Stephanie L. McNulty addresses the global context of participatory reforms in developing nations. She observes and interprets what happens after greater citizen involvement is mandated in seventeen countries, with close case studies of Guatemala, Bolivia, and Peru. The first cross-national comparison on this issue,Democracy from Above? explores whether the reforms effectively redress the persistent problems of discrimination, elite capture, clientelism, and corruption in the countries that adopt them. As officials and reformers around the world and at every level of government look to strengthen citizen involvement and confidence in the political process, McNulty provides a clear understanding of the possibilities and limitations of nationally mandated participatory reforms…(More)”.

Societal costs and benefits of high-value open government data: a case study in the Netherlands


Paper by F.M. Welle Donker and B. van Loenen: “Much research has emphasised the benefits of open government data, and especially high-value data. The G8 Open Data Charter defines high-value data as data that improve democracy and encourage the innovative reuse of the particular data. Thus, governments worldwide invest resources to identify potential high-value datasets and to publish these data as open data. However, while the benefits of open data are well researched, the costs of publishing data as open data are less researched. This research examines the relationship between the costs of making data suitable for publication as (linked) open data and the societal benefits thereof. A case study of five high-value datasets was carried out in the Netherlands to provide a societal cost-benefit analysis of open high-value data. Different options were investigated, ranging from not publishing the dataset at all to publishing the dataset as linked open data.

In general, it can be concluded that the societal benefits of (linked) open data are higher than the costs. The case studies show that there are differences between the datasets. In many cases, costs for open data are an integral part of general data management costs and hardly lead to additional costs. In certain cases, however, the costs to anonymize /aggregate the data are high compared to the potential value of an open data version of the dataset. Although, for these datasets, this leads to a less favourable relationship between costs and benefits, the societal benefits would still be higher than without an open data version….(More)”.

Index: Open Data


By Alexandra Shaw, Michelle Winowatan, Andrew Young, and Stefaan Verhulst

The Living Library Index – inspired by the Harper’s Index – provides important statistics and highlights global trends in governance innovation. This installment focuses on open data and was originally published in 2018.

Value and Impact

  • The projected year at which all 28+ EU member countries will have a fully operating open data portal: 2020

  • Between 2016 and 2020, the market size of open data in Europe is expected to increase by 36.9%, and reach this value by 2020: EUR 75.7 billion

Public Views on and Use of Open Government Data

  • Number of Americans who do not trust the federal government or social media sites to protect their data: Approximately 50%

  • Key findings from The Economist Intelligence Unit report on Open Government Data Demand:

    • Percentage of respondents who say the key reason why governments open up their data is to create greater trust between the government and citizens: 70%

    • Percentage of respondents who say OGD plays an important role in improving lives of citizens: 78%

    • Percentage of respondents who say OGD helps with daily decision making especially for transportation, education, environment: 53%

    • Percentage of respondents who cite lack of awareness about OGD and its potential use and benefits as the greatest barrier to usage: 50%

    • Percentage of respondents who say they lack access to usable and relevant data: 31%

    • Percentage of respondents who think they don’t have sufficient technical skills to use open government data: 25%

    • Percentage of respondents who feel the number of OGD apps available is insufficient, indicating an opportunity for app developers: 20%

    • Percentage of respondents who say OGD has the potential to generate economic value and new business opportunity: 61%

    • Percentage of respondents who say they don’t trust governments to keep data safe, protected, and anonymized: 19%

Efforts and Involvement

  • Time that’s passed since open government advocates convened to create a set of principles for open government data – the instance that started the open data government movement: 10 years

  • Countries participating in the Open Government Partnership today: 79 OGP participating countries and 20 subnational governments

  • Percentage of “open data readiness” in Europe according to European Data Portal: 72%

    • Open data readiness consists of four indicators which are presence of policy, national coordination, licensing norms, and use of data.

  • Number of U.S. cities with Open Data portals: 27

  • Number of governments who have adopted the International Open Data Charter: 62

  • Number of non-state organizations endorsing the International Open Data Charter: 57

  • Number of countries analyzed by the Open Data Index: 94

  • Number of Latin American countries that do not have open data portals as of 2017: 4 total – Belize, Guatemala, Honduras and Nicaragua

  • Number of cities participating in the Open Data Census: 39

Demand for Open Data

  • Open data demand measured by frequency of open government data use according to The Economist Intelligence Unit report:

    • Australia

      • Monthly: 15% of respondents

      • Quarterly: 22% of respondents

      • Annually: 10% of respondents

    • Finland

      • Monthly: 28% of respondents

      • Quarterly: 18% of respondents

      • Annually: 20% of respondents

    •  France

      • Monthly: 27% of respondents

      • Quarterly: 17% of respondents

      • Annually: 19% of respondents

        •  
    • India

      • Monthly: 29% of respondents

      • Quarterly: 20% of respondents

      • Annually: 10% of respondents

    • Singapore

      • Monthly: 28% of respondents

      • Quarterly: 15% of respondents

      • Annually: 17% of respondents 

    • UK

      • Monthly: 23% of respondents

      • Quarterly: 21% of respondents

      • Annually: 15% of respondents

    • US

      • Monthly: 16% of respondents

      • Quarterly: 15% of respondents

      • Annually: 20% of respondents

  • Number of FOIA requests received in the US for fiscal year 2017: 818,271

  • Number of FOIA request processed in the US for fiscal year 2017: 823,222

  • Distribution of FOIA requests in 2017 among top 5 agencies with highest number of request:

    • DHS: 45%

    • DOJ: 10%

    • NARA: 7%

    • DOD: 7%

    • HHS: 4%

Examining Datasets

  • Country with highest index score according to ODB Leaders Edition: Canada (76 out of 100)

  • Country with lowest index score according to ODB Leaders Edition: Sierra Leone (22 out of 100)

  • Number of datasets open in the top 30 governments according to ODB Leaders Edition: Fewer than 1 in 5

  • Average percentage of datasets that are open in the top 30 open data governments according to ODB Leaders Edition: 19%

  • Average percentage of datasets that are open in the top 30 open data governments according to ODB Leaders Edition by sector/subject:

    • Budget: 30%

    • Companies: 13%

    • Contracts: 27%

    • Crime: 17%

    • Education: 13%

    • Elections: 17%

    • Environment: 20%

    • Health: 17%

    • Land: 7%

    • Legislation: 13%

    • Maps: 20%

    • Spending: 13%

    • Statistics: 27%

    • Trade: 23%

    • Transport: 30%

  • Percentage of countries that release data on government spending according to ODB Leaders Edition: 13%

  • Percentage of government data that is updated at regular intervals according to ODB Leaders Edition: 74%

  • Number of datasets available through:

  • Number of datasets classed as “open” in 94 places worldwide analyzed by the Open Data Index: 11%

  • Percentage of open datasets in the Caribbean, according to Open Data Census: 7%

  • Number of companies whose data is available through OpenCorporates: 158,589,950

City Open Data

  • New York City

  • Singapore

    • Number of datasets published in Singapore: 1,480

    • Percentage of datasets with standardized format: 35%

    • Percentage of datasets made as raw as possible: 25%

  • Barcelona

    • Number of datasets published in Barcelona: 443

    • Open data demand in Barcelona measured by:

      • Number of unique sessions in the month of September 2018: 5,401

    • Quality of datasets published in Barcelona according to Tim Berners Lee 5-star Open Data: 3 stars

  • London

    • Number of datasets published in London: 762

    • Number of data requests since October 2014: 325

  • Bandung

    • Number of datasets published in Bandung: 1,417

  • Buenos Aires

    • Number of datasets published in Buenos Aires: 216

  • Dubai

    • Number of datasets published in Dubai: 267

  • Melbourne

    • Number of datasets published in Melbourne: 199

Sources

  • About OGP, Open Government Partnership. 2018.  

Can a set of equations keep U.S. census data private?


Jeffrey Mervis at Science: “The U.S. Census Bureau is making waves among social scientists with what it calls a “sea change” in how it plans to safeguard the confidentiality of data it releases from the decennial census.

The agency announced in September 2018 that it will apply a mathematical concept called differential privacy to its release of 2020 census data after conducting experiments that suggest current approaches can’t assure confidentiality. But critics of the new policy believe the Census Bureau is moving too quickly to fix a system that isn’t broken. They also fear the changes will degrade the quality of the information used by thousands of researchers, businesses, and government agencies.

The move has implications that extend far beyond the research community. Proponents of differential privacy say a fierce, ongoing legal battle over plans to add a citizenship question to the 2020 census has only underscored the need to assure people that the government will protect their privacy....

Differential privacy, first described in 2006, isn’t a substitute for swapping and other ways to perturb the data. Rather, it allows someone—in this case, the Census Bureau—to measure the likelihood that enough information will “leak” from a public data set to open the door to reconstruction.

“Any time you release a statistic, you’re leaking something,” explains Jerry Reiter, a professor of statistics at Duke University in Durham, North Carolina, who has worked on differential privacy as a consultant with the Census Bureau. “The only way to absolutely ensure confidentiality is to release no data. So the question is, how much risk is OK? Differential privacy allows you to put a boundary” on that risk....

In the case of census data, however, the agency has already decided what information it will release, and the number of queries is unlimited. So its challenge is to calculate how much the data must be perturbed to prevent reconstruction....

A professor of labor economics at Cornell University, Abowd first learned that traditional procedures to limit disclosure were vulnerable—and that algorithms existed to quantify the risk—at a 2005 conference on privacy attended mainly by cryptographers and computer scientists. “We were speaking different languages, and there was no Rosetta Stone,” he says.

He took on the challenge of finding common ground. In 2008, building on a long relationship with the Census Bureau, he and a team at Cornell created the first application of differential privacy to a census product. It is a web-based tool, called OnTheMap, that shows where people work and live….

The three-step process required substantial computing power. First, the researchers reconstructed records for individuals—say, a 55-year-old Hispanic woman—by mining the aggregated census tables. Then, they tried to match the reconstructed individuals to even more detailed census block records (that still lacked names or addresses); they found “putative matches” about half the time.

Finally, they compared the putative matches to commercially available credit databases in hopes of attaching a name to a particular record. Even if they could, however, the team didn’t know whether they had actually found the right person.

Abowd won’t say what proportion of the putative matches appeared to be correct. (He says a forthcoming paper will contain the ratio, which he calls “the amount of uncertainty an attacker would have once they claim to have reidentified a person from the public data.”) Although one of Abowd’s recent papers notes that “the risk of re-identification is small,” he believes the experiment proved reidentification “can be done.” And that, he says, “is a strong motivation for moving to differential privacy.”…

Such arguments haven’t convinced Ruggles and other social scientists opposed to applying differential privacy on the 2020 census. They are circulating manuscripts that question the significance of the census reconstruction exercise and that call on the agency to delay and change its plan....

Ruggles, meanwhile, has spent a lot of time thinking about the kinds of problems differential privacy might create. His Minnesota institute, for instance, disseminates data from the Census Bureau and 105 other national statistical agencies to 176,000 users. And he fears differential privacy will put a serious crimp in that flow of information…

There are also questions of capacity and accessibility. The centers require users to do all their work onsite, so researchers would have to travel, and the centers offer fewer than 300 workstations in total....

Abowd has said, “The deployment of differential privacy within the Census Bureau marks a sea change for the way that official statistics are produced and published.” And Ruggles agrees. But he says the agency hasn’t done enough to equip researchers with the maps and tools needed to navigate the uncharted waters….(More)”.