How Integrated Data Can Support COVID-19 Crisis and Recovery


Blog by Actionable Intelligence for Social Policy (AISP): “…State and local leaders are called upon to respond to the immediate harms of COVID-19. Yet with a looming recession threatening to undo gains among marginalized groups — particularly the Black middle class — tools to understand and disrupt long-term impacts on economic mobility and well-being are also urgently needed.

Administrative data[3] — the information collected during the course of routine service delivery, program administration, and business operations — provide an essential tool to help policymakers, community leaders, and researchers understand short- and long-term impacts of the pandemic. Several jurisdictions now have the capacity to link administrative data across programs in order to better understand how individuals interact with multiple systems, study longitudinal outcomes, and identify vulnerable subpopulations. As the COVID-19 crisis reveals weaknesses in the U.S. social safety net, states and localities with integrated administrative data infrastructure can use their capacity to identify populations and needs otherwise overlooked. Youth who “age out” of the child welfare system or individuals experiencing chronic homelessness often remain invisible when using traditional methods, aggregate data, or administrative records from a single source.

This blogpost demonstrates how nimble state and local data integration efforts have leveraged their capacity to quickly respond to and understand the impacts of COVID-19, while also reflecting on what can be done to mitigate harm and shift thinking about social welfare and the safety net….(More)”.

Personal data, public data, privacy & power: GDPR & company data


Open Corporates: “…there are three other aspects which are relevant when talking about access to EU company data.

Cargo-culting GDPR

The first, is a tendency to take this complex and subtle legislation that is GDPR and use a poorly understood version in other legislation and regulation, even if that regulation is already covered by GDPR. This actually undermines the GDPR regime, and prevents it from working effectively, and should strongly be resisted. In the tech world, such approaches are called ‘cargo-culting’.

Similarly GDPR is often used as an excuse for not releasing company information as open data, even when the same data is being sold to third parties apparently without concerns — if one is covered by GDPR, the other certainly should be.

Widened power asymmetries

The second issue is the unintended consequences of GDPR, specifically the way it increases asymmetries of power and agency. For example, something like the so-called Right To Be Forgotten takes very significant resources to implement, and so actually strengthens the position of the giant tech companies — for such companies, investing millions in large teams to decide who should and should not be given the Right To Be Forgotten is just a relatively small cost of doing business.

Another issue is the growth of a whole new industry dedicated to removing traces of people’s past from the internet (2), which is also increasing the asymmetries of power. The vast majority of people are not directors of companies, or beneficial owners, and it is only the relatively rich and powerful (including politicians and criminals) who can afford lawyers to stifle free speech, or remove parts of their past they would rather not be there, from business failures to associations with criminals.

OpenCorporates, for example, was threatened with a lawsuit from a member of one of the wealthiest families in Europe for reproducing a gazette notice from the Luxembourg official gazette (a publication that contains public notices). We refused to back down, believing we had a good case in law and in the public interest, and the other side gave up. But such so-called SLAPP suits are becoming increasingly common, although unlike many US states there are currently no defences in place to resist these in the EU, despite pressure from civil society to address this….

At the same time, the automatic assumption that all Personally Identifiable Information (PII), someone’s name for example, is private is highly problematic, confusing both citizens and policy makers, and further undermining democracies and fair societies. As an obvious case, it’s critical that we know the names of our elected representatives, and those in positions of power, otherwise we would have an opaque society where decisions are made by nameless individuals with opaque agendas and personal interests — such as a leader awarding a contract to their brother’s company, for example.

As the diagram below illustrates, there is some personally identifiable information that it’s strongly in the public interest to know. Take the director or beneficial owner of a company, for example, of course their details are PII — clearly you need to know their name (and other information too), otherwise what actually do you know about them, or the company (only that some unnamed individual has been given special protection under law to be shielded from the company’s debts and actions, and yet can benefit from its profits)?

On the other hand, much of the data which is truly about our privacy — the profiles, inferences and scores that companies store on us — is explicitly outside GDPR, if it doesn’t contain PII.

Image for post

Hopefully, as awareness of the issues increases, we will develop a more nuanced, deeper, understanding of privacy, such that case law around GDPR, and successors to this legislation begin to rebalance and case law starts to bring clarity to the ambiguities of the GDPR….(More)”.

The EU is launching a market for personal data. Here’s what that means for privacy.


Anna Artyushina at MIT Tech Review: “The European Union has long been a trendsetter in privacy regulation. Its General Data Protection Regulation (GDPR) and stringent antitrust laws have inspired new legislation around the world. For decades, the EU has codified protections on personal data and fought against what it viewed as commercial exploitation of private information, proudly positioning its regulations in contrast to the light-touch privacy policies in the United States.

The new European data governance strategy (pdf) takes a fundamentally different approach. With it, the EU will become an active player in facilitating the use and monetization of its citizens’ personal data. Unveiled by the European Commission in February 2020, the strategy outlines policy measures and investments to be rolled out in the next five years.

This new strategy represents a radical shift in the EU’s focus, from protecting individual privacy to promoting data sharing as a civic duty. Specifically, it will create a pan-European market for personal data through a mechanism called a data trust. A data trust is a steward that manages people’s data on their behalf and has fiduciary duties toward its clients.

The EU’s new plan considers personal data to be a key asset for Europe. However, this approach raises some questions. First, the EU’s intent to profit from the personal data it collects puts European governments in a weak position to regulate the industry. Second, the improper use of data trusts can actually deprive citizens of their rights to their own data.

The Trusts Project, the first initiative put forth by the new EU policies, will be implemented by 2022. With a €7 million budget, it will set up a pan-European pool of personal and nonpersonal information that should become a one-stop shop for businesses and governments looking to access citizens’ information.

Global technology companies will not be allowed to store or move Europeans’ data. Instead, they will be required to access it via the trusts. Citizens will collect “data dividends,” which haven’t been clearly defined but could include monetary or nonmonetary payments from companies that use their personal data. With the EU’s roughly 500 million citizens poised to become data sources, the trusts will create the world’s largest data market.

For citizens, this means the data created by them and about them will be held in public servers and managed by data trusts. The European Commission envisions the trusts as a way to help European businesses and governments reuse and extract value from the massive amounts of data produced across the region, and to help European citizens benefit from their information. The project documentation, however, does not specify how individuals will be compensated.

Data trusts were first proposed by internet pioneer Sir Tim Berners Lee in 2018, and the concept has drawn considerable interest since then. Just like the trusts used to manage one’s property, data trusts may serve different purposes: they can be for-profit enterprises, or they can be set up for data storage and protection, or to work for a charitable cause.

IBM and Mastercard have built a data trust to manage the financial information of their European clients in Ireland; the UK and Canada have employed data trusts to stimulate the growth of the AI industries there; and recently, India announced plans to establish its own public data trust to spur the growth of technology companies.

The new EU project is modeled on Austria’s digital system, which keeps track of information produced by and about its citizens by assigning them unique identifiers and storing the data in public repositories.

Unfortunately, data trusts do not guarantee more transparency. The trust is governed by a charter created by the trust’s settlor, and its rules can be made to prioritize someone’s interests. The trust is run by a board of directors, which means a party that has more seats gains significant control.

The Trusts Project is bound to face some governance issues of its own. Public and private actors often do not see eye to eye when it comes to running critical infrastructure or managing valuable assets. Technology companies tend to favor policies that create opportunity for their own products and services. Caught in a conflict of interest, Europe may overlook the question of privacy….(More)”.

The Risks and Rewards of Data Sharing for Smart Cities


Study by Massimo Russo and Tian Feng: “…To develop innovative solutions to problems old and new, many cities are aggregating and sharing more and more data, establishing platforms to facilitate private-sector participation, and holding “hackathons” and other digital events to invite public help. But digital solutions carry their own complications. Technology-led innovation often depends on access to data from a wide variety of sources to derive correlations and insights. Questions regarding data ownership, amalgamation, compensation, and privacy can be flashing red lights.

Smart cities are on the leading edge of the trend toward greater data sharing. They are also complex generators and users of data. Companies, industries, governments, and others are following in their wake, sharing more data in order to foster innovation and address such macro-level challenges as public health and welfare and climate change. Smart cities thus provide a constructive laboratory for studying the challenges and benefits of data sharing.

WHY CITIES SHARE DATA

BCG examined some 75 smart-city applications that use data from a variety of sources, including connected equipment (that is, the Internet of Things, or IoT). Nearly half the applications require data sourced from multiple industries or platforms. (See Exhibit 1.) For example, a parking reservation app assembles garage occupancy data, historical traffic data, current weather data, and information on upcoming public events to determine real-time parking costs. We also looked at a broader set of potential future applications and found that an additional 40% will likewise require cross-industry data aggregation.

Because today’s smart solutions are often sponsored by individual municipal departments, many IoT-enabled applications rely on limited, siloed data. But given the potential value of applications that require aggregation across sources, it’s no surprise that many cities are pursuing partnerships with tech providers to develop platforms and other initiatives that integrate data from multiple sources….(More)”.

What privacy preserving techniques make possible: for transport authorities


Blog by Georgina Bourke: “The Mayor of London listed cycling and walking as key population health indicators in the London Health Inequalities Strategy. The pandemic has only amplified the need for people to use cycling as a safer and healthier mode of transport. Yet as the majority of cyclists are white, Black communities are less likely to get the health benefits that cycling provides. Groups like Transport for London (TfL) should monitor how different communities cycle and who is excluded. Organisations like the London Office of Technology and Innovation (LOTI) could help boroughs procure privacy preserving technology to help their efforts.

But at the moment, it’s difficult for public organisations to access mobility data held by private companies. One reason is because mobility data is sensitive. Even if you remove identifiers like name and address, there’s still a risk you can reidentify someone by linking different data sets together. This means you could track how an individual moved around a city. I wrote more about the privacy risks with mobility data in a previous blog post. The industry’s awareness of privacy issues in using and sharing mobility data is rising. In the case of Los Angeles Department of Transport’s Mobility Data Specification (LADOT), Uber is concerned about sharing anonymised data because of the privacy risk. Both organisations are now involved in a legal battle to see which has the rights to the data. This might have been avoided if Uber had applied privacy preserving techniques….

Privacy preserving techniques can help mobility providers share important insights with authorities without compromising peoples’ privacy.

Instead of requiring access to all customer trip data, authorities could ask specific questions like, where are the least popular places to cycle? If mobility providers apply techniques like randomised response, an individual’s identity is obscured by the noise added to the data. This means it’s highly unlikely that someone could be reidentified later on. And because this technique requires authorities to ask very specific questions – for randomised response to work, the answer has to be binary, ie Yes or No – authorities will also be practicing data minimisation by default.

It’s easy to imagine transport authorities like TfL combining privacy preserved mobility data from multiple mobility providers to compare insights and measure service provision. They could cross reference the privacy preserved bike trip data with demographic data in the local area to learn how different communities cycle. The first step to addressing inequality is being able to measure it….(More)”.

The co-ops that electrified Depression-era farms are now building rural internet


Nicolás Rivero at Quartz: “In 2017, Mark McKinney decided enough was enough. The head of the Jackson County Rural Electric Membership Corporation in southern Indiana, a co-op that provides electricity to a rural community of 20,000 members, McKinney was still living without a reliable internet connection. No internet service provider would build the infrastructure to get him or his neighbors online.

“We realized no one was interested due to the capital expense and limited number of members per mile,” says McKinney, “so the board made the decision to go at it on our own.”

The coronavirus pandemic quickly proved the wisdom of their decision: Thanks to their new fiber optic connection, McKinney and his wife were able to self-quarantine without missing work after they were exposed to the virus. Their son finished the spring semester at home after his university shut down in March. “We could not have done that without this connection,” he said.

Across the rural US, more than 100 cooperatives, first launched to provide electric and telephone services as far back as the 1930s, are now laying miles of fiber optic cable to connect their members to high speed internet. Many started building their own networks after failing to convince established internet service providers to cover their communities.

But while rural fiber optic networks have spread swiftly over the past five years, their progress has been uneven. In North Dakota, for example, fiber optic co-ops cover 82% of the state’s landmass, while Nevada has just one co-op. And in the states where the utilities do exist, they tend to serve the whitest communities….(More)”.

The Truth Is Paywalled But The Lies Are Free


Essay by Nathan J. Robinson: “…This means that a lot of the most vital information will end up locked behind the paywall. And while I am not much of a New Yorker fan either, it’s concerning that the Hoover Institute will freely give you Richard Epstein’s infamous article downplaying the threat of coronavirus, but Isaac Chotiner’s interview demolishing Epstein requires a monthly subscription, meaning that the lie is more accessible than its refutation. Eric Levitz of New York is one of the best and most prolific left political commentators we have. But unless you’re a subscriber of New York, you won’t get to hear much of what he has to say each month. 

Possibly even worse is the fact that so much academic writing is kept behind vastly more costly paywalls. A white supremacist on YouTube will tell you all about race and IQ but if you want to read a careful scholarly refutation, obtaining a legal PDF from the journal publisher would cost you $14.95, a price nobody in their right mind would pay for one article if they can’t get institutional access. (I recently gave up on trying to access a scholarly article because I could not find a way to get it for less than $39.95, though in that case the article was garbage rather than gold.) Academic publishing is a nightmarish patchwork, with lots of articles advertised at exorbitant fees on one site, and then for free on another, or accessible only through certain databases, which your university or public library may or may not have access to. (Libraries have to budget carefully because subscription prices are often nuts. A library subscription to the Journal of Coordination Chemistryfor instance, costs $11,367 annually.) 

Of course, people can find their ways around paywalls. SciHub is a completely illegal but extremely convenient means of obtaining academic research for free. (I am purely describing it, not advocating it.) You can find a free version of the article debunking race and IQ myths on ResearchGate, a site that has engaged in mass copyright infringement in order to make research accessible. Often, because journal publishers tightly control access to their copyrighted work in order to charge those exorbitant fees for PDFs, the versions of articles that you can get for free are drafts that have not yet gone through peer review, and have thus been subjected to less scrutiny. This means that the more reliable an article is, the less accessible it is. On the other hand, pseudo-scholarhip is easy to find. Right-wing think tanks like the Cato Institute, the Foundation for Economic Education, the Hoover Institution, the Mackinac Center, the American Enterprise Institute, and the Heritage Foundation pump out slickly-produced policy documents on every subject under the sun. They are utterly untrustworthy—the conclusion is always going to be “let the free market handle the problem,” no matter what the problem or what the facts of the case. But it is often dressed up to look sober-minded and non-ideological. 

It’s not easy or cheap to be an “independent researcher.” When I was writing my first book, Superpredator, I wanted to look through newspaper, magazine, and journal archives to find everything I could about Bill Clinton’s record on race. I was lucky I had a university affiliation, because this gave me access to databases like LexisNexis. If I hadn’t, the cost of finding out what I wanted to find out would likely have run into the thousands of dollars.  

A problem beyond cost, though, is convenience. I find that even when I am doing research through databases and my university library, it is often an absolute mess: the sites are clunky and constantly demanding login credentials. The amount of time wasted in figuring out how to obtain a piece of research material is a massive cost on top of the actual pricing. The federal court document database, PACER, for instance, charges 10 cents a page for access to records, which adds up quickly since legal research often involves looking through thousands of pages. They offer an exemption if you are a researcher or can’t afford it, but to get the exemption you have to fill out a three page form and provide an explanation of both why you need each document and why you deserve the exemption. This is a waste of time that inhibits people’s productivity and limits their access to knowledge.

In fact, to see just how much human potential is being squandered by having knowledge dispensed by the “free market,” let us briefly picture what “totally democratic and accessible knowledge” would look like…(More)”.

A Time for More Democracy Not Less


Graham Smith at Involve: “As part of the “A democratic response to COVID-19” project, we have been scanning print and social media to get a sense of how arguments for participation and deliberation are resonating in public debates….

Researchers from the Institute for Development Studies point to learning from previous pandemics. Drawing from their experience of working on the ebola epidemic in West Africa, they argue that pandemics are not just technical problems to be solved, but are social in character. They call for more deliberation and participation to ensure that decisions reflect not only the diversity of expert opinion, but also respond to the experiential knowledge of the most vulnerable….

A number of these proposals call for citizens’ assemblies, perhaps to the detriment of other participatory and deliberative processes. The Carnegie Trust offers a broader agenda, reminding us of the pressing contemporary significance of their pre-COVID-19 calls for co-design and co-production. 

The Nuffield Council offers some simple guidance to government about how to act:

  • Show us (the public) what it is doing and thinking across the range of issues of concern
  • Set out the ethical considerations that inform(ed) its judgements
  • Explain how it has arrived at decisions (including taking advice from e.g. SAGE, MEAG), and not that it is just ‘following the science’
  • Invite a broad range of perspectives into the room, including wider public representation 
  • Think ahead – consult and engage other civic interests

We have found only a small number of examples of specific initiatives taking a participatory or deliberative approach to bringing in a broader range of voices in response to the pandemic. Our Covid Voices is gathering written statements of the experience of COVID-19 from those with health conditions or disabilities. The thinktank Demos is running a ‘People’s Commission’, inviting stories of lockdown life. It is not only reflections or stories. The Scottish Government invited ideas on how to tackle the virus, receiving and synthesising 4,000 suggestions. The West Midlands Combined Authority has established a citizens’ panel to guide its recovery work. The UK Citizens’ Assembly (and the French Convention) produced recommendations on how commitments to reach net zero carbon emissions need to be central to a post-COVID-19 recovery. We are sure that these examples only touch the surface of activity and that there will be many more initiatives that we are yet to hear about.

Of course, in one area, citizens have already taken matters into their own hands, with the huge growth in mutual-aid groups to ensure people’s emergency needs are met. The New Local Government Network has considered how public authorities could best support and work with such groups, and Danny Kruger MP was invited by the Prime Minister to investigate how to build on this community-level response.

The call for a more participatory and deliberative approach to governance needs to be more than a niche concern. As the Financial Times recognises, we need a “new civic contract” between government and the people….(More)”.

Resetting the state for the post-covid digital age


Blog by Carlos Santiso: “The COVID-19 crisis is putting our global digital resilience to the test. It has revealed the importance of a country’s digital infrastructure as the backbone of the economy, not just as an enabler of the tech economy. Digitally advanced governments, such as Estonia, have been able to put their entire bureaucracies in remote mode in a matter of days, without major disruption. And some early evidence even suggests that their productivity increased during lockdown.

With the crisis, the costs of not going digital have largely surpassed the risks of doing so. Countries and cities lagging behind have realised the necessity to boost their digital resilience and accelerate their digital transformation. Spain, for example, adopted an ambitious plan to inject 70 billion euro into in its digital transformation over the next five years, with a Digital Spain 2025 agenda comprising 10 priorities and 48 measures. In the case of Brazil, the country was already taking steps towards the digital transformation of its public sector before the COVID-19 crisis hit. The crisis is accelerating this transformation.

The great accelerator

Long before the crisis hit, the data-driven digital revolution has been challenging governments to modernise and become more agile, open and responsive. Progress has nevertheless been uneven, hindered by a variety of factors, from political resistance to budget constraints. Going digital requires the sort of whole-of government reforms that need political muscle and long-term vision to break-up traditional data silos within bureaucracies, jealous to preserve their power. In bureaucracies, information is power. Now, information has become ubiquitous and governing data, a critical challenge.

Cutting red tape will be central to the recovery. Many governments are fast-tracking regulatory simplification and administrative streamlining to reboot hard-hit economic sectors. Digitalisation is resetting the relationship between states and citizens, a Copernican revolution for our rule-based bureaucracies….(More)“.

Data for Policy: Junk-Food Diet or Technological Frontier?


Blog by Ed Humpherson at Data & Policy: “At the Office for Statistics Regulation, thinking about these questions is our day job. We set the standards for Government statistics and data through our Code of Practice for Statistics. And we review how Government departments are living up to these standards when they publish data and statistics. We routinely look at Government statistics are used in public debate.

Based on this, I would propose four factors that ensure that new data sources and tools serve the public good. They do so when:

  1. When data quality is properly tested and understood:

As my colleague Penny Babb wrote recently in a blog“‘Don’t trust the data. If you’ve found something interesting, something has probably gone wrong!”. People who work routinely with data develop a sort of innate scepticism, which Penny’s blog captures neatly. Understanding the limitations of both the data, and the inferences you make about the data, are the starting point for any appropriate role for data and policy. Accepting results and insights from new data at face value is a mistake. Much better to test the quality, explore the risks of mistakes, and only then to share findings and conclusions.

2. When the risks of misleadingness are considered:

At OSR, we have an approach to misleadingness that focuses on whether a misuse of data might lead a listener to a wrong conclusion. In fact, by “wrong” we don’t mean in some absolute sense of objective truth; more that if they received the data presented in a different and more faithful way, they would change their mind. Here’s a really simple example: someone might hear that, of two neighbouring countries, one has a much lower fatality rate, when comparing deaths to positive tests for Covid-19. …

3. When the data fill gaps

Data gaps come in several forms. One gap, highlighted by the interest in real-time economic indicators, is timing. Economic statistics don’t really tell us what’s going on right now. Figures like GDP, trade and inflation tells us about some point in the (admittedly quite) recent past. This is the attraction of the real-time economic indicators, which the Bank of England have drawn on in their decisions during the pandemic. They give policymakers a much more real-time feel by filling in this timing gap.

Other gaps are not about time but about coverage….

4. When the data are available

Perhaps the most important thing for data and policy is to democratise the notion of who the data are for. Data (and policy itself) are not just for decision-making elites. They are a tool to help people make sense of their world, what is going on in their community, helping frame and guide the choices they make.

For this reason, I often instinctively recoil at narratives of data that focus on the usefulness of data to decision-makers. Of course, we are all decision-makers of one kind or another, and data can help us all. But I always suspect that the “data for decision-makers” narrative harbours an assumption that decisions are made by senior, central, expert people, who make decisions on behalf of society; people who are, in the words of the musical Hamilton, in the room where it happens. It’s this implication that I find uncomfortable.

That’s why, during the pandemic, our work at the Office for Statistics Regulation has repeatedly argued that data should be made available. We have published a statement that any management information referred to by a decision maker should be published clearly and openly. We call this equality of access.

We fight for equality of access. We have secured the publication of lots of data — on positive Covid-19 cases in England’s Local Authorities, on Covid-19 in prisons, on antibody testing in Scotland…. and several others.

Data and policy are a powerful mix. They offer huge benefits to society in terms of defining, understanding and solving problems, and thereby in improving lives. We should be pleased that the coming together of data and policy is being sped-up by the pandemic.

But to secure these benefits, we need to focus on four things: quality, misleadingness, gaps, and public availability….(More)”