Why We Should Care About Bad Data


Blog by Stefaan G. Verhulst: “At a time of open and big data, data-led and evidence-based policy making has great potential to improve problem solving but will have limited, if not harmful, effects if the underlying components are riddled with bad data.

Why should we care about bad data? What do we mean by bad data? And what are the determining factors contributing to bad data that if understood and addressed could prevent or tackle bad data? These questions were the subject of my short presentation during a recent webinar on  Bad Data: The Hobgoblin of Effective Government, hosted by the American Society for Public Administration and moderated by Richard Greene (Partner, Barrett and Greene Inc.). Other panelists included Ben Ward (Manager, Information Technology Audits Unit, California State Auditor’s Office) and Katherine Barrett (Partner, Barrett and Greene Inc.). The webinar was a follow-up to the excellent Special Issue of Governing on Bad Data written by Richard and Katherine….(More)”

Design Thinking for the Greater Good


New Book by Jeanne Liedtka, Randy Salzman, and Daisy Azer:  “Facing especially wicked problems, social sector organizations are searching for powerful new methods to understand and address them. Design Thinking for the Greater Good goes in depth on both the how of using new tools and the why. As a way to reframe problems, ideate solutions, and iterate toward better answers, design thinking is already well established in the commercial world. Through ten stories of struggles and successes in fields such as health care, education, agriculture, transportation, social services, and security, the authors show how collaborative creativity can shake up even the most entrenched bureaucracies—and provide a practical roadmap for readers to implement these tools.

The design thinkers Jeanne Liedtka, Randy Salzman, and Daisy Azer explore how major agencies like the Department of Health and Human Services and the Transportation and Security Administration in the United States, as well as organizations in Canada, Australia, and the United Kingdom, have instituted principles of design thinking. In each case, these groups have used the tools of design thinking to reduce risk, manage change, use resources more effectively, bridge the communication gap between parties, and manage the competing demands of diverse stakeholders. Along the way, they have improved the quality of their products and enhanced the experiences of those they serve. These strategies are accessible to analytical and creative types alike, and their benefits extend throughout an organization. This book will help today’s leaders and thinkers implement these practices in their own pursuit of creative solutions that are both innovative and achievable….(More)”.

Waste Is Information


Book by Dietmar Offenhuber: “Waste is material information. Landfills are detailed records of everyday consumption and behavior; much of what we know about the distant past we know from discarded objects unearthed by archaeologists and interpreted by historians. And yet the systems and infrastructures that process our waste often remain opaque. In this book, Dietmar Offenhuber examines waste from the perspective of information, considering emerging practices and technologies for making waste systems legible and how the resulting datasets and visualizations shape infrastructure governance. He does so by looking at three waste tracking and participatory sensing projects in Seattle, São Paulo, and Boston.

Offenhuber expands the notion of urban legibility—the idea that the city can be read like a text—to introduce the concept of infrastructure legibility. He argues that infrastructure governance is enacted through representations of the infrastructural system, and that these representations stem from the different stakeholders’ interests, which drive their efforts to make the system legible. The Trash Track project in Seattle used sensor technology to map discarded items through the waste and recycling systems; the Forager project looked at the informal organization processes of waste pickers working for Brazilian recycling cooperatives; and mobile systems designed by the city of Boston allowed residents to report such infrastructure failures as potholes and garbage spills. Through these case studies, Offenhuber outlines an emerging paradigm of infrastructure governance based on a complex negotiation among users, technology, and the city….(More)”.

Innovation@DFID: Crowdsourcing New Ideas at the UK’s Department for International Development


Paper by Anke Schwittay and Paul Braund: “Over the last decade, traditional development institutions have joined market-based actors in embracing inclusive innovation to ensure the sector’s relevance and impacts. In 2014, the UK’s Department for International Development’s (DFID) Innovation Hub launched Amplify as its own flagship initiative. The programme, which is managed by IDEO, a Silicon Valley-based design consultancy, aims to crowdsource new ideas to various development challenges from a broad and diverse group of actors, including poor people themselves. By examining the direction, diversity and distribution of Amplify’s work, we argue that while development innovation can generate more inclusive practices, its transformative potential is constrained by broader developmental logics and policy regimes….(More)”

America is not a true democracy. But it could be with the help of technology


Nicole Softness at Quartz: “Many Americans aren’t aware they don’t live in a direct democracy. But with a little digital assistance, they could be….Once completely cut off from the global community, Estonia is now considered a world leader for its efforts to integrate technology with government administration. While standing in line for coffee, you could file your tax return, confirm sensitive personal medical information, and register a new company in just a few swipes, all on Estonia’s free wifi.

What makes this possible without the risk of fraud? Digital trust. Using a technology called blockchain, which verifies online communications and transactions at every step (and essentially eliminates the possibility of online fraud), Estonian leadership has moved the majority of citizenship processes online. Startups have now created new channels for democratic participation, like Rahvaalgatus, an online crowdsourcing platform that allows users to discuss and digitally vote on policy proposals submitted to the Estonian parliament.

Brazil has also utilized this trust quite valiantly. The country’s constitution, passed in 1988, legislated that signatures from 1% of a population could force the Brazilian leadership to recognize any signed document as an official draft bill and vote. Until recently, the notion of getting sufficient signatures on paper would have been laughable: that’s just over 2 million physical signatures. However, votes can now be cast online, which makes gathering digital signatures all the more easy. As a result, Brazilians now have more control over the legislature being brought before parliament.

 Blockchain technology creates an immutable record of signatures tied to the identities of voters. Again, blockchain technology is key here, as it creates an immutable record of signatures tied to the identities of voters. The government knows which voters are legitimate citizens, and citizens can be sure their votes remain accurate. When Brazilians are able to participate in this manner, their democracy shifts towards the sort of “direct” democracy that, until now, seemed logistically impossible in modern society.

Australian citizens have engaged in a slightly different experiment, dubbed “Government 2.0.” In March 2016, technology experts convened a new political party called Flux, which they describe as “democracy for the information age.” The party platform argues that bureaucracy stymies key government functions, which cannot process the requisite information required to govern.

If elected to government, members of Flux would vote on bills scheduled to appear before parliament based on the digital ballots of the supporters who voted them in. Voters could choose to participate in casting their vote for that bill themselves, or transfer their votes to trusted experts. Flux representatives in parliament would then cast their votes 100% based on the results of these member participants. (They are yet to win any seats in government, however.)

These solutions show us that bureaucratic boundaries no longer have to limit our access to a true democracy. The technology is here to make direct democracy the reality that the Greeks once imagined.

More so, increasing democratic participation will have positive ripple effects beyond participation in a direct democracy: Informed voting is the gateway to more active civic engagement and a more informed electorate, all of which raises the level of debate in a political environment desperately in need of participation….(More)”

The accuracy of farmer-generated data in an agricultural citizen science methodology


Jonathan Steinke, Jacob van Etten and Pablo Mejía Zelan in Agronomy for Sustainable Development: “Over the last decades, participatory approaches involving on-farm experimentation have become more prevalent in agricultural research. Nevertheless, these approaches remain difficult to scale because they usually require close attention from well-trained professionals. Novel large-N participatory trials, building on recent advances in citizen science and crowdsourcing methodologies, involve large numbers of participants and little researcher supervision. Reduced supervision may affect data quality, but the “Wisdom of Crowds” principle implies that many independent observations from a diverse group of people often lead to highly accurate results when taken together. In this study, we test whether farmer-generated data in agricultural citizen science are good enough to generate valid statements about the research topic. We experimentally assess the accuracy of farmer observations in trials of crowdsourced crop variety selection that use triadic comparisons of technologies (tricot). At five sites in Honduras, 35 farmers (women and men) participated in tricot experiments. They ranked three varieties of common bean (Phaseolus vulgaris L.) for Plant vigorPlant architecturePest resistance, and Disease resistance. Furthermore, with a simulation approach using the empirical data, we did an order-of-magnitude estimation of the sample size of participants needed to produce relevant results. Reliability of farmers’ experimental observations was generally low (Kendall’s W 0.174 to 0.676). But aggregated observations contained information and had sufficient validity (Kendall’s tau coefficient 0.33 to 0.76) to identify the correct ranking orders of varieties by fitting Mallows-Bradley-Terry models to the data. Our sample size simulation shows that low reliability can be compensated by engaging higher numbers of observers to generate statistically meaningful results, demonstrating the usefulness of the Wisdom of Crowds principle in agricultural research. In this first study on data quality from a farmer citizen science methodology, we show that realistic numbers of less than 200 participants can produce meaningful results for agricultural research by tricot-style trials….(More)”.

Crowdsourcing Expertise to Increase Congressional Capacity


Austin Seaborn at Beeck Center: “Members of Congress have close connections with their districts, and information arising from local organizations, such as professional groups, academia, industry as well as constituents with relevant expertise (like retirees, veterans or students) is highly valuable to them.  Today, congressional staff capacity is at a historic low, while at the same time, constituents in districts are often well equipped to address the underlying policy questions that Congress seeks to solve….

In meetings we have had with House and Senate staffers, they repeatedly express both the difficulty managing their substantial area-specific work loads and their interest in finding ways to substantively engage constituents to find good nuggets of information to help them in their roles as policymakers. At the same time, constituents are demanding more transparency and dialogue from their elected representatives. In many cases, our project brings these two together. It allows Members to tap the expertise in their districts while at the same time creating an avenue for constituents to contribute their knowledge and area expertise to the legislative process. It’s a win for constituents and a win for Member of Congress and their staffs.

It is important to note that the United States lags behind other democracies in experimenting with more inclusive methods during the policymaking process. In the United Kingdom, for example, the UK Parliament has experimented with a variety of new digital tools to engage with constituents. These methods range from Twitter hashtags, which are now quite common given the rise in social media use by governments and elected officials, to a variety of web forums on a variety of platforms. Since June of 2015, they have also been doing digital debates, where questions from the general public are crowdsourced and later integrated into a parliamentary debate by the Member of Parliament leading the debate. Estonia, South Africa, Taiwan, France also…notable examples.

One promising new development we hope to explore more thoroughly is the U.S. Library of Congress’s recently announced legislative data App Challenge. This competition is distinct from the many hackathons that have been held on behalf of Congress in the past, in that this challenge seeks new methods not only to innovate, but also to integrate and legislate. In his announcement, the Library’s Chief Information Officer, Bernard A. Barton, Jr., stated, “An informed citizenry is better able to participate in our democracy, and this is a very real opportunity to contribute to a better understanding of the work being done in Washington.  It may even provide insights for the people doing the work around the clock, both on the Hill, and in state and district offices.  Your innovation and integration may ultimately benefit the way our elected officials legislate for our future.” We believe these sorts of new methods will play a crucial role in the future of engaging citizens in their democracies….(More)”.

The DeepMind debacle demands dialogue on data


Hetan Shah in Nature: “Without public approval, advances in how we use data will stall. That is why a regulator’s ruling against the operator of three London hospitals is about more than mishandling records from 1.6 million patients. It is a missed opportunity to have a conversation with the public about appropriate uses for their data….

What can be done to address this deficit? Beyond meeting legal standards, all relevant institutions must take care to show themselves trustworthy in the eyes of the public. The lapses of the Royal Free hospitals and DeepMind provide, by omission, valuable lessons.

The first is to be open about what data are transferred. The extent of data transfer between the Royal Free and DeepMind came to light through investigative journalism. In my opinion, had the project proceeded under open contracting, it would have been subject to public scrutiny, and to questions about whether a company owned by Google — often accused of data monopoly — was best suited to create a relatively simple app.

The second lesson is that data transfer should be proportionate to the task. Information-sharing agreements should specify clear limits. It is unclear why an app for kidney injury requires the identifiable records of every patient seen by three hospitals over a five-year period.

Finally, governance mechanisms must be strengthened. It is shocking to me that the Royal Free did not assess the privacy impact of its actions before handing over access to records. DeepMind does deserve credit for (belatedly) setting up an independent review panel for health-care projects, especially because the panel has a designated budget and has not required members to sign non-disclosure agreements. (The two groups also agreed a new contract late last year, after criticism.)

More is needed. The Information Commissioner asked the Royal Free to improve its processes but did not fine it or require it to rescind data. This rap on the knuckles is unlikely to deter future, potentially worse, misuses of data. People are aware of the potential for over-reach, from the US government’s demands for state voter records to the Chinese government’s alleged plans to create a ‘social credit’ system that would monitor private behaviour.

Innovations such as artificial intelligence, machine learning and the Internet of Things offer great opportunities, but will falter without a public consensus around the role of data. To develop this, all data collectors and crunchers must be open and transparent. Consider how public confidence in genetic modification was lost in Europe, and how that has set back progress.

Public dialogue can build trust through collaborative efforts. A 14-member Citizen’s Reference Panel on health technologies was convened in Ontario, Canada in 2009. The Engage2020 programme incorporates societal input in the Horizon2020 stream of European Union science funding….(More)”

How open data can help the Global South, from disaster relief to voter turnout


Stefaan Verhulst and Andrew Young in The Conversation Global: “The modern era is marked by growing faith in the power of data. “Big data”, “open data”, and “evidence-based decision-making” have become buzzwords, touted as solutions to the world’s most complex and persistent problems, from corruption and famine to the refugee crisis.

While perhaps most pronounced in higher income countries, this trend is now emerging globally. In Africa, Latin America, Asia and beyond, hopes are high that access to data can help developing economies by increasing transparency, fostering sustainable development, building climate resiliency and the like.

This is an exciting prospect, but can opening up data actually make a difference in people’s lives?

Getting data-driven about data

The GovLab at New York University spent the last year trying to answer that question….

Our conclusion: the enthusiasm is justified – as long as it’s tempered with a good measure of realism, too. Here are our six major takeaways:

1. We need a framework – Overall, there is still little evidence to substantiate the enthusiastic claims that open data can foment sustainable development and transform governance. That’s not surprising given the early stage of most open data initiatives.

It may be early for impact evaluation, but it’s not too soon to develop a model that will eventually allow us to assess the impact of opening up data over time.

To that end, the GovLab has created an evidence-based framework that aims to better capture the role of open data in developing countries. The Open Data Logic Framework below focuses on various points in the open data value cycle, from data supply to demand, use and impact.

Logic model of open data. The GovLab

2. Open data has real promise – Based on this framework and the underlying evidence that fed into it, we can guardedly conclude that open data does in fact spur development – but only under certain conditions and within the right supporting ecosystem.

One well-known success took place after Nepal’s 2015 earthquake when open data helped NGOs map important landmarks such as health facilities and road networks, among other uses.

And in Colombia, the International Centre for Tropical Agriculture launched Aclímate Colombia, a tool that gives smallholder farmers data-driven insight into planting strategies that makes them more resilient to climate change….

3. Open data can improve people’s lives Examining projects in a number of sectors critical to development, including health, humanitarian aid, agriculture, poverty alleviation, energy and education, we found four main ways that data can have an impact….

4. Data can be an asset in development While these impacts are apparent in both developed and developing countries, we believe that open data can have a particularly powerful role in developing economies.

Where data is scarce, as it often is in poorer countries, open data can lead to an inherently more equitable and democratic distribution of information and knowledge. This, in turn, may activate a wider range of expertise to address complex problems; it’s what we in the field call “open innovation”.

This quality can allow resource-starved developing economies to access and leverage the best minds around.

And because trust in government is quite low in many developing economies, the transparency bred of releasing data can have after-effects that go well beyond the immediate impact of the data itself…

5. The ingredients matter To better understand why some open data projects fail while others succeed, we created a “periodic table” of open data (below), which includes 27 enabling factors divided into five broad categories….

6. We can plan for impact Our report ends by identifying how development organisations can catalyse the release and use of open data to make a difference on the ground.

Recommendations include:

· Define the problem, understand the user, and be aware of local conditions;

· Focus on readiness, responsiveness and change management;

· Nurture an open data ecosystem through collaboration and partnerships;

· Have a risk mitigation strategy;

· Secure resources and focus on sustainability; and

· Build a strong evidence base and support more research.

Next steps

In short, while it may still be too early to fully capture open data’s as-of-yet muted impact on developing economies, there are certainly reasons for optimism.

Much like blockchaindrones and other much-hyped technical advances, it’s time to start substantiating the excitement over open data with real, hard evidence.

The next step is to get systematic, using the kind of analytical framework we present here to gain comparative and actionable insight into if, when and how open data works. Only by getting data-driven about open data can we help it live up to its potential….(More)

Why and How Open Data Matters for Developing Economies


The GovLab is pleased to announce the launch of a new report that assesses and explores ways open data can be used in developing economies, “Open Data in Developing Economies: Toward Building an Evidence Base on What Works and How.” The new report co-authored by Stefaan Verhulst and Andrew Young seeks to :

(1) provides an evidence-based framework governments, NGOs, donors, and others can use to assess the impacts resulting from the use of open data in developing economies;

(2) outlines four key impact areas gleaned from 12 case studies that feature real-world examples from 12 countries, ranging from Colombia to Nepal; and

(3) identifies 27 critical factors that help to determine the success (or failure) of open data uses in developing economies, organized into a newly-created Periodic Table of Open Data, along with recommendations for both practitioners and decision-makers, including donor agencies.

table

The “Open Data in Developing Economies” report is the result of a months-long collaboration between The GovLab and several partners, including the U.S. Agency for International Development, FHI 360, and the World Wide Web Foundation.

The full report and cases studies are available at ODimpact.org…”