Open Data Is Changing the World in Four Ways…


 at The GovLab Blog: “New repository of case studies documents the impact of open data globally: odimpact.org.

odimpact-tweet-3

Despite global commitments to and increasing enthusiasm for open data, little is actually known about its use and impact. What kinds of social and economic transformation has open data brought about, and what is its future potential? How—and under what circumstances—has it been most effective? How have open data practitioners mitigated risks and maximized social good?

Even as proponents of open data extol its virtues, the field continues to suffer from a paucity of empiricalevidence. This limits our understanding of open data and its impact.

Over the last few months, The GovLab (@thegovlab), in collaboration with Omidyar Network(@OmidyarNetwork), has worked to address these shortcomings by developing 19 detailed open data case studies from around the world. The case studies have been selected for their sectoral and geographic representativeness. They are built in part from secondary sources (“desk research”), and also from more than60 first-hand interviews with important players and key stakeholders. In a related collaboration withOmidyar Network, Becky Hogge(@barefoot_techie), an independent researcher, has developed an additional six open data case studies, all focused on the United Kingdom.  Together, these case studies, seek to provide a more nuanced understanding of the various processes and factors underlying the demand, supply, release, use and impact of open data.

Today, after receiving and integrating comments from dozens of peer reviewers through a unique open process, we are delighted to share an initial batch of 10 case studies, as well three of Hogge’s UK-based stories. These are being made available at a new custom-built repository, Open Data’s Impact (http://odimpact.org), that will eventually house all the case studies, key findings across the studies, and additional resources related to the impact of open data. All this information will be stored in machine-readable HTML and PDF format, and will be searchable by area of impact, sector and region….(More)

Data Science ethics


Gov.uk blog: “If Tesco knows day-to-day how poorly the nation is, how can Government access  similar  insights so it can better plan health services? If Airbnb can give you a tailored service depending on your tastes, how can Government provide people with the right support to help them back into work in a way that is right for them? If companies are routinely using social media data to get feedback from their customers to improve their services, how can Government also use publicly available data to do the same?

Data science allows us to use new types of data and powerful tools to analyse this more quickly and more objectively than any human could. It can put us in the vanguard of policymaking – revealing new insights that leads to better and more tailored interventions. And  it can help reduce costs, freeing up resource to spend on more serious cases.

But some of these data uses and machine-learning techniques are new and still relatively untested in Government. Of course, we operate within legal frameworks such as the Data Protection Act and Intellectual Property law. These are flexible but don’t always talk explicitly about the new challenges data science throws up. For example, how are you to explain the decision making process of a deep learning black box algorithm? And if you were able to, how would you do so in plain English and not a row of 0s and 1s?

We want data scientists to feel confident to innovate with data, alongside  the policy makers and operational staff who make daily decisions on the data that the analysts provide –. That’s why we are creating an ethical framework which brings together the relevant parts of the law and ethical considerations into a simple document that helps Government officials decide what it can do and what it should do. We have a moral responsibility to maximise the use of data – which is never more apparent than after incidents of abuse or crime are left undetected – as well as to pay heed to the potential risks of these new tools. The guidelines are draft and not formal government policy, but we want to share them more widely in order to help iterate and improve them further….

So what’s in the framework? There is more detail in the fuller document, but it is based around six key principles:

  1. Start with a clear user need and public benefit: this will help you justify the level of data sensitivity and method you use
  2. Use the minimum level of data necessary to fulfill the public benefit: there are many techniques for doing so, such as de-identification, aggregation or querying against data
  3. Build robust data science models: the model is only as good as the data it contains and while machines are less biased than humans they can get it wrong. It’s critical to be clear about the confidence of the model and think through unintended consequences and biases contained within the data
  4. Be alert to public perceptions: put simply, what would a normal person on the street think about the project?
  5. Be as open and accountable as possible: Transparency is the antiseptic for unethical behavior. Aim to be as open as possible (with explanations in plain English), although in certain public protection cases the ability to be transparent will be constrained.
  6. Keep data safe and secure: this is not restricted to data science projects but we know that the public are most concerned about losing control of their data….(More)”

Controlling the crowd? Government and citizen interaction on emergency-response platforms


 at the Policy and Internet Blog: “My interest in the role of crowdsourcing tools and practices in emergency situations was triggered by my personal experience. In 2010 I was one of the co-founders of the Russian “Help Map” project, which facilitated volunteer-based response to wildfires in central Russia. When I was working on this project, I realized that a crowdsourcing platform can bring the participation of the citizen to a new level and transform sporadic initiatives by single citizens and groups into large-scale, relatively well coordinated operations. What was also important was that both the needs and the forms of participation required in order to address these needs be defined by the users themselves.

To some extent the citizen-based response filled the gap left by the lack of a sufficient response from the traditional institutions.[1] This suggests that the role of ICTs in disaster response should be examined within the political context of the power relationship between members of the public who use digital tools and the traditional institutions. My experience in 2010 was the first time I was able to see that, while we would expect that in a case of natural disaster both the authorities and the citizens would be mostly concerned about the emergency, the actual situation might be different.

Apparently the emergence of independent, citizen-based collective action in response to a disaster was considered as some type of threat by the institutional actors. First, it was a threat to the image of these institutions, which didn’t want citizens to be portrayed as the leading responding actors. Second, any type of citizen-based collective action, even if not purely political, may be an issue of concern in authoritarian countries in particular. Accordingly, one can argue that, while citizens are struggling against a disaster, in some cases the traditional institutions may make substantial efforts to restrain and contain the action of citizens. In this light, the role of information technologies can include not only enhancing citizen engagement and increasing the efficiency of the response, but also controlling the digital crowd of potential volunteers.

The purpose of this paper was to conceptualize the tension between the role of ICTs in the engagement of the crowd and its resources, and the role of ICTs in controlling the resources of the crowd. The research suggests a theoretical and methodological framework that allows us to explore this tension. The paper focuses on an analysis of specific platforms and suggests empirical data about the structure of the platforms, and interviews with developers and administrators of the platforms. This data is used in order to identify how tools of engagement are transformed into tools of control, and what major differences there are between platforms that seek to achieve these two goals. That said, obviously any platform can have properties of control and properties of engagement at the same time; however the proportion of these two types of elements can differ significantly.

One of the core issues for my research is how traditional actors respond to fast, bottom-up innovation by citizens.[2]. On the one hand, the authorities try to restrict the empowerment of citizens by the new tools. On the other hand, the institutional actors also seek to innovate and develop new tools that can restore the balance of power that has been challenged by citizen-based innovation. The tension between using digital tools for the engagement of the crowd and for control of the crowd can be considered as one of the aspects of this dynamic.

That doesn’t mean that all state-backed platforms are created solely for the purpose of control. One can argue, however, that the development of digital tools that offer a mechanism of command and control over the resources of the crowd is prevalent among the projects that are supported by the authorities. This can also be approached as a means of using information technologies in order to include the digital crowd within the “vertical of power”, which is a top-down strategy of governance. That is why this paper seeks to conceptualize this phenomenon as “vertical crowdsourcing”.

The question of whether using a digital tool as a mechanism of control is intentional is to some extent secondary. What is important is that the analysis of platform structures relying on activity theory identifies a number of properties that allow us to argue that these tools are primarily tools of control. The conceptual framework introduced in the paper is used in order to follow the transformation of tools for the engagement of the crowd into tools of control over the crowd. That said, some of the interviews with the developers and administrators of the platforms may suggest the intentional nature of the development of tools of control, while crowd engagement is secondary….Read the full article: Asmolov, G. (2015) Vertical Crowdsourcing in Russia: Balancing Governance of Crowds and State–Citizen Partnership in Emergency Situations.”

 

Data enriched research, data enhanced impact: the importance of UK data infrastructure.


Matthew Woollard at LSE Impact Blog: “…Data made available for reuse, such as those in the UK Data Service collection have huge potential. They can unlock new discoveries in research, provide evidence for policy decisions and help promote core data skills in the next generation of researchers. By being part of a single infrastructure, data owners and data creators can work together with the UK Data Service – rather than duplicating efforts – to engage with the people who can drive the impact of their research further to provide real benefit to society. As a service we are also identifying new ways to understand and promote our impact, and our Impact Fellow and Director of Impact and Communications, Victoria Moody, is focusing on raising the visibility of the UK Data Service holdings and developing and promoting the use and impact of the data and resources in policy-relevant research, especially to new audiences such as policymakers, government sectors, charities, the private sector and the media…..

We are improving how we demonstrate the impact of both the Service and the data which we hold, by focusing on generating more and more authentic user corroboration. Our emphasis is on drawing together evidence about the reach and significance of the impact of our data and resources, and of the Service as a whole through our infrastructure and expertise. Headline impact indicators through which we will better understand our impact cover a range of areas (outlined above) where the Service brings efficiency to data access and re-use, benefit to its users and a financial and social return on investment.

We are working to understand more about how Service data contributes to impact by tracking the use of Service data in a range of initiatives focused on developing impact from research and by developing our insight into usage of our data by our users. Data in the collection have featured in a range of impact case studies in the Research Excellence Framework 2014. We are also developing a focus on understanding the specific beneficial effect, rather than simply that data were used in an output, that is – as it appears in policy, debate or the evidential process (although important). Early thoughts in developing this process are where (ideally) cited data can be tracked through the specific beneficial outcome and on to an evidenced effect, corroborated by the end user.

data service 1

Our impact case studies demonstrate how the data have supported research which has led to policy change in a range of areas including; the development of mathematical models for Practice based Commissioning budgets for adult mental health in the UK and informing public policy on obesity; both using the Health Survey for England. Service data have also informed the development of impact around understanding public attitudes towards the police and other legal institutions using the Crime Survey for England and Wales and research to support the development of the national minimum wage using the Labour Force Survey. The cutting-edge new Demos Integration Hub maps the changing face of Britain’s diversity, revealing a mixed picture in the integration and upward mobility of ethnic minority communities and uses 2011 Census aggregate data (England and Wales) and Understanding Society….(More)”

Looking for Open Data from a different country? Try the European Data portal


Wendy Carrara in DAE blog: “The Open Data movement is reaching all countries in Europe. Data Portals give you access to re-usable government information. But have you ever tried to find Open Data from another country whose language you do not speak? Or have you tried to see whether data from one country exist also in a similar way in another? The European Data Portal that we just launched can help you….

The European Data Portal project main work streams is the development of a new pan-European open data infrastructure. Its goal is to be a gateway offering access to data published by administrations in countries across Europe, from the EU and beyond.
The portal is launched during the European Data Forum in Luxembourg.

Additionally we will support public administrations in publishing more data as open data and have targeted actions to stimulate re-use. By taking a look at the data released by other countries and made available on the European Data Portal, governments can also be inspired to publish new data sets they had not though about in the first place.

The re-use of Open Data will further boost the economy. The benefits of Open Data are diverse and range from improved performance of public administrations and economic growth in the private sector to wider social welfare. The economic studyconducted by the European Data Portal team estimates that between 2016 and 2020, the market size of Open Data is expected to increase by 36.9% to a value of 75.7 bn EUR in 2020.

For data to be re-used, it has to be accessible

Currently, the portal includes over 240.000 datasets from 34 European countries. Information about the data available is structured into thirteen different categories ranging from agriculture to transport, including science, justice, health and so on. This enables you to quickly browse through categories and feel inspired by the data made accessible….(More)”

An Introduction to System Mapping


Joelle Cook at FSG: “Much has been written recently about the importance of using a system lens, or focusing on system change, to make real progress against some of society’s toughest challenges. But what does that really mean? The following definition of system change resonated with us, fromNPC’s 2015 guide:

“Systems change is an intentional process designed to alter the status quo by shifting the function or structure of an identified system with purposeful interventions. It is a journey which can require a radical change in people’s attitudes as well as in the ways people work. Systems change aims to bring about lasting change by altering underlying structures and supporting mechanisms which make the system operate in a particular way. These can include policies, routines, relationships, resources, power structures and values.”

However, to change the system, you need to first understand the system, and mapping is a great way to do that. A “system,” as described by Julia Coffman in her 2007 framework for evaluating system change, is “a group of interacting, interrelated, and interdependent components that form a complex and unified whole.” A system’s overall purpose or goal is achieved through the actions and interactions of its components.

As you can imagine, there are a number of different ways you might approach mapping the system to represent system elements and connections. For example, you might create:

  • Actor maps, to show which individuals and/or organizations are key players in the space and how they are connected
  • Mind maps, that highlight various trends in the external environment that influence the issue at hand
  • Issue maps, which lay out the political, social, or economic issues affecting a given geography or constituency (often used by advocacy groups)
  • Causal-loop diagrams, that  focus on explicating the feedback loops (positive and negative) that lead to system behavior or functioning

For more information, see a blog from Innovation Network on systems mapping, Jeff Wasbes’blog on causal loop diagrams, and an example from the Hewlett Foundation’s Madison Initiative….(More)”

Open Data as Open Educational Resources: Case studies of emerging practice


Book edited by Javiera Atenas and Leo Havemann: “…is the outcome of a collective effort that has its origins in the 5th Open Knowledge Open Education Working Group call, in which the idea of using Open Data in schools was mentioned. It occurred to us that Open Data and open educational resources seemed to us almost to exist in separate open worlds.

We decided to seek out evidence in the use of open data as OER, initially by conducting a bibliographical search. As we could not find published evidence, we decided to ask educators if they were in fact, using open data in this way, and wrote a post for this blog (with Ernesto Priego) explaining our perspective, called The 21st Century’s Raw Material: Using Open Data as Open Educational Resources. We ended the post with a link to an exploratory survey, the results of which indicated a need for more awareness of the existence and potential value of Open Data amongst educators…..

the case studies themselves. They have been provided by scholars and practitioners from different disciplines and countries, and they reflect different approaches to the use of open data. The first case study presents an approach to educating both teachers and students in the use of open data for civil monitoring via Scuola di OpenCoesione in Italy, and has been written by Chiara Ciociola and Luigi Reggi. The second case, by Tim Coughlan from the Open University, UK, showcases practical applications in the use of local and contextualised open data for the development of apps. The third case, written by Katie Shamash, Juan Pablo Alperin & Alessandra Bordini from Simon Fraser University, Canada, demonstrates how publishing students can engage, through data analysis, in very current debates around scholarly communications and be encouraged to publish their own findings. The fourth case by Alan Dix from Talis and University of Birmingham, UK, and Geoffrey Ellis from University of Konstanz, Germany, is unique because the data discussed in this case is self-produced, indeed ‘quantified self’ data, which was used with students as material for class discussion and, separately, as source data for another student’s dissertation project. Finally, the fifth case, presented by Virginia Power from University of the West of England, UK, examines strategies to develop data and statistical literacies in future librarians and knowledge managers, aiming to support and extend their theoretical understanding of the concept of the ‘knowledge society’ through the use of Open Data….(More)

The book can be downloaded here Open Data as Open Educational Resources

Role of Citizens in India’s Smart Cities Challenge


Florence Engasser and Tom Saunders at the World Policy Blog: “India faces a wide range of urban challenges — from serious air pollution and poor local governance, to badly planned cities and a lack of decent housing. India’s Smart Cities Challenge, which has now selected 98 of the 100 cities that will receive funding, could go a long way in addressing these issues.

According to Prime Minister Narendra Modi, there are five key instruments that make a “smart” city: the use of clean technologies, the use of information and communications technology (ICT), private sector involvement, citizen participation and smart governance. There are good examples of new practices for each of these pillars.

For example, New Delhi recently launched a program to replace streetlights with energy efficient LEDs. The Digital India program is designed to upgrade the country’s IT infrastructure and includes plans to build “broadband highways” across the country. As for private sector participation, the Indian government is trying to encourage it by listing sectors and opportunities for public-private partnerships.

Citizen participation is one of Modi’s five key instruments, but this is an area where smart city pilots around the world have tended to perform least well on. While people are the implied beneficiaries of programs that aim to improve efficiency and reduce waste, they are rarely given a chance to participate in the design or delivery of smart city projects, which are usually implemented and managed by experts who have only a vague idea of the challenges that local communities face.

Citizen Participation

Engaging citizens is especially important in an Indian context because there have already been several striking examples of failed urban redevelopments that have blatantly lacked any type of community consultation or participation….

In practice, how can Indian cities engage residents in their smart city projects?

There are many tools available to policymakers — from traditional community engagement activities such as community meetings, to websites like Mygov.in that ask for feedback on policies. Now, there are a number of reasons to think smartphones could be an important tool to help improve collaboration between residents and city governments in Indian cities.

First, while only around 10 percent of Indians currently own a smartphone, this is predicted to rise to around half by 2020, and will be much higher in urban areas. A key driver of this is local manufacturing giants like Micromax, which have revolutionized low-cost technology in India, with smartphones costing as little as $30 (compared to around $800 for the newest iPhone).

Second, smartphone apps give city governments the potential to interact directly with citizens to make the most of what they know and feel about their communities. This can happen passively, for example, the Waze Connected Citizens program, which shares user location data with city governments to help improve transport planning. It can also be more active, for example, FixMyStreet, which allows people to report maintenance issues like potholes to their city government.

Third, smartphones are one of the main ways for people to access social media, and researchers are now developing a range of new and innovative solutions to address urban challenges using these platforms. This includes Petajakarta, which creates crowd-sourced maps of flooding in Jakarta by aggregating tweets that mention the word ‘flood.’

Made in India

Considering some of the above trends, it is interesting to think about the role smartphones could play in the governance of Indian cities and in better engaging communities. India is far from being behind in the field, and there are already a few really good examples of innovative smartphone applications made in India.

Swachh Bharat Abhiyan (translated as Clean India Initiative) is a campaign launched by Modi in October 2014, covering over 4,000 towns all over the country, with the aim to clean India’s streets. The Clean India mobile application, launched at the end of 2014 to coincide with Modi’s initiative, was developed by Mahek Shah and allows users to take pictures to report, geo-locate, and timestamp streets that need cleaning or problems to be fixed by the local authorities.

Similar to FixMyStreet, users are able to tag their reports with keywords to categorize problems. Today, Clean India has been downloaded over 12,000 times and has 5,000 active users. Although still at a very early stage, Clean India has great potential to facilitate the complaint and reporting process by empowering people to become the eyes and ears of municipalities on the ground, who are often completely unaware of issues that matter to residents.

In Bangalore, an initiative by the MOD Institute, a local nongovernmental organization, enabled residents to come together, online and offline, to create a community vision for the redevelopment of Shanthinagar, a neighborhood of the city. The project, Next Bengaluru, used new technologies to engage local residents in urban planning and tap into their knowledge of the area to promote a vision matching their real needs.

The initiative was very successful. In just three months, between December 2014 and March 2015, over 1,200 neighbors and residents visited the on-site community space, and the team crowd-sourced more than 600 ideas for redevelopment and planning both on-site and through the Next Bangalore website.

The MOD Institute now intends to work with local urban planners to try get these ideas adopted by the city government. The project has also developed a pilot app that will enable people to map abandoned urban spaces via smartphone and messaging service in the future.

Finally, Safecity India is a nonprofit organization providing a platform for anyone to share, anonymously or not, personal stories of sexual harassment and abuse in public spaces. Men and women can report different types of abuses — from ogling, whistles and comments, to stalking, groping and sexual assault. The aggregated data is then mapped, allowing citizens and governments to better understand crime trends at hyper-local levels.

Since its launch in 2012, SafeCity has received more than 4,000 reports of sexual crime and harassment in over 50 cities across India and Nepal. SafeCity helps generate greater awareness, breaks the cultural stigma associated with reporting sexual abuse and gives voice to grassroots movements and campaigns such as SayftyProtsahan, or Stop Street Harassment, forcing authorities to take action….(More)

Cleaning Up Lead Poisoning One Tweet at a Time


WorldPolicy Blog: “At first, no one knew why the children of Bagega in Zamfara state were dying. In the spring of 2010, hundreds of kids in and around the northern Nigerian village were falling ill, having seizures and going blind, many of them never to recover. A Médecins Sans Frontières‎ team soon discovered the causes: gold and lead.

With the global recession causing the price of precious metals to soar, impoverished villagers had turned to mining the area’s gold deposits. But the gold veins were mingled with lead, and as a result the villagers’ low-tech mining methods were sending clouds of lead-laced dust into the air. The miners, unknowingly carrying the powerful toxin on their clothes and skin, brought it into their homes where their children breathed it in.

The result was perhaps the worst outbreak of lead poisoning in history, killing over 400 children in Bagega and neighboring villages. In response, the Nigerian government pledged to cleanup the lead-contaminated topsoil and provide medical care to the stricken children. But by mid-2012, there was no sign of the promised funds. Digitally savvy activists with the organization Connected Development (CODE) stepped in to make sure that the money was disbursed.

A group of young Nigerians founded CODE in 2010 in the capital Abuja, with the mission of empowering local communities to hold the government to account by improving their access to information and helping their voices to be heard. “In 2010, we were working to connect communities with data for advocacy programs,” says CODE co-founder Oludotun Babayemi, a former country director of a World Wildlife Fund project in Nigeria. “When we heard about Bagega, we thought this was an opportunity for us.”

In 2012, CODE launched a campaign dubbed ‘Follow the Money Nigeria’ aimed at applying pressure on the government to release the promised funds. “Eighty percent of the less developed parts of Nigeria have zero access to Twitter, let alone Facebook, so it’s difficult for them to convey their stories,” says Babayemi. “We collect all the videos and testimonies and take it global.”

CODE members travelled to the lead-afflicted area to gather information. They then posted their findings online, and publicized them with a #SaveBagegahashtag, which they tweeted to members of the government, local and international organizations and the general public. CODE hosted a 48-hour ‘tweet-a-thon’, joined by a senator, to support the campaign….

By July 2014, CODE reported that the clean-up was complete and that over 1,000 children had been screened and enrolled in lead treatment programs. Bagega’s health center has also been refurbished and the village’s roads improved. “There are thousands of communities like Bagega,” says Babayemi. “They just need someone to amplify their voice.”….

Key lessons

  • Revealing information is not enough; change requires a real-world campaign driven by that information and civil society champions who can leverage their status and networks to draw international attention to the issues and maintain pressure.
  • Building relationships with sympathetic members of government is key.
  • Targeted online campaigns can help amplify the message of marginalized communities offline to achieve impact (More)”

Git for Law Revisited


Ari Hershowitz at Linked Legislation: “Laws change. Each time a new U.S. law is enacted, it enters a backdrop of approximately 22 million words of existing law. The new law may strike some text, add some text, and make other adjustments that trickle through the legal corpus. Seeing these changes in context would help lawmakers and the public better understand their impact.

To software engineers, this problem sounds like a perfect application for automated change management. Input an amendment, output tracked changes (see sample below). In the ideal system such changes could be seen as soon as the law is enacted — or even while a bill is being debated. We are now much closer to this ideal.

Changes to 16 U.S.C. 3835 by law 113-79

On Quora, on this blog, and elsewhere, I’ve discussed some of the challenges to using git, an automated change management system, to track laws. The biggest technical challenge has been that most laws, and most amendments to those laws, have not been structured in a computer friendly way. But that is changing.

The Law Revision Counsel (LRC) compiles the U.S. Code, through careful analysis of new laws, identifying the parts of existing law that will be changed (in a process called Classification), and making those changes by hand. The drafting and revision process takes great skill and legal expertise.

So, for example, the LRC makes changes to the current U.S. Code, following the language of a law such as this one:

Sample provision, 113-79 section 2006(a)

LRC attorneys identify the affected provisions of the U.S. Code and then carry out each of these instructions (strike “The Secretary”, insert “During fiscal year”…”). Since 2011, the LRC is using and publishing the final result of this analysis in XML format. One of the consequences of this format change is that it becomes feasible to automatically match the “before” to the “after” text, and produce a redlined version as seen above, showing the changes in context.

To produce this redlined version, I ran xml_diff, an open-source program written by Joshua Tauberer of govtrack.us, who also works with my company, Xcential, on modernization projects for the U.S. House. The results can be remarkably accurate. As a pre-requisite, it is necessary to have a “before” and “after” version in XML format and a small enough stretch of text to make the comparison manageable….(More)”