The Next 5 Years in Open Data: 3 Key Trends to Watch


Kevin Merritt (Socrata Inc.) at GovTech:2014 was a pivotal year in the evolution of open data for one simple and powerful reason – it went mainstream and was widely adopted on just about every continent. Open data is now table stakes. Any government that is not participating in open data is behind its peers…The move toward data-driven government will absolutely accelerate between 2015 and 2020, thanks to three key trends.

1. Comparative Analytics for Government Employees

The first noteworthy trend that will drive open data change in 2015 is that open data technology offerings will deliver first-class benefits to public-sector employees. This means government employees will be able to derive enormous insights from their own data and act on them in a deep, meaningful and analytical way. Until only recently, the primary beneficiaries of open data initiatives were external stakeholders: developers and entrepreneurs; scientists, researchers, analysts, journalists and economists; and ordinary citizens lacking technical training. The open data movement, until now, has ignored an important class of stakeholders – government employees….

2. Increased Global Expansion for Open Data

The second major trend fueling data-driven government is that 2015 will be a year of accelerating adoption of open data internationally.
Right now, for example, open data is being adopted prolifically in Europe, Latin America, Australia, New Zealand and Canada.
….
We will continue to see international governments adopt open data in 2015 for a variety of reasons. Northern European governments, for instance, are interested in efficiency and performance right now; Southern European governments, on the other hand, are currently focused on transparency, trust, and credibility. Despite the different motivations, the open data technology solutions are the same. And, looking out beyond 2015, it’s important to note that Southern European governments will also adopt open data to help increase job creation and improve delivery of services.

3. “Open Data” Will Simply Become “Government Data”

The third trend that we’ll see in the arena of open data lies a little further out on the horizon, and it will be surprising. In my opinion, the term “open data” may disappear within a decade; and in its place will simply be the term “government data.”
That’s because virtually all government data will be open data by 2020; and government data will be everywhere it needs to be – available to the public as fast as it’s created, processed and accumulated….(More).”

The Black Box Society


New book by Frank Pasquale on “The Secret Algorithms That Control Money and Information”: “Every day, corporations are connecting the dots about our personal behavior—silently scrutinizing clues left behind by our work habits and Internet use. The data compiled and portraits created are incredibly detailed, to the point of being invasive. But who connects the dots about what firms are doing with this information? The Black Box Society argues that we all need to be able to do so—and to set limits on how big data affects our lives.
Hidden algorithms can make (or ruin) reputations, decide the destiny of entrepreneurs, or even devastate an entire economy. Shrouded in secrecy and complexity, decisions at major Silicon Valley and Wall Street firms were long assumed to be neutral and technical. But leaks, whistleblowers, and legal disputes have shed new light on automated judgment. Self-serving and reckless behavior is surprisingly common, and easy to hide in code protected by legal and real secrecy. Even after billions of dollars of fines have been levied, underfunded regulators may have only scratched the surface of this troubling behavior….(More).”

The Participatory Approach to Open Data


at the SmartChicagoCollaborative: “…Having vast stores of government data is great, but to make this data useful – powerful – takes a different type of approach. The next step in the open data movement will be about participatory data.

Systems that talk back

One of the great advantages behind Chicago’s 311 ServiceTracker is that when you submit something to the system, the system has the capacity to talk back giving you a tracking number and an option to get email updates about your request. What also happens is that as soon as you enter your request, the data get automatically uploaded into the city’s data portal giving other 311 apps like SeeClickFix and access to the information as well…

Participatory Legislative Apps

We already see a number of apps that allow user to actively participate using legislative data.
At the Federal level, apps like PopVox allow users to find and track legislation that’s making it’s way through Congress. The app then allows users to vote if they approve or disapprove of a particular bill. You can then send explain your reasoning in a message that will be sent to all of your elected officials. The app makes it easier for residents to send feedback on legislation by creating a user interface that cuts through the somewhat difficult process of keeping tabs on legislation.
At the state level, New York’s OpenLegislation site allows users to search for state legislation and provide commentary on each resolution.
At the local level, apps like Councilmatic allows users to post comments on city legislation – but these comments aren’t mailed or sent to alderman the same way PopVox does. The interaction only works if the alderman are also using Councilmatic to receive feedback…

Crowdsourced Data

Chicago has hardwired several datasets into their computer systems, meaning that this data is automatically updated as the city does the people’s business.
But city governments can’t be everywhere at once. There are a number of apps that are designed to gather information from residents to better understand what’s going on their cities.
In Gary, the city partnered with the University of Chicago and LocalData to collect information on the state of buildings in Gary, IN. LocalData is also being used in Chicago, Houston, and Detroit by both city governments and non-profit organizations.
Another method the City of Chicago has been using to crowdsource data has been to put several of their datasets on GitHub and accept pull requests on that data. (A pull request is when one developer makes a change to a code repository and asks the original owner to merge the new changes into the original repository.) An example of this is bikers adding private bike rack locations to the city’s own bike rack dataset.

Going from crowdsourced to participatory

Shareabouts is a mapping platform by OpenPlans that gives city the ability to collect resident input on city infrastructure. Chicago’s Divvy Bikeshare program is using the tool to collect resident feedback on where the new Divvy stations should go. The app allows users to comment on suggested locations and share the discussion on social media.
But perhaps the most unique participatory app has been piloted by the City of South Bend, Indiana. CityVoice is a Code for America fellowship project designed to get resident feedback on abandoned buildings in South Bend…. (More)”

Three Things that a Good Infographic Needs


Antonio Salazar at BBVA Open Mind: “…A good infographic can save lives. Or almost, like the map drawn by Dr. Snow, who in 1854 tried to prove —with not much success, it’s true— that cholera spread by water. Not all graphics can be so ambitious, but when they hope to become a very powerful vehicle for knowledge, these are three elements that an excellent infographic should have. In order:

1. Rigor

This is taken for granted. Or not. This chart from Pravda (not Miuccia, the Soviet daily) is manipulated graphically to give the impression of growth:
OpenMind-antonio-salazar-infografia-tips-2
But rigor does not only mean accuracy, it means respect for the data in a broader sense. Letting them speak beyond the ornamentation, avoiding the “ducks” referred to by Edward R. Tufte.  Guiding our eye so the reader can discover them one step at a time. This brings us to the second quality:

2. Depth

The beauty of the data sometimes lies in their implications and their connections. This is why a good  infographic should be understandable at first sight, but also, when viewed for a second time, enable the reader to take pleasure in the data. Perhaps in the process of reasoning, both analytical and visual, that a graphic requires, we may come to a conclusion we hadn’t imagined. This is why a good infographic should have layers of information. Any infographic by Francesco Franchi is a wonder of visual richness and provides hours of entertainment and learning.

OpenMind-antonio-salazar-infografia-tips-3

Infographic by Francesco Franchi and Alessandro Giberti on Chinese imports published in IL

3. Narration

Storytelling in today’s parlance. When time, space and even meteorology are combined skillfully to tell a story through the data….”

Big Data in Action for Development


New report by the Worldbank: “Data provide critical inputs in designing effective development policy recommendations, supporting their implementation, and evaluating results. In this new report “Big Data in Action for Development,” the World Bank Group collaborated with Second Muse, a global innovation agency, to explore big data’s transformative potential for socioeconomic development.  The report develops a conceptual framework to work with big data in the development sector and presents a variety of case studies that lay out big data’s innovations, challenges, and opportunities.”

Democratizing Inequalities: Dilemmas of the New Public Participation


New book edited by Caroline W. Lee, Michael McQuarrie and Edward T. Walker: “Opportunities to “have your say,” “get involved,” and “join the conversation” are everywhere in public life. From crowdsourcing and town hall meetings to government experiments with social media, participatory politics increasingly seem like a revolutionary antidote to the decline of civic engagement and the thinning of the contemporary public sphere. Many argue that, with new technologies, flexible organizational cultures, and a supportive policymaking context, we now hold the keys to large-scale democratic revitalization.
Democratizing Inequalities shows that the equation may not be so simple. Modern societies face a variety of structural problems that limit potentials for true democratization, as well as vast inequalities in political action and voice that are not easily resolved by participatory solutions. Popular participation may even reinforce elite power in unexpected ways. Resisting an oversimplified account of participation as empowerment, this collection of essays brings together a diverse range of leading scholars to reveal surprising insights into how dilemmas of the new public participation play out in politics and organizations. Through investigations including fights over the authenticity of business-sponsored public participation, the surge of the Tea Party, the role of corporations in electoral campaigns, and participatory budgeting practices in Brazil, Democratizing Inequalities seeks to refresh our understanding of public participation and trace the reshaping of authority in today’s political environment.”

Businesses dig for treasure in open data


Lindsay Clark in ComputerWeekly: “Open data, a movement which promises access to vast swaths of information held by public bodies, has started getting its hands dirty, or rather its feet.
Before a spade goes in the ground, construction and civil engineering projects face a great unknown: what is down there? In the UK, should someone discover anything of archaeological importance, a project can be halted – sometimes for months – while researchers study the site and remove artefacts….
During an open innovation day hosted by the Science and Technologies Facilities Council (STFC), open data services and technology firm Democrata proposed analytics could predict the likelihood of unearthing an archaeological find in any given location. This would help developers understand the likely risks to construction and would assist archaeologists in targeting digs more accurately. The idea was inspired by a presentation from the Archaeological Data Service in the UK at the event in June 2014.
The proposal won support from the STFC which, together with IBM, provided a nine-strong development team and access to the Hartree Centre’s supercomputer – a 131,000 core high-performance facility. For natural language processing of historic documents, the system uses two components of IBM’s Watson – the AI service which famously won the US TV quiz show Jeopardy. The system uses SPSS modelling software, the language R for algorithm development and Hadoop data repositories….
The proof of concept draws together data from the University of York’s archaeological data, the Department of the Environment, English Heritage, Scottish Natural Heritage, Ordnance Survey, Forestry Commission, Office for National Statistics, the Land Registry and others….The system analyses sets of indicators of archaeology, including historic population dispersal trends, specific geology, flora and fauna considerations, as well as proximity to a water source, a trail or road, standing stones and other archaeological sites. Earlier studies created a list of 45 indicators which was whittled down to seven for the proof of concept. The team used logistic regression to assess the relationship between input variables and come up with its prediction….”

The Emerging Science of Human-Data Interaction


Emerging Technology From the arXiv: “The rapidly evolving ecosystems associated with personal data is creating an entirely new field of scientific study, say computer scientists. And this requires a much more powerful ethics-based infrastructure….
Now Richard Mortier at the University of Nottingham in the UK and a few pals say the increasingly complex, invasive and opaque use of data should be a call to arms to change the way we study data, interact with it and control its use. Today, they publish a manifesto describing how a new science of human-data interaction is emerging from this “data ecosystem” and say that it combines disciplines such as computer science, statistics, sociology, psychology and behavioural economics.
They start by pointing out that the long-standing discipline of human-computer interaction research has always focused on computers as devices to be interacted with. But our interaction with the cyber world has become more sophisticated as computing power has become ubiquitous, a phenomenon driven by the Internet but also through mobile devices such as smartphones. Consequently, humans are constantly producing and revealing data in all kinds of different ways.
Mortier and co say there is an important distinction between data that is consciously created and released such as a Facebook profile; observed data such as online shopping behaviour; and inferred data that is created by other organisations about us, such as preferences based on friends’ preferences.
This leads the team to identify three key themes associated with human-data interaction that they believe the communities involved with data should focus on.
The first of these is concerned with making data, and the analytics associated with it, both transparent and comprehensible to ordinary people. Mortier and co describe this as the legibility of data and say that the goal is to ensure that people are clearly aware of the data they are providing, the methods used to draw inferences about it and the implications of this.
Making people aware of the data being collected is straightforward but understanding the implications of this data collection process and the processing that follows is much harder. In particular, this could be in conflict with the intellectual property rights of the companies that do the analytics.
An even more significant factor is that the implications of this processing are not always clear at the time the data is collected. A good example is the way the New York Times tracked down an individual after her seemingly anonymized searches were published by AOL. It is hard to imagine that this individual had any idea that the searches she was making would later allow her identification.
The second theme is concerned with giving people the ability to control and interact with the data relating to them. Mortier and co describe this as “agency”. People must be allowed to opt in or opt out of data collection programs and to correct data if it turns out to be wrong or outdated and so on. That will require simple-to-use data access mechanisms that have yet to be developed
The final theme builds on this to allow people to change their data preferences in future, an idea the team call “negotiability”. Something like this is already coming into force in the European Union where the Court of Justice has recently begun to enforce the “right to be forgotten”, which allows people to remove information from search results under certain circumstances….”
Ref: http://arxiv.org/abs/1412.6159  Human-Data Interaction: The Human Face of the Data-Driven Society

Can Business And Tech Transform The Way Our Government Works By 2020?


Ben Schiller at Co.Exist: “The rise of open data, crowd-sourcing, predictive analytics, and other big tech trends, aren’t just for companies to contend with. They’re also a challenge for government. New technology gives public agencies the opportunity to develop and deliver services in new ways, track results more accurately, and open up decision-making.
Deloitte’s big new Government 2020 report looks at the trends impacting government and lays out a bunch of ideas for how they can innovate. We picked out a few below. There are more infographics in the slide show.

Consumerization of public services

Deloitte expects entrepreneurs to “develop innovative and radically user-friendly approaches to satisfy unmet consumer demand for better public services.” Startups like Uber or Lyft “reinvigorated transportation.” Now it expects a similar “focus on seamless customer experiences” in education and health care.

Open workforce

Deloitte expects governments to become looser: collections of people doing a job, rather than large hierarchical structures. “Governments [will] expand their talent networks to include ‘partnership talent’ (employees who are parts of joint ventures), ‘borrowed talent’ (employees of contractors), ‘freelance talent’ (independent, individual contractors) and ‘open-source talent,'” the report says.

Outcome based legislation

Just as big data analytics allows companies to measure the effectiveness of marketing campaigns, so it allows governments to measure how well legislation and regulation is working. They can “shift from a concentration on processes to the achievement of specific targets.” And, if the law isn’t working, someone has the data to throw it out….”

Data is Law


Mark Headd at Civic Innovations: The Future is Open: “In his famous essay on the importance of the technological underpinnings of the Internet, Lawrence Lessig described the potential threat if the architecture of cyberspace was built on values that diverged from those we believe are important to the proper functioning of our democracy. The central point of this seminal work seems to grow in importance each day as technology and the Internet become more deeply embedded into our daily lives.
But increasingly, another kind of architecture is becoming central to the way we live and interact with each other – and to the way in which we are governed and how we interact with those that govern us. This architecture is used by governments at the federal, state and local level to share data with the public.
This data – everything from weather data, economic data, education data, crime data, environmental data – is becoming increasingly important for how we view the world around us and our perception of how we are governed. It is quite easy for us to catalog the wide range of personal decisions – some rote, everyday decisions like what to wear based on the weather forecast, and some much more substantial like where to live or where to send our children to school – that are influenced by data collected, maintained or curated by government.
It seems to me that Lessig’s observations from a decade and a half ago about the way in which the underlying architecture of the Internet may affect our democracy can now be applied to data. Ours is the age of data – it pervades every aspect of our lives and influences how we raise our children, how we spend our time and money and who we elect to public office.
But even more fundamental to our democracy, how well our government leaders are performing the job we empower them to do depends on data. How effective is policing in reducing the number of violent crimes? How effective are environmental regulations in reducing dangerous emissions? How well are programs performing to lift people out of poverty and place them in gainful employment? How well are schools educating our children?
These are all questions that we answer – in whole or in part – by looking at data. Data that governments themselves are largely responsible for compiling and publishing….
Having access to open data is no longer an option for participating effectively in our modern democracy, it’s a requirement. Data – to borrow Lessig’s argument – has become law.”