Opportunities and Challenges of Policy Informatics: Tackling Complex Problems through the Combination of Open Data, Technology and Analytics


Gabriel Puron-Cid et al in the International Journal on Public Administration in the Digital Age: “Contemporary societies face complex problems that challenge the sustainability of their social and economic systems. Such problems may require joint efforts from the public and private sectors as well as from the society at large in order to find innovative solutions. In addition, the open government movement constitutes a revitalized wave of access to data to promote innovation through transparency, participation and collaboration. This paper argues that currently there is an opportunity to combine emergent information technologies, new analytical methods, and open data in order to develop innovative solutions to some of the pressing problems in modern societies. Therefore, the objective is to propose a conceptual model to better understand policy innovations based on three pillars: data, information technologies, and analytical methods and techniques. The potential benefits generated from the creation of organizations with advanced analytical capabilities within governments, universities, and non-governmental organizations are numerous and the expected positive impacts on society are significant. However, this paper also discusses some important political, organizational, and technical challenges…(More).

 

#IAmAResearchParasite


Marcia McNutt in Science: “In the midst of steady progress in policies for data sharing, a recent editorial expressed a contrarian view.* The authors described the concern of some scientists about the rise of an underclass of “research parasites” who exploit data sets that are collected and curated by others. Even worse, these parasites might use such data to try to disprove the conclusions posited in the data’s original source studies. The editorial raised the points of how anyone not involved in the original study could use the data without misrepresenting it, and the danger of perhaps arriving at erroneous conclusions. The editorial advised instead that data sharing be implemented by involving the authors of the original study as coauthors in follow-up research. The research community immediately took to Twitter under the hashtag #IAmAResearchParasite to voice opposition to the editorial.

Much of what we know about the large-scale features of this planet is apparent thanks to widespread data-sharing practices and the early establishment of data banks in the geosciences. Aspects such as determining the shape of the ocean floor, ocean chemistry, the internal structure of Earth’s deep interior, the physics and chemistry of the atmosphere, and many other topics could not have been ascertained from a single investigator’s field program…

There are costs to implementing data reuse, but there are also costs for irreproducible research and for recollecting data for new uses. And no amount of funding can reconstruct lost ephemeral or time-dependent phenomena for which the data were not well curated. No more excuses: Let’s step up to data sharing…(More)”

Do Universities, Research Institutions Hold the Key to Open Data’s Next Chapter


Ben Miller at Government Technology: “Government produces a lot of data — reams of it, roomfuls of it, rivers of it. It comes in from citizen-submitted forms, fleet vehicles, roadway sensors and traffic lights. It comes from utilities, body cameras and smartphones. It fills up servers and spills into the cloud. It’s everywhere.

And often, all that data sits there not doing much. A governing entity might have robust data collection and it might have an open data policy, but that doesn’t mean it has the computing power, expertise or human capital to turn those efforts into value.

The amount of data available to government and the computing public promises to continue to multiply — the growing smart cities trend, for example, installs networks of sensors on everything from utility poles to garbage bins.

As all this happens, a movement — a new spin on an old concept — has begun to take root: partnerships between government and research institutes. Usually housed within universities and laboratories, these partnerships aim to match strength with strength. Where government has raw data, professors and researchers have expertise and analytics programs.

Several leaders in such partnerships, spanning some of the most tech-savvy cities in the country, see increasing momentum toward the concept. For instance, the John D. and Catherine T. MacArthur Foundation in September helped launch the MetroLab Network, an organization of more than 20 cities that have partnered with local universities and research institutes for smart-city-oriented projects….

Two recurring themes in projects that universities and research organizations take on in cooperation with government are project evaluation and impact analysis. That’s at least partially driven by the very nature of the open data movement: One reason to open data is to get a better idea of how well the government is operating….

Open data may have been part of the impetus for city-university partnerships, in that the availability of more data lured researchers wanting to work with it and extract value. But those partnerships have, in turn, led to government officials opening more data than ever before for useful applications.

Sort of.

“I think what you’re seeing is not just open data, but kind of shades of open — the desire to make the data open to university researchers, but not necessarily the broader public,” said Beth Noveck, co-founder of New York University’s GovLab.


shipping+crates

GOVLAB: DOCKER FOR DATA 

Much of what GovLab does is about opening up access to data, and that is the whole point of Docker for Data. The project aims to simplify and quicken the process of extracting and loading large data sets so they will respond to Structured Query Language commands by moving the computing power of that process to the cloud. The docker can be installed with a single line of code, and its website plays host to already-extracted data sets. Since its inception, the website has grown to include more than 100 gigabytes of data from more than 8,000 data sets. From Baltimore, for example, one can easily find information on public health, water sampling, arrests, senior centers and more. Photo via Shutterstock.


That’s partially because researchers are a controlled group who can be forced to sign memorandums of understanding and trained to protect privacy and prevent security breaches when government hands over sensitive data. That’s a top concern of agencies that manage data, and it shows in the GovLab’s work.

It was something Noveck found to be very clear when she started working on a project she simply calls “Arnold” because of project support from the Laura and John Arnold Foundation. The project involves building a better understanding of how different criminal justice jurisdictions collect, store and share data. The motivation is to help bridge the gaps between people who manage the data and people who should have easy access to it. When Noveck’s center conducted a survey among criminal justice record-keepers, the researchers found big differences between participants.

“There’s an incredible disparity of practices that range from some jurisdictions that have a very well established, formalized [memorandum of understanding] process for getting access to data, to just — you send an email to a guy and you hope that he responds, and there’s no organized way to gain access to data, not just between [researchers] and government entities, but between government entities,” she said….(More)

Value public information so we can trust it, rely on it and use it


Speech by David Fricker, the director general of the National Archives of Australia: “No-one can deny that we are in an age of information abundance. More and more we rely on information from a variety of sources and channels. Digital information is seductive, because it’s immediate, available and easy to move around. But digital information can be used for nefarious purposes. Social issues can be at odds with processes of government in this digital age. There is a tension between what is the information, where it comes from and how it’s going to be used.

How do we know if the information has reached us without being changed, whether that’s intentional or not?

How do we know that government digital information will be the authoritative source when the pace of information exchange is so rapid? In short, how do we know what to trust?

“It’s everyone’s responsibly to contribute to a transparent government, and that means changes in our thinking and in our actions.”

Consider the challenges and risks that come with the digital age: what does it really mean to have transparency and integrity of government in today’s digital environment?…

What does the digital age mean for government? Government should be delivering services online, which means thinking about location, timeliness and information accessibility. It’s about getting public-sector data out there, into the public, making it available to fuel the digital economy. And it’s about a process of change across government to make sure that we’re breaking down all of those silos, and the duplication and fragmentation which exist across government agencies in the application of information, communications, and technology…..

The digital age is about the digital economy, it’s about rethinking the economy of the nation through the lens of information that enables it. It’s understanding that a nation will be enriched, in terms of culture life, prosperity and rights, if we embrace the digital economy. And that’s a weighty responsibility. But the responsibility is not mine alone. It’s a responsibility of everyone in the government who makes records in their daily work. It’s everyone’s responsibly to contribute to a transparent government. And that means changes in our thinking and in our actions….

What has changed about democracy in the digital age? Once upon a time if you wanted to express your anger about something, you might write a letter to the editor of the paper, to the government department, or to your local member and then expect some sort of an argument or discussion as a response. Now, you can bypass all of that. You might post an inflammatory tweet or blog, your comment gathers momentum, you pick the right hashtag, and off we go. It’s all happening: you’re trending on Twitter…..

If I turn to transparency now, at the top of the list is the basic recognition that government information is public information. The information of the government belongs to the people who elected that government. It’s a fundamental of democratic values. It also means that there’s got to be more public participation in the development of public policy, which means if you’re going to have evidence-based, informed, policy development; government information has to be available, anywhere, anytime….

Good information governance is at the heart of managing digital information to provide access to that information into the future — ready access to government information is vital for transparency. Only when information is digital and managed well can government share it effectively with the Australian community, to the benefit of society and the economy.

There are many examples where poor information management, or poor information governance, has led to failures — both in the private and public sectors. Professor Peter Shergold’s recent report, Learning from Failure, why large government policy initiatives have gone so badly wrong in the past and how the chances of success in the future can be improved, highlights examples such as the Home Insulation Program, the NBN and Building the Education Revolution….(Full Speech)

Ebola: A Big Data Disaster


Study by Sean Martin McDonald: “…undertaken with support from the Open Society Foundation, Ford Foundation, and Media Democracy Fund, explores the use of Big Data in the form of Call Detail Record (CDR) data in humanitarian crisis.

It discusses the challenges of digital humanitarian coordination in health emergencies like the Ebola outbreak in West Africa, and the marked tension in the debate around experimentation with humanitarian technologies and the impact on privacy. McDonald’s research focuses on the two primary legal and human rights frameworks, privacy and property, to question the impact of unregulated use of CDR’s on human rights. It also highlights how the diffusion of data science to the realm of international development constitutes a genuine opportunity to bring powerful new tools to fight crisis and emergencies.

Analysing the risks of using CDRs to perform migration analysis and contact tracing without user consent, as well as the application of big data to disease surveillance is an important entry point into the debate around use of Big Data for development and humanitarian aid. The paper also raises crucial questions of legal significance about the access to information, the limitation of data sharing, and the concept of proportionality in privacy invasion in the public good. These issues hold great relevance in today’s time where big data and its emerging role for development, involving its actual and potential uses as well as harms is under consideration across the world.

The paper highlights the absence of a dialogue around the significant legal risks posed by the collection, use, and international transfer of personally identifiable data and humanitarian information, and the grey areas around assumptions of public good. The paper calls for a critical discussion around the experimental nature of data modelling in emergency response due to mismanagement of information has been largely emphasized to protect the contours of human rights….

See Sean Martin McDonald – “Ebola: A Big Data Disaster” (PDF).

 

A machine intelligence commission for the UK


Geoff Mulgan at NESTA: ” This paper makes the case for creating a Machine Intelligence Commission – a new public institution to help the development of new generations of algorithms, machine learning tools and uses of big data, ensuring that the public interest is protected.

I argue that new institutions of this kind – which can interrogate, inspect and influence technological development – are a precondition for growing informed public trust. That trust will, in turn, be essential if we are to reap the full potential public and economic benefits from new technologies. The proposal draws on lessons from fields such as human fertilisation, biotech and energy, which have shown how trust can be earned, and how new industries can be grown.  It also draws on lessons from the mistakes made in fields like GM crops and personal health data, where lack of trust has impeded progress….(More)”

Meet your Matchmaker: New crowdsourced sites for rare diseases


Carina Storrs at CNN: “Angela’s son Jacob was born with a number of concerning traits. He had an extra finger, and a foot and hip that were abnormally shaped. The doctors called in geneticists to try to diagnose his unusual condition. “That started our long, 12-year journey,” said Angela, who lives in the Baltimore area.

As geneticists do, they studied Jacob’s genes, looking for mutations in specific regions of the genome that could point to a problem. But there were no leads.

In the meantime, Jacob developed just about every kind of health problem there is. He has cognitive delays, digestive problems, muscle weakness, osteoporosis and other ailments.

“It was extremely frustrating, it was like being on a roller coaster. You wait six to eight weeks for the (gene) test and then it comes back as showing nothing,” recalled Angela, who asked that their last name not be used to protect her son’s privacy. “How do we go about treating until we get at what it is?”

Finally a test last year, which was able to take a broad look at all of Jacob’s genes, revealed a possible genetic culprit, but it still did not shed any light on his condition. “Nothing was known about the gene,” said Dr. Antonie Kline, director of pediatric genetics at the Greater Baltimore Medical Center, who had been following Jacob since birth.

Fortunately, Kline knew about an online program called GeneMatcher, which launched in December 2013. It would allow her to enter the new mystery gene into a database and search for other clinicians in the world who work with patients who have mutations in the same gene….

the search for “someone else on the planet” can be hard, Hamosh said. The diseases in GeneMatcher are rare, affecting fewer than 200,000 people in the United States, and it can be difficult for clinicians with similar patients to find each other just through word of mouth and professional connections. Au, the Canadian researcher with a patient similar to Jacob, is actually a friend of Kline’s, but the two had never realized their patients’ similarities.

It was not just Hamosh and her colleagues who were struck by the need for something like GeneMatcher. At the same time they were developing their program, researchers in Canada and the UK were creating PhenomeCentral and Decipher, respectively.

The three are collectively known as matchmaker programs. They connect patients with rare diseases which clinicians may never have seen before. In the case of PhenomeCentral, however, clinicians do not have to have a genetic culprit and can search only for other patients with similar traits or symptoms.

In the summer of 2015, it got much easier for clinicians all over the world to use these programs, when a clearinghouse site called Matchmaker Exchange was launched. They can now enter the patient information one time and search all three databases….(More)

Facebook Is Making a Map of Everyone in the World


Robinsion Meyer at The Atlantic: “Americans inhabit an intricately mapped world. Type “Burger King” into an online box, and Google will cough up a dozen nearby options, each keyed to a precise latitude and longitude.

But throughout much of the world, local knowledge stays local. While countries might conduct censuses, the data doesn’t go much deeper than the county or province level.

Take population data, for instance: More than 7.4 billion humans sprawl across this planet of ours. They live in dense urban centers, in small towns linked by farms, and alone on the outskirts of jungles. But no one’s sure where, exactly, many of them live.

Now, Facebook says it has mapped almost 2 billion people better than any previous project. The company’s Connectivity Labs announced this week that it created new, high-resolution population-distribution maps of 20 countries, most of which are developing. It won’t release most of the maps until later this year,but if they’re accurate, they will be the best-quality population maps ever made for most of those places.

The maps will be notable for another reason, too: If they’re accurate, they ‘ll signal the arrival of a new, AI-aided age of cartography.

In the rich world, reliable population information is taken for granted.  But elsewhere, population-distribution maps have dozens of applications in different fields. Urban planners need to estimate city density so they can place and improve roads. Epidemiologists and public-health workers use them to track outbreaks or analyze access to health care. And after a disaster, population maps can be used (along with crisis mapping) to prioritize where emergency aid gets sent….(More)

Big Data Visualization: Review of 20 Tools


Edoardo L’Astorina at BluFrame: “Big Data is amazing. It describes our everyday behavior, keeps track of the places we go, stores what we like to do and how much time we spend doing our favorite activities.

Big Data is made of numbers, and I think we all agree when we say: Numbers are difficult to look at. Enter Big Data visualization….Data visualization lets you interact with data. It goes beyond analysis. Visualization brings a presentation to life. It keeps your audience’s eyes on the screen. And gets people interested….

We made everything easy for you and prepared a series of reviews that cover all the features of the best data visualization tools out there. And we divided our reviews in two sections: data visualization tools for presentations and data visualization tools for developers.

Here are reviews of our 20 best tools for Big Data visualization.

Data Visualization Tools for Presentations: Zero Coding Required:…

Tableau.. is the big data visualization tool for corporate. Tableau lets you create charts, graphs, maps and many other graphics. A desktop app is available for visual analytics….

Infogram…lets you link their visualizations and infographics to real time big data…

ChartBlocks… is an easy-to-use online tool that requires no coding, and builds visualizations from spreadsheets, databases… and live feeds….

Datawrapper.. is aimed squarely at publishers and journalists…

Plotly…will help you create a sharp and slick chart in just a few minutes, starting from a simple spreadsheet….

RAW… boasts on its homepage to be “the missing link between spreadsheets and vector graphics”….

Visual.ly… is a visual content service….

Data Visualization Tools for Developers: JavaScript libraries

D3.js…runs on JavaScript and uses HTML, CSS and SVG. D3.js is open-source and applies data-driven transformation to a webpage and – as you can see from their examples – allows for beautiful and fast visualizations….

Ember Charts is – as the name suggests – based on the Ember.js framework and uses D3.js under the hood….

NVD3…runs on top of D3.js –surprise surprise– and aims to build re-usable charts and components….

Google Charts… runs on HTML5 and SVG and aims at Android, iOS and total cross-browser compatibility, including older Internet Explorer versions supported via VML

FusionCharts is – according to their site – the most comprehensive JavaScript charting library, and includes over 90 charts and 900 maps….

Highcharts…is a JavaScript API that integrates easily with jQuery and boasts being used by 61 out of the world’s 100 largest companies….

Chart.js…For a small chart project, Chart.js is your go-to place….

Leaflet… leveragesOpenStreetMap data and adds HTML5/CSS3 visualizations and interactivity on top to ensure everything is responsive and mobile ready….

Chartist.js.. is born out of a community effort to blow all other JavaScript charting libraries out of the water…

n3-charts…is for the AngularJS lovers out there….

Sigma JS…is what you want for interactivity….

Polymaps…visualizes…. you guessed it: maps….

Processing.js …is a JavaScript library that sits on top of the Processing visual programming language…(More)

The impact of a move towards Open Data in West Africa


 at the Georgetown Journal of International Affairs:  “The concept of “open data” is not new, but its definition is quite recent. Since computers began communicating through networks, engineers have been developing standards to share data. The open data philosophy holds that some data should be freely available for use, reuse, distribute and publish without copyright and patent controls. Several mechanisms can also limit access to data like restricted database access, use of proprietary technologies or encryption. Ultimately, open data buttresses government initiatives to boost innovation, support transparency, empower citizens, encourage accountability, and fight corruption.

West Africa is primed for open data. The region experienced a 6% growth in 2014, according to the Africa Development Bank. Its Internet user network is also growing: 17% of the sub-Saharan population owned a unique smartphone in 2013, a number projected to grow to 37% by 2020 according to the GSMA. To improve the quality of governance and services in the digital age, the region must develop new infrastructures, revise digital strategies, simplify procurement procedures, adapt legal frameworks, and allow access to public data. Open data can enhance local economies and the standard of living.

This paper speaks towards the impact of open data in West Africa. First it assesses open data as a positive tool for governance and civil society. Then, it analyzes the current situation of open data across the region. Finally, it highlights specific best practices for enhancing impact in the future….(More)”