Crowdsourced pollution data via smartphones


Springwise: “Citizens in eleven cities in Europe were recently recruited to help crowdsource pollution measurements, as part of the large-scale research project iSPEX-EU. Participants used their smartphones, an app and a lens called a spectropolarimeter, to collect data about air quality across the continent, which will be used by iSPEX to make comprehensive maps.

The project ran for six weeks and saw thousands of measurements taken in Athens, Barcelona, Belgrade, Berlin, Copenhagen, London, Manchester, Milan, Rome, and Toulouse. To contribute, citizens registered their interest, downloaded the free app and were sent an iSPEX lens. Then, on a clear day they placed the lens over their smartphone camera and photographed the sky in multiple directions. The app registered the location and direction of each picture and measured the light spectrum and the polarization of the light.
From the data, iSPEX are able to calculate how much fine dust — known as aerosols — there is in the atmosphere in that place and create a map showing levels of air pollution across Europe. The crowdsourced data can be used to aid government research by filling in any blank spaces and ensuring that the official data is honest.

We’ve seen attempts at similar projects before, such asSmart Citizen, but iSPEX EU benefits from the flexibility and simplicity of its tools. Smartphones have been successfully harnessed as scientific apparatus, enabling researchers to crowdsource data about issues including cancer and tree disease….(More)”

New traffic app and disaster prevention technology road tested


Psych.org: “A new smartphone traffic app tested by citizens in Dublin, Ireland allows users to give feedback on traffic incidents, enabling traffic management centres to respond quicker when collisions and other incidents happen around the city. The ‘CrowdAlert’ app, which is now available for download, is one of the key components utilised in the EU-funded INSIGHT project and a good example of how smartphones and social networks can be harnessed to improve public services and safety.

‘We are witnessing an explosion in the quantity, quality, and variety of available information, fuelled in large part by advances in sensor networking, the availability of low-cost sensor-enabled devices and by the widespread adoption of powerful smart-phones,’ explains  coordinator professor Dimitrios Gunopulos from the National and Kapodistrian University of Athens. ‘These revolutionary technologies are driving the development and adoption of applications where mobile devices are used for continuous data sensing and analysis.’

The project also developed a novel citywide real-time traffic monitoring tool, the ‘INSIGHT System’, which was tested in real conditions in the Dublin City control room, along with nationwide disaster monitoring technologies. The INSIGHT system was shown to provide early warnings to experts at situation centres, enabling them to monitor situations in real-time, including disasters with potentially nation-wide impacts such as severe weather conditions, floods and subsequent knock-on events such as fires and power outages.

The project’s results will be of interest to public services, which have until now lacked the necessary infrastructure for handling and integrating miscellaneous data streams, including data from static and mobile sensors as well as information coming from social network sources, in real-time. Providing cities with the ability to manage emergency situations with enhanced capabilities will also open up new markets for network technologies….(More)”

Introducing Government as a Platform


Peter Williams, Jan Gravesen and Trinette Brownhill in Government Executive: “Governments around the world are facing competitive pressures and expectations from their constituents that are prompting them to innovate and dissolve age-old structures. Many governments have introduced a digital strategy in which at least one of the goals is aimed at bringing their organizations closer to citizens and businesses.

To achieve this, ideally IT and data in government would not be constrained by the different functional towers that make up the organization, as is often the case. They would not be constrained by complex, monolithic application design philosophies and lengthy implementation cycles, nor would development be constrained by the assumption that all activity has to be executed by the government itself.

Instead, applications would be created rapidly and cheaply, and modules would be shared as reusable blocks of code and integrated data. It would be relatively straightforward to integrate data from multiple departments to enable a focus on the complex needs of, say, a single parent who is diabetic and a student. Delivery would be facilitated in the manner best required, or preferred, by the citizen. Third parties would also be able to access these modules of code and data to build higher value government services that multiple agencies would then buy into. The code would run on a cloud infrastructure that maximizes the efficiency in which processing resources are used.

GaaP an organized set of ideas and principles that allows organizations to approach these ideals. It allows governments to institute more efficient sharing of IT resources as well as unlock data and functionality via application programming interfaces to allow third parties to build higher value citizen services. In doing so, security plays a crucial role protecting the privacy of constituents and enterprise assets.

We see increasingly well-established examples of GaaP services in many parts of the world. The notion has significantly influenced strategic thinking in the UK, Australia, Denmark, Canada and Singapore. In particular, it has evolved in a deliberate way in the UK’s Government Data Services, building on the Blairite notion of “joined up government”; in Australia’s e-government strategy and its myGov program; and as a significant influencer in Singapore’s entire approach to building its “smarter nation” infrastructure.

Collaborative Government

GaaP assumes a transformational shift in efficiency, effectiveness and transparency, in which agencies move toward a collaborative government and away from today’s siloed approach. That collaboration may be among agencies, but also with other entities (nongovernmental organizations, the private sector, citizens, etc.).

GaaP’s focus on collaboration enables public agencies to move away from their traditional towered approach to IT and increasingly make use of shared and composable services offered by a common – usually a virtualized, cloud-enabled – platform. This leads to more efficient use of development resources, platforms and IT support. We are seeing examples of this already with a group of townships in New York state and also with two large Spanish cities that are embarking on this approach.

While efficient resource and service sharing is central to the idea of GaaP, it is not sufficient. The idea is that GaaP must allow app developers, irrespective of whether they are citizens, private organizations or other public agencies, to develop new value-added services using published government data and APIs. In this sense, the platform becomes a connecting layer between public agencies’ systems and data on the one hand, and private citizens, organizations and other public agencies on the other.

In its most fundamental form, GaaP is able to:

  • Consume data and government services from existing departmental systems.
  • Consume syndicated services from platform-as-a-service or software-as-a-service providers in the public marketplace.
  • Securely unlock these data and services and allow third parties –citizens, private organizations or other agencies – to combine services and data into higher-order services or more citizen-centric or business-centric services.

It is the openness, the secure interoperability, and the ability to compose new services on the basis of existing services and data that define the nature of the platform.

The Challenges

At one time, the challenge of creating a GaaP structure would have been technology: Today, it is governance….(More)”

What we can learn from the failure of Google Flu Trends


David Lazer and Ryan Kennedy at Wired: “….The issue of using big data for the common good is far more general than Google—which deserves credit, after all, for offering the occasional peek at their data. These records exist because of a compact between individual consumers and the corporation. The legalese of that compact is typically obscure (how many people carefully read terms and conditions?), but the essential bargain is that the individual gets some service, and the corporation gets some data.

What is left out that bargain is the public interest. Corporations and consumers are part of a broader society, and many of these big data archives offer insights that could benefit us all. As Eric Schmidt, CEO of Google, has said, “We must remember that technology remains a tool of humanity.” How can we, and corporate giants, then use these big data archives as a tool to serve humanity?

Google’s sequel to GFT, done right, could serve as a model for collaboration around big data for the public good. Google is making flu-related search data available to the CDC as well as select research groups. A key question going forward will be whether Google works with these groups to improve the methodology underlying GFT. Future versions should, for example, continually update the fit of the data to flu prevalence—otherwise, the value of the data stream will rapidly decay.

This is just an example, however, of the general challenge of how to build models of collaboration amongst industry, government, academics, and general do-gooders to use big data archives to produce insights for the public good. This came to the fore with the struggle (and delay) for finding a way to appropriately share mobile phone data in west Africa during the Ebola epidemic (mobile phone data are likely the best tool for understanding human—and thus Ebola—movement). Companies need to develop efforts to share data for the public good in a fashion that respects individual privacy.

There is not going to be a single solution to this issue, but for starters, we are pushing for a “big data” repository in Boston to allow holders of sensitive big data to share those collections with researchers while keeping them totally secure. The UN has its Global Pulse initiative, setting up collaborative data repositories around the world. Flowminder, based in Sweden, is a nonprofit dedicated to gathering mobile phone data that could help in response to disasters. But these are still small, incipient, and fragile efforts.

The question going forward now is how build on and strengthen these efforts, while still guarding the privacy of individuals and the proprietary interests of the holders of big data….(More)”

Researchers wrestle with a privacy problem


Erika Check Hayden at Nature: “The data contained in tax returns, health and welfare records could be a gold mine for scientists — but only if they can protect people’s identities….In 2011, six US economists tackled a question at the heart of education policy: how much does great teaching help children in the long run?

They started with the records of more than 11,500 Tennessee schoolchildren who, as part of an experiment in the 1980s, had been randomly assigned to high- and average-quality teachers between the ages of five and eight. Then they gauged the children’s earnings as adults from federal tax returns filed in the 2000s. The analysis showed that the benefits of a good early education last for decades: each year of better teaching in childhood boosted an individual’s annual earnings by some 3.5% on average. Other data showed the same individuals besting their peers on measures such as university attendance, retirement savings, marriage rates and home ownership.

The economists’ work was widely hailed in education-policy circles, and US President Barack Obama cited it in his 2012 State of the Union address when he called for more investment in teacher training.

But for many social scientists, the most impressive thing was that the authors had been able to examine US federal tax returns: a closely guarded data set that was then available to researchers only with tight restrictions. This has made the study an emblem for both the challenges and the enormous potential power of ‘administrative data’ — information collected during routine provision of services, including tax returns, records of welfare benefits, data on visits to doctors and hospitals, and criminal records. Unlike Internet searches, social-media posts and the rest of the digital trails that people establish in their daily lives, administrative data cover entire populations with minimal self-selection effects: in the US census, for example, everyone sampled is required by law to respond and tell the truth.

This puts administrative data sets at the frontier of social science, says John Friedman, an economist at Brown University in Providence, Rhode Island, and one of the lead authors of the education study “They allow researchers to not just get at old questions in a new way,” he says, “but to come at problems that were completely impossible before.”….

But there is also concern that the rush to use these data could pose new threats to citizens’ privacy. “The types of protections that we’re used to thinking about have been based on the twin pillars of anonymity and informed consent, and neither of those hold in this new world,” says Julia Lane, an economist at New York University. In 2013, for instance, researchers showed that they could uncover the identities of supposedly anonymous participants in a genetic study simply by cross-referencing their data with publicly available genealogical information.

Many people are looking for ways to address these concerns without inhibiting research. Suggested solutions include policy measures, such as an international code of conduct for data privacy, and technical methods that allow the use of the data while protecting privacy. Crucially, notes Lane, although preserving privacy sometimes complicates researchers’ lives, it is necessary to uphold the public trust that makes the work possible.

“Difficulty in access is a feature, not a bug,” she says. “It should be hard to get access to data, but it’s very important that such access be made possible.” Many nations collect administrative data on a massive scale, but only a few, notably in northern Europe, have so far made it easy for researchers to use those data.

In Denmark, for instance, every newborn child is assigned a unique identification number that tracks his or her lifelong interactions with the country’s free health-care system and almost every other government service. In 2002, researchers used data gathered through this identification system to retrospectively analyse the vaccination and health status of almost every child born in the country from 1991 to 1998 — 537,000 in all. At the time, it was the largest study ever to disprove the now-debunked link between measles vaccination and autism.

Other countries have begun to catch up. In 2012, for instance, Britain launched the unified UK Data Service to facilitate research access to data from the country’s census and other surveys. A year later, the service added a new Administrative Data Research Network, which has centres in England, Scotland, Northern Ireland and Wales to provide secure environments for researchers to access anonymized administrative data.

In the United States, the Census Bureau has been expanding its network of Research Data Centers, which currently includes 19 sites around the country at which researchers with the appropriate permissions can access confidential data from the bureau itself, as well as from other agencies. “We’re trying to explore all the available ways that we can expand access to these rich data sets,” says Ron Jarmin, the bureau’s assistant director for research and methodology.

In January, a group of federal agencies, foundations and universities created the Institute for Research on Innovation and Science at the University of Michigan in Ann Arbor to combine university and government data and measure the impact of research spending on economic outcomes. And in July, the US House of Representatives passed a bipartisan bill to study whether the federal government should provide a central clearing house of statistical administrative data.

Yet vast swathes of administrative data are still inaccessible, says George Alter, director of the Inter-university Consortium for Political and Social Research based at the University of Michigan, which serves as a data repository for approximately 760 institutions. “Health systems, social-welfare systems, financial transactions, business records — those things are just not available in most cases because of privacy concerns,” says Alter. “This is a big drag on research.”…

Many researchers argue, however, that there are legitimate scientific uses for such data. Jarmin says that the Census Bureau is exploring the use of data from credit-card companies to monitor economic activity. And researchers funded by the US National Science Foundation are studying how to use public Twitter posts to keep track of trends in phenomena such as unemployment.

 

….Computer scientists and cryptographers are experimenting with technological solutions. One, called differential privacy, adds a small amount of distortion to a data set, so that querying the data gives a roughly accurate result without revealing the identity of the individuals involved. The US Census Bureau uses this approach for its OnTheMap project, which tracks workers’ daily commutes. ….In any case, although synthetic data potentially solve the privacy problem, there are some research applications that cannot tolerate any noise in the data. A good example is the work showing the effect of neighbourhood on earning potential3, which was carried out by Raj Chetty, an economist at Harvard University in Cambridge, Massachusetts. Chetty needed to track specific individuals to show that the areas in which children live their early lives correlate with their ability to earn more or less than their parents. In subsequent studies5, Chetty and his colleagues showed that moving children from resource-poor to resource-rich neighbourhoods can boost their earnings in adulthood, proving a causal link.

Secure multiparty computation is a technique that attempts to address this issue by allowing multiple data holders to analyse parts of the total data set, without revealing the underlying data to each other. Only the results of the analyses are shared….(More)”

Algorithm predicts and prevents train delays two hours in advance


Springwise: “Transport apps such as Ototo make it easier than ever for passengers to stay informed about problems with public transport, but real-time information can only help so much — by the time users find out about a delayed service, it is often too late to take an alternative route. Now, Stockholmstag — the company that runs Sweden’s trains — have found a solution in the form of an algorithm called ‘The Commuter Prognosis’, which can predict network delays up to two hours in advance, giving train operators time to issue extra services or provide travelers with adequate warning.
The system was created by mathematician Wilhelm Landerholm. It uses historical data to predict how a small delay, even as little as two minutes, will affect the running of the rest of the network. Often the initial late train causes a ripple effect, with subsequent services being delayed to accommodate new platform arrival time, which then affect subsequent trains, and so on. But soon, using ‘The Commuter Prognosis’, Stockholmstag train operators will be able to make the necessary adjustments to prevent this. In addition, the information will be relayed to commuters, enabling them to take a different train and therefore reducing overcrowding. The prediction tool is expected to be put into use in Sweden by the end of the year….(More)”

Revolution Delayed: The Impact of Open Data on the Fight against Corruption


Report by RiSSC – Research Centre on Security and Crime (Italy): “In the recent years, the demand for Open Data picked up stream among stakeholders to increasing transparency and accountability of the Public Sector. Governments are supporting Open Data supply, to achieve social and economic benefits, return on investments, and political consensus.

While it is self-evident that Open Data contributes to greater transparency – as it makes data more available and easy to use by the public and governments, its impact on fighting corruption largely depends on the ability to analyse it and develop initiatives that trigger both social accountability mechanisms, and government responsiveness against illicit or inappropriate behaviours.

To date, Open Data Revolution against corruption is delayed. The impact of Open Data on the prevention and repression of corruption, and on the development of anti- corruption tools, appears to be limited, and the return on investments not yet forthcoming. Evidence remains anecdotal, and a better understanding on the mechanisms and dynamics of using Open Data against corruption is needed.

The overall objective of this exploratory study is to provide evidence on the results achieved by Open Data, and recommendations for the European Commission and Member States’ authorities, for the implementation of effective anti-corruption strategies based on transparency and openness, to unlock the potential impact of “Open Data revolution” against Corruption.

The project has explored the legal framework and the status of implementation of Open Data policies in four EU Countries – Italy, United Kingdom, Spain, and Austria. TACOD project has searched for evidence on Open Data role on law enforcement cooperation, anti-corruption initiatives, public campaigns, and investigative journalism against corruption.

RiSSC – Research Centre on Security and Crime (Italy), the University of Oxford and the University of Nottingham (United Kingdom), Transparency International (Italy and United Kingdom), the Institute for Conflict Resolution (Austria), and Blomeyer&Sanz (Spain), have carried out the research between January 2014 and February 2015, under an agreement with the European Commission – DH Migration and Home Affairs. The project has been coordinated by RiSSC, with the support of a European Working Group of Experts, chaired by prof. Richard Rose, and an external evaluator, Mr. Andrea Menapace, and it has benefited from the contribution of many experts, activists, representatives of Institutions in the four Countries….(More)

The Website That Visualizes Human Activity in Cities Across the World


Emerging Technology From the arXiv: “The data from mobile phones is revolutionizing our understanding of human activity. In recent years, it has revealed commuting patterns in major cities, wealth distribution in African countries, and even reproductive strategies in western societies. That has provided unprecedented insight for economists, sociologists, and city planners among others.

But this kind of advanced research is just a first step in a much broader trend. Phone data is set to become a standard resource that almost anyone can use to study and watch humanity continuously, much as they can now watch the weather unfold anywhere on the planet almost in real time.

But one thing is holding them back—the lack of powerful computational tools that can gather, crunch, and present the data in meaningful ways.

Today, that looks set to change to the work of Dániel Kondor and a few pals at the SENSEable City Laboratory, part of MIT, and at Ericsson, a company that produces network infrastructure technologies. These guys have unveiled a powerful online tool that uses mobile phone data to visualize human activity in cities all over the world.

This new tool, called ManyCities, allows anybody to study human activity in various cities with unprecedented detail.  But the key is that it organizes and presents the data in intuitive ways that quickly reveals trends and special events….

ManyCities then presents the data in three simple ways. The first shows how phone usage varies over time, revealing clear daily and weekly patterns as well as longer term trends. For example, ManyCities clearly shows a steady, long-term increase in data traffic, the effect of holidays, and how usage patterns change dramatically during important events like the Wimbledon tennis championship in London.

ManyCities also allows user to drill down into the data to compare patterns in different neighborhoods or in different cities. It shows, for example, that text message activity peaks in the morning in Hong Kong, in the evening in New York and at midday in London….Kondor and co have made it available at www.ManyCities.org for anybody to try.

This kind of tool is clearly evolving into a real time analytics tool. It’s not hard to imagine how people could use it to plan events such as conferences, sporting contests, or concerts or to plan emergency city infrastructure. One day people may even tune in to a “smartphone forecast” to find out if their phone will work when the big game kicks off that evening.

Ref: arxiv.org/abs/1509.00459 : Visualizing Signatures Of Human Activity In Cities Across The Globe”

Can the crowd deliver more open government?


  at GovernmentNews: “…Crowdsourcing and policy making was the subject of a lecture by visiting academic Dr Tanja Aitamurto at Victoria’s Swinburne University of Technology earlier this month. Dr Aitamurto wrote “Crowdsourcing for Democracy: New Era in Policy-Making” and led the design and implementation of the Finnish Experiment, a pioneering case study in crowdsourcing policy making.

She spoke about how Scandinavian countries have used crowdsourcing to “tap into the collective intelligence of a large and diverse crowd” in an “open ended knowledge information search process” in an open call for anybody to participate online and complete a task.

It has already been used widely and effectively by companies  such as Proctor and Gamble who offer a financial reward for solutions to their R&D problems.

The Finnish government recently used crowdsourcing when it came to reform the country’s Traffic Act following a rash of complaints to the Minister of the Environment about it. The Act, which regulates issues such as off-road traffic, is an emotive issue in Finland where snow mobiles are used six months of the year and many people live in remote areas.

The idea was for people to submit problems and solutions online, covering areas such as safety, noise, environmental protection, the rights of snowmobile owners and landowners’ rights. Everyone could see what was written and could comment on it.

Dr Aitamurto said crowdsourcing had four stages:

• The problem mapping space, where people were asked to outline the issues that needed solving
• An appeal for solutions
• An expert panel evaluated the comments received based on the criteria of: effectiveness, cost efficiency, ease of implementation and fairness. The crowd also had the chance to evaluate and rank solutions online
• The findings were then handed over to the government for the law writing process

Dr Aitamurto said active participation seemed to create a strong sense of empowerment for those involved.

She said some people reported that it was the first time in their lives they felt they were really participating in democracy and influencing decision making in society. They said it felt much more real than voting in an election, which felt alien and remote.

“Participation becomes a channel for advocacy, not just for self-interest but a channel to hear what others are saying and then also to make yourself heard. People expected a compromise at the end,” Dr Aitamurto said.

Being able to participate online was ideal for people who lived remotely and turned crowdsourcing into a democratic innovation which brought citizens closer to policy and decision making between elections.

Other benefits included reaching out to tap into new pools of knowledge, rather than relying on a small group of homogenous experts to solve the problem.

“When we use crowdsourcing we actually extend our knowledge search to multiple, hundreds of thousands of distant neighbourhoods online and that can be the power of crowdsourcing: to find solutions and information that we wouldn’t find otherwise. We find also unexpected information because it’s a self-selecting crowd … people that we might not have in our networks already,” Dr Aitamurto said.

The process can increase transparency as people interact on online platforms and where the government keeps feedback loops going.

Dr Aitamurto is also a pains to highlight what crowdsourcing is not and cannot be, because participants are self-selecting and not statistically representative.

“The crowd doesn’t make decisions, it provides information. It’s not a method or tool for direct democracy and it’s not a public opinion poll either”.

Crowdsourcing has fed into policy in other countries too, for example, during Iceland’s constitutional reform and in the United States where the federal Emergency Management Agency overhauled its strategy after a string of natural disasters.

Australian government has been getting in on the act using cloud-based software Citizen Space to gain input into a huge range of topics. While much of it is technically consultation, rather than feeding into actual policy design, it is certainly a step towards more open government.

British company Delib, which is behind the software, bills it as “managing, publicising and archiving all of your organisation’s consultation activity”.

One council who has used Citizens Space is Wyong Shire on the NSW Central Coast. The council has used the consultation hub to elicit ratepayers’ views on a number of topics, including a special rate variation, community precinct forums, strategic plans and planning decisions.

One of Citizen Space’s most valuable features is the section ‘we asked, you said, we did’….(More)”

A new journal wants to publish your research ideas


at ScienceInsider: “Do you have a great idea for a study that you want to share with the world? A new journal will gladly publish it. Research Ideas and Outcomes(RIO) will also publish papers on your methods, workflows, data, reports, and software—in short, “all outputs of the research cycle.” RIO, an open-access (OA) journal, was officially launched today and will start accepting submissions in November.

“We’re interested in making the full process of science open,” says RIO founding editor Ross Mounce, a researcher at the Natural History Museum in London. Many good research proposals fall by the wayside because funding agencies have limited budgets, Mounce says; RIO is a way to give them another chance. Mounce hopes that funders will use the journal to spot interesting new projects.

Publishing proposals can also help create links between research teams, Mounce says. “Let’s say you’re going to Madagascar for 6 months to sample turtle DNA,” he suggests. ”If you can let other researchers know ahead of time, you can agree to do things together.”

RIO‘s idea to publish research proposals is “exactly what we need if we really want to have open science,” says Iryna Kuchma, the OA program manager at the nonprofit organization Electronic Information for Libraries in Rome. Pensoft, the publishing company behind RIO, is a “strong open-access publishing venue” that has proven its worth with more than a dozen journals in the biodiversity field, Kuchma says.

The big question is, of course: Will researchers want to share promising ideas, at the risk that rivals run with them?…(More)”