100 parliaments as open data, ready for you to use


Myfanwy Nixon at mySociety’s blog and OpeningParliament: “If you need data on the people who make up your parliament, another country’s parliament, or indeed all parliaments, you may be in luck.

Every Politician, the latest Poplus project, aims to collect, store and share information about every parliament in the world, past and present—and it already contains 100 of them.

What’s more, it’s all provided as Open Data to anyone who would like to use it to power a civic tech project. We’re thinking parliamentary monitoring organisations, journalists, groups who run access-to-democracy sites like our own WriteToThem, and especially researchers who want to do analysis across multiple countries.

But isn’t that data already available?

Yes and no. There’s no doubt that you can find details of most parliaments online, either on official government websites, on Wikipedia, or on a variety of other places online.

But, as you might expect from data that’s coming from hundreds of different sources, it’s in a multitude of different formats. That makes it very hard to work with in any kind of consistent fashion.

Every Politician standardises all of its data into the Popolo standard and then provides it in two simple downloadable formats:

  • csv, which contains basic data that’s easy to work with on spreadsheets
  • JSON which contains richer data on each person, and is ideal for developers

This standardisation means that it should now be a lot easier to work on projects across multiple countries, or to compare one country’s data with another. It also means that data works well with other Poplus Components….(More)”

Beyond the Common Rule: Ethical Structures for Data Research in Non-Academic Settings


Future of Privacy Forum: “In the wake of last year’s news about the Facebook “emotional contagion” study and subsequent public debate about the role of A/B Testing and ethical concerns around the use of Big Data, FPF Senior Fellow Omer Tene participated in a December symposum on corporate consumer research hosted by Silicon Flatirons. This past month, the Colorado Technology Law Journal published a series of papers that emerged out of the symposium, including “Beyond the Common Rule: Ethical Structures for Data Research in Non-Academic Settings.”

“Beyond the Common Rule,” by Jules Polonetsky, Omer Tene, and Joseph Jerome, continues the Future of Privacy Forum’s effort to build on the notion of consumer subject review boards first advocated by Ryan Calo at FPF’s 2013 Big Data symposium. It explores how researchers, increasingly in corporate settings, are analyzing data and testing theories using often sensitive personal information. Many of these new uses of PII are simply natural extensions of current practices, and are either within the expectations of individuals or the bounds of the FIPPs. Yet many of these projects could involve surprising applications or uses of data, exceeding user expectations, and offering notice and obtaining consent could may not be feasible.

This article expands on ideas and suggestions put forward around the recent discussion draft of the White House Consumer Privacy Bill of Rights, which espouses “Privacy Review Boards” as a safety value for noncontextual data uses. It explores how existing institutional review boards within the academy and for human testing research could offer lessons for guiding principles, providing accountability and enhancing consumer trust, and offers suggestions for how companies — and researchers — can pursue both knowledge and data innovation responsibly and ethically….(More)”

Local open data ecosystems – a prototype map


Ed Parkes and Gail Dawes at Nesta: “It is increasingly recognised that some of the most important open data is published by local authorities (LAs) – data which is important to us like bin collection days, planning applications and even where your local public toilet is. Also given the likely move towards greater decentralisation, firstly through devolution to cities, the importance of the publication of local open data could arguably become more important over the next couple of years. In addition, as of 1st April, there is a new transparency code for local government requiring local authorities to publish further information on things like spending to local land assets. To pre-empt this likely renewed focus on local open data we have begun to develop a prototype map to highlight the UK’s local open data ecosystem.

Already there is some great practice in the publication of open data at a local level – such as Leeds Data Mill, London Datastore, and Open Data Sheffield. This regional activity is also characterised not just by high quality data publication, but also by pulling together through hackdays, challenges and meetups a community interested in the power of open data. This creates an ecosystem of publishers and re-users at a local level. Some of the best practice in relation to developing such an ecosystem was recognised by the last government in the announcement of a group of Local Authority Open Data Champions. Some of these were also recipients of the funding for projects from both the Cabinet Office and through the Open Data User Group.

Outside of this best practice it isn’t always easy to understand how developed smaller, less urban open data agendas are. Other than looking at each councils’ website or increasingly on the data portals that forwarding thinking councils are providing, there is a surprisingly large number of places that local authorities could make their open data available. The most well known of these is the Openly Local project but at the time of writing this now seems to be retired. Perhaps the best catalogue of local authority data is on Data.gov.uk itself. This has 1,449 datasets published by LAs across 200 different organisations. Following that there is the Open Data Communities website which hosts links to LA linked datasets. Using data from the latter, Steve Peters has developed the local data dashboard (which was itself based on the UK Local Government Open Data resource map from Owen Boswarva). In addition, local authorities can also register their open data in the LGA’s Open Data Inventory Service and take it through the ODI’s data certification process.

Prototype map of local open data eco-systems

To try to highlight patterns in local authority open data publication we decided to make a map of activity around the country (although in the first instance we’ve focused on England)….(More)

Yelp’s Consumer Protection Initiative: ProPublica Partnership Brings Medical Info to Yelp


Yelp Official Blog: “…exists to empower and protect consumers, and we’re continually focused on how we can enhance our service while enhancing the ability for consumers to make smart transactional decisions along the way.

A few years ago, we partnered with local governments to launch the LIVES open data standard. Now, millions of consumers find restaurant inspection scores when that information is most relevant: while they’re in the middle of making a dining decision (instead of when they’re signing the check). Studies have shown that displaying this information more prominently has a positive impact.

Today we’re excited to announce we’ve joined forces with ProPublica to incorporate health care statistics and consumer opinion survey data onto the Yelp business pages of more than 25,000 medical treatment facilities. Read more in today’s Washington Post story.

We couldn’t be more excited to partner with ProPublica, the Pulitzer Prize winning non-profit newsroom that produces investigative journalism in the public interest.

The information is compiled by ProPublica from their own research and the Centers for Medicare and Medicaid Services (CMS) for 4,600 hospitals, 15,000 nursing homes, and 6,300 dialysis clinics in the US and will be updated quarterly. Hover text on the business page will explain the statistics, which include number of serious deficiencies and fines per nursing home and emergency room wait times for hospitals. For example, West Kendall Baptist Hospital has better than average doctor communication and an average 33 minute ER wait time, Beachside Nursing Center currently has no deficiencies, and San Mateo Dialysis Center has a better than average patient survival rate.

Now the millions of consumers who use Yelp to find and evaluate everything from restaurants to retail will have even more information at their fingertips when they are in the midst of the most critical life decisions, like which hospital to choose for a sick child or which nursing home will provide the best care for aging parents….(More)

Print Wikipedia


Print Wikipedia is a both a utilitarian visualization of the largest accumulation of human knowledge and a poetic gesture towards the futility of the scale of big data. Michael Mandiberg has written software that parses the entirety of the English-language Wikipedia database and programmatically lays out 7600 volumes, complete with covers, and then uploads them to Lulu.com. In addition, he has compiled a Wikipedia Table of Contents, and a Wikipedia Contributor Appendix…..

Michael Mandiberg is an interdisciplinary artist, scholar, and educator living in Brooklyn, New York. He received his M.F.A. from the California Institute of the Arts and his B.A. from Brown University. His work traces the lines of political and symbolic power online, working on the Internet in order to comment on and intercede in the real flows of information. His work lives at Mandiberg.com.

Print Wikipedia by Michael Mandiberg from Lulu.com on Vimeo.”

 

Four things policy-makers need to know about social media data and real time analytics.


Ella McPherson at LSE’s Impact Blog: “I recently gave evidence to the House of Commons Science and Technology Select Committee. This was based on written evidence co-authored with my colleague, Anne Alexander, and submitted to their ongoing inquiry into social media data and real time analytics. Both Anne and I research the use of social media during contested times; Anne looks at its use by political activists and labour movement organisers in the Arab world, and I look at its use in human rights reporting. In both cases, the need to establish facticity is high, as is the potential for the deliberate or inadvertent falsification of information. Similarly to the case that Carruthers makes about war reporting, we believe that the political-economic, methodological, and ethical issues raised by media dynamics in the context of crisis are bellwethers for the dynamics in more peaceful and mundane contexts.

From our work we have learned four crucial lessons that policy-makers considering this issue should understand:

1.  Social media information is vulnerable to a variety of distortions – some typical of all information, and others more specific to the characteristics of social media communications….

2.  If social media information is used to establish events, it must be verified; while technology can hasten this process, it is unlikely to ever occur real time due to the subjective, human element of judgment required….

 

3.  Verifying social media information may require identifying its source, which has ethical implications related to informed consent and anonymisation….

4.  Another way to think about social media information is as what Hermida calls an ‘awareness system,’ which reduces the need to collect source identities; under this approach, researchers look at volume rather than veracity to recognise information of interest… (More)

How We’re Changing the Way We Respond to Petitions


Jason Goldman (White House) at Medium: “…In 2011 (years before I arrived at the White House), the team here developed a petitions platform called We the People. It provided a clear and easy way for the American people to petition their government — along with a threshold for action. Namely — once a petition gains 100,000 signatures.

This was a new system for the United States government, announced as a flagship effort in the first U.S. Open Government National Action Plan. Right now it exists only for the White House (Hey, Congress! We have anopen API! Get in touch!) Some other countries, including Germany and theUnited Kingdom, do online petitions, too. In fact, the European Parliamenthas even started its own online petitioning platform.

For the most part, we’ve been pretty good about responding — before today, the Obama Administration had responded to 255 petitions that had collectively gathered more than 11 million signatures. That’s more than 91 percent of the petitions that have met our threshold requiring a response. Some responses have taken a little longer than others. But now, I’m happy to say, we have caught up.

Today, the White House is responding to every petition in our We the Peoplebacklog — 20 in all.

This means that nearly 2.5 million people who had petitioned us to take action on something heard back today. And it’s our goal to make that response the start of the conversation, not the final page. The White House is made up of offices that research and analyze the kinds of policy issues raised by these petitions, and leaders from those offices will be taking questions today, and in the weeks to come, from petition signers, on topics such as vaccination policy, community policing, and other petition subjects.

Take a look at more We the People stats here.

We’ll start the conversation on Twitter. Follow @WeThePeople, and join the conversation using hashtag #WeThePeople. (I’ll be personally taking your questions on @Goldman44 about how we’re changing the platform specifically at 3:30 p.m. Eastern.)

We the People, Moving Forward

We’re going to be changing a few things about We the People.

  1. First, from now on, if a petition meets the signature goal within a designated period of time, we will aim to respond to it — with an update or policy statement — within 60 days wherever possible. You can read about the details of our policy in the We the People Terms of Participation.
  2. Second, other outside petitions platforms are starting to tap into the We the People platform. We’re excited to announce today that Change.org is choosing to integrate with the We the People platform, meaning the future signatures of its 100 million users will count toward the threshold for getting an official response from the Administration. We’re also opening up the code behind petitions.whitehouse.gov on Drupal.org and GitHub, which empowers other governments and outside organizations to create their own versions of this platform to engage their own citizens and constituencies.
  3. Third, and most importantly, the process of hearing from us about your petition is going to look a little different. We’ve assembled a team of people responsible for taking your questions and requests and bringing them to the right people — whether within the White House or in an agency within the Administration — who may be in a position to say something about your request….(More)

Quantifying Crowd Size with Mobile Phone and Twitter Data


, , and Being able to infer the number of people in a specific area is of extreme importance for the avoidance of crowd disasters and to facilitate emergency evacuations. Here, using a football stadium and an airport as case studies, we present evidence of a strong relationship between the number of people in restricted areas and activity recorded by mobile phone providers and the online service Twitter. Our findings suggest that data generated through our interactions with mobile phone networks and the Internet may allow us to gain valuable measurements of the current state of society….(More)”

Transforming public services the right way with impact management


Emily Bazalgette at FutureGov: “…Impact evaluation involves using a range of research methodologies to investigate whether our products and services are having an impact on users’ lives. ….Rigorous academic impact evaluation wasn’t really designed for rapidly iterating products made by a fast-moving digital and design company like FutureGov. Our products can change significantly over short periods of time — for instance, in a single workshop Doc Ready evolved from a feature-rich social media platform to a stripped-down checklist builder — and that can create a tension between our agile process and traditional evaluation methodologies, which tend to require a fixed product to support a long-term evaluation plan.

We’ve decided to embrace this tension by using Theories of Change, a useful evaluation tool recommended to us by our investors and partners Nesta Impact Investments. To give you a flavour (excuse the pun), below we have Casserole Club’s Theory of Change.

Casserole toc

The problem we’re trying to solve (reducing social isolation) doesn’t tend to change, but the way we solve it might (the inputs and short to medium-term outcomes). In future, we may find that we need to adapt to serve new user groups, or operate in different channels, or that there are mediating outcomes for social isolation that Casserole Club produces other than social contact with a Casserole Club cook. Theories of Change allow us to stay focused on big-picture outcomes, while being flexible about how the product delivers on these outcomes.

Another lesson is to make evaluation everyone’s business. Like many young-ish companies, FutureGov is not at the stage where we have the resources to support a full-time, dedicated Head of Impact. But we’ve found that you can get pretty far if you’ve got a flat structure and lots of passionate people (both of which, luckily, we have). Our lack of hierarchy means that anyone can take up a project and run with it, and collaboration across the company is encouraged. Product impact evaluation is owned by the product teams who manage the product over time. This means we can get more done, that research design benefits from the deep knowledge of our product teams, and that evaluation skills (like how to design a decent survey or depth interview) have started to spread across the organisation….(More)”

Transform Government From The Outside In


Review by GCN of a new report by Forrester: “Agencies struggles to match the customer experience available from the private sector, and that causes citizens to become dissatisfied with government. In fact, seven of the 10 worst organizations in the Forrester’s U.S. Customer Experience Index are federal agencies, and only a third of Americans say their experience with the government meets expectations.

FINDINGS: To keep up with public expectations, Forrester found governments must embrace mobile, turn big data into actionable insights, improve the customer experience and accelerate digital government.  Among the recommendations:

Agencies must shift their thinking to make mobile the primary platform for connection between citizens and government.  Government staff should also have mobile access to the tools and resources needed to complete tasks in the field. Agencies should learn what mobile methods work best for citizens, ensure all citizen services are mobile-friendly and use the mobile platform for sharing information with the public and gathering incident reports and sentiments. By building mobile-friendly infrastructure and processes, like municipal Wi-Fi hotspots, the government (and its services) can be constantly connected to its citizens and businesses.

Governments must find ways to integrate, share and use the large amounts of data and analytics it collects. By aggregating citizen-driven data from precinct-level or agency-specific databases and data collected by systems already in place, the government can increase responsiveness, target areas in need and make better short-term decisions and long-term plans. Opening data to researchers, the private sector and citizens can also spark innovation across industries.

Better customer experience has a ripple effect through government, improving the efficacy of legislation, compliance, engagement and the effectiveness of government offices. This means making processes such as applying for healthcare, registering a car or paying taxes easier and available with highly functioning user-friendly websites.  Such improvements in  communication and digital customer service, will save citizens’ time, increase the use of government services and reduce agencies’ workloads….(More)”