The Royal Statistical Society Data Manifesto


ePSiplatform: “A Data Manifesto released by the Royal Statistical Society describes ten recommendations that focus on how the next UK government can improve data for policymaking, democracy and for prosperity…..the Society calls for official statistics to be at the heart of policy debate and recommends that the Office for National Statistics and the wider Government Statistical Service be given adequate resources, as well as calling for greater investment in research, science and innovation.
The document shows that the Society is broadly supportive of the open data agenda; in particular the opening up of government data and giving citizens greater access to quality local data.
It calls for greater data sharing between government departments for statistics and research purposes and believes the private sector should be encouraged to share data with researchers for the same purpose. It also calls for an end to pre-release access to official statistics….Download the Data Manifesto in PDF format.”

Digital Enlightenment Yearbook 2014


Book edited O’Hara, K. , Nguyen, M-H.C., Haynes, P.: “Tracking the evolution of digital technology is no easy task; changes happen so fast that keeping pace presents quite a challenge. This is, nevertheless, the aim of the Digital Enlightenment Yearbook.
This book is the third in the series which began in 2012 under the auspices of the Digital Enlightenment Forum. This year, the focus is on the relationship of individuals with their networks, and explores “Social networks and social machines, surveillance and empowerment”. In what is now the well-established tradition of the yearbook, different stakeholders in society and various disciplinary communities (technology, law, philosophy, sociology, economics, policymaking) bring their very different opinions and perspectives to bear on this topic.
The book is divided into four parts: the individual as data manager; the individual, society and the market; big data and open data; and new approaches. These are bookended by a Prologue and an Epilogue, which provide illuminating perspectives on the discussions in between. The division of the book is not definitive; it suggests one narrative, but others are clearly possible.
The 2014 Digital Enlightenment Yearbook gathers together the science, social science, law and politics of the digital environment in order to help us reformulate and address the timely and pressing questions which this new environment raises. We are all of us affected by digital technology, and the subjects covered here are consequently of importance to us all. (Contents)”

Access to Scientific Data in the 21st Century: Rationale and Illustrative Usage Rights Review


Paper by James Campbell  in Data Science Journal: “Making scientific data openly accessible and available for re-use is desirable to encourage validation of research results and/or economic development. Understanding what users may, or may not, do with data in online data repositories is key to maximizing the benefits of scientific data re-use. Many online repositories that allow access to scientific data indicate that data is “open,” yet specific usage conditions reviewed on 40 “open” sites suggest that there is no agreed upon understanding of what “open” means with respect to data. This inconsistency can be an impediment to data re-use by researchers and the public. (More)”

Open Government: Origin, Development, and Conceptual Perspectives


Paper by Bernd W. Wirtz & Steven Birkmeyer in the International Journal of Public Administration: “The term “open government” is frequently used in practice and science. Since President Obama’s Memorandum for the Heads of Executive Departments and Agencies in March 2009, open government has attracted an enormous amount of public attention. It is applied by authors from diverse areas, leading to a very heterogeneous comprehension of the concept. Against this background, this article screens the current open government literature to deduce an integrative definition of open government. Furthermore, this article analyzes the empirical and conceptual literature of open government to deduce an open government framework. In general, this article provides a clear understanding of the open government concept. (More)”

With a Few Bits of Data, Researchers Identify ‘Anonymous’ People


in the New York Times: “Even when real names and other personal information are stripped from big data sets, it is often possible to use just a few pieces of the information to identify a specific person, according to a study to be published Friday in the journal Science.

In the study, titled “Unique in the Shopping Mall: On the Reidentifiability of Credit Card Metadata,” a group of data scientists analyzed credit card transactions made by 1.1 million people in 10,000 stores over a three-month period. The data set contained details including the date of each transaction, amount charged and name of the store.

Although the information had been “anonymized” by removing personal details like names and account numbers, the uniqueness of people’s behavior made it easy to single them out.

In fact, knowing just four random pieces of information was enough to reidentify 90 percent of the shoppers as unique individuals and to uncover their records, researchers calculated. And that uniqueness of behavior — or “unicity,” as the researchers termed it — combined with publicly available information, like Instagram or Twitter posts, could make it possible to reidentify people’s records by name.

“The message is that we ought to rethink and reformulate the way we think about data protection,” said Yves-Alexandre de Montjoye, a graduate student in computational privacy at the M.I.T. Media Lab who was the lead author of the study. “The old model of anonymity doesn’t seem to be the right model when we are talking about large-scale metadata.”

The analysis of large data sets containing details on people’s behavior holds great potential to improve public health, city planning and education.

But the study calls into question the standard methods many companies, hospitals and government agencies currently use to anonymize their records. It may also give ammunition to some technologists and privacy advocates who have challenged the consumer-tracking processes used by advertising software and analytics companies to tailor ads to so-called anonymous users online….(More).”

How Network Science Is Changing Our Understanding of Law


Emerging Technology From the arXiv: “One of the more fascinating areas of science that has emerged in recent years is the study of networks and their application to everyday life. It turns out that many important properties of our world are governed by networks with very specific properties.
These networks are not random by any means. Instead, they are often connected in the now famous small world pattern in which any part of the network can be reached in a relatively small number of steps. These kinds of networks lie behind many natural phenomena such as earthquakes, epidemics and forest fires and are equally ubiquitous in social phenomena such as the spread of fashions, languages, and even wars.
So it should come as no surprise that the same kind of network should exist in the legal world. Today, Marios Koniaris and pals at the National Technical University of Athens in Greece show that the network of links between laws follows exactly the same pattern. They say their network approach provides a unique insight into the nature of the law, the way it has emerged and how changes may influence it in the future.
The work of Koniaris and co focuses entirely on the law associated with the European Union. They begin by pointing out that this legal network is different from many other types of networks in two important ways….
The network can also be using for visualizing the nature of the legal world. It reveals clusters and related connections and can help legislators determine the effect of proposed changes…..It also shows how network science is spreading to every corner of scientific and social research.
Ref:  arxiv.org/abs/1501.05237 : Network Analysis In The Legal Domain: A Complex Model For European Union Legal Sources”

The new scientific revolution: Reproducibility at last


in the Washington Post:”…Reproducibility is a core scientific principle. A result that can’t be reproduced is not necessarily erroneous: Perhaps there were simply variables in the experiment that no one detected or accounted for. Still, science sets high standards for itself, and if experimental results can’t be reproduced, it’s hard to know what to make of them.
“The whole point of science, the way we know something, is not that I trust Isaac Newton because I think he was a great guy. The whole point is that I can do it myself,” said Brian Nosek, the founder of a start-up in Charlottesville, Va., called the Center for Open Science. “Show me the data, show me the process, show me the method, and then if I want to, I can reproduce it.”
The reproducibility issue is closely associated with a Greek researcher, John Ioannidis, who published a paper in 2005 with the startling title “Why Most Published Research Findings Are False.”
Ioannidis, now at Stanford, has started a program to help researchers improve the reliability of their experiments. He said the surge of interest in reproducibility was in part a reflection of the explosive growth of science around the world. The Internet is a factor, too: It’s easier for researchers to see what everyone else is doing….
Errors can potentially emerge from a practice called “data dredging”: When an initial hypothesis doesn’t pan out, the researcher will scan the data for something that looks like a story. The researcher will see a bump in the data and think it’s significant, but the next researcher to come along won’t see it — because the bump was a statistical fluke….
So far about 7,000 people are using that service, and the center has received commitments for $14 million in grants, with partners that include the National Science Foundation and the National Institutes of Health, Nosek said.
Another COS initiative will help researchers register their experiments in advance, telling the world exactly what they plan to do, what questions they will ask. This would avoid the data-dredging maneuver in which researchers who are disappointed go on a deep dive for something publishable.
Nosek and other reformers talk about “publication bias.” Positive results get reported, negative results ignored. Someone reading a journal article may never know about all the similar experiments that came to naught….(More).”

How to Convince Men to Help the Poor


at Pacific Standard: “Please give. It’s a plea we are confronted with constantly, as a variety of charities implore us to help them help the less fortunate.

Whether we get out our checkbook or throw the request in the recycling bin is determined, in part, by the specific way the request is framed. But a new study suggests non-profits might want to create two separate appeals: One aimed at men, and another at women.

A research team led by Stanford University sociologist Robb Willer reports empathy-based appeals tend to be effective with women. But as a rule, men—who traditionally give somewhat less to anti-poverty charities—need to be convinced that their self-interest aligns with that of the campaign.

“Framing poverty as an issue that negatively affects all Americans increased men’s willingness to donate to the cause, eliminating the gender gap,” the researchers write in the journal Social Science Research….

“While this reframing resonated with men, who were otherwise less likely to spontaneously express concern about poverty,” Willer and his colleagues write, “it had the opposite effect for women, who might have felt less motivated to express concern about poverty when doing so seemed inconsistent with feeling empathy for the poor.”…(More)”

The Cathedral of Computation


at the Atlantic: “We’re not living in an algorithmic culture so much as a computational theocracy.  Algorithms are everywhere, supposedly. We are living in an “algorithmic culture,” to use the author and communication scholar Ted Striphas’s name for it. Google’s search algorithms determine how we access information. Facebook’s News Feed algorithms determine how we socialize. Netflix’s and Amazon’s collaborative filtering algorithms choose products and media for us. You hear it everywhere. “Google announced a change to its algorithm,” a journalist reports. “We live in a world run by algorithms,” a TED talk exhorts. “Algorithms rule the world,” a news report threatens. Another upgrades rule to dominion: “The 10 Algorithms that Dominate Our World.”…
It’s part of a larger trend. The scientific revolution was meant to challenge tradition and faith, particularly a faith in religious superstition. But today, Enlightenment ideas like reason and science are beginning to flip into their opposites. Science and technology have become so pervasive and distorted, they have turned into a new type of theology.
The worship of the algorithm is hardly the only example of the theological reversal of the Enlightenment—for another sign, just look at the surfeit of nonfiction books promising insights into “The Science of…” anything, from laughter to marijuana. But algorithms hold a special station in the new technological temple because computers have become our favorite idols….
Once you adopt skepticism toward the algorithmic- and the data-divine, you can no longer construe any computational system as merely algorithmic. Think about Google Maps, for example. It’s not just mapping software running via computer—it also involves geographical information systems, geolocation satellites and transponders, human-driven automobiles, roof-mounted panoramic optical recording systems, international recording and privacy law, physical- and data-network routing systems, and web/mobile presentational apparatuses. That’s not algorithmic culture—it’s just, well, culture….(More).”

Would You Share Private Data for the Good of City Planning?


Henry Grabar at NextCity: “The proliferation of granular data on automobile movement, drawn from smartphones, cab companies, sensors and cameras, is sharpening our sense of how cars travel through cities. Panglossian seers believe the end of traffic jams is nigh.
This information will change cities beyond their roads. Real-time traffic data may lead to reworked intersections and new turning lanes, but understanding cars is in some ways a stand-in for understanding people. There’s traffic as traffic and traffic as proxy, notes Brett Goldstein, an urban science fellow at the University of Chicago who served as that city’s first data officer from 2011 to 2013. “We’d be really naive, in thinking about how we make cities better,” he says, “to only consider traffic for what it is.”
Even a small subset of a city’s car data goes a long way. Consider the raft of discrete findings that have emerged from the records of New York City taxis.
Researchers at the Massachusetts Institute of Technology, led by Paolo Santi, showed that cab-sharing could reduce taxi mileage by 40 percent. Their counterparts at NYU, led by Claudio Silva, mapped activity around hubs like train stations and airports and during hurricanes.
“You start to build actual models of how people move, and where they move,” observes Silva, the head of disciplines at NYU’s Center for Science and Urban Progress (CUSP). “The uses of this data for non-traffic engineering are really substantial.”…
Many of these ideas are hypothetical, for the moment, because so-called “granular” data is so hard to come by. That’s one reason the release of New York’s taxi cab data spurred so many studies — it’s an oasis of information in a desert of undisclosed records. Corporate entreaties, like Uber’s pending data offering to Boston, don’t always meet researchers’ standards. “It’s going to be a lot of superficial data, and it’s not clear how usable it’ll be at this point,” explains Sarah Kaufman, the digital manager at NYU’s Rudin Center for Transportation….
Yet Americans seem much more alarmed by the collection of location data than other privacy breaches.
How can data utopians convince the hoi polloi to share their comings and goings? One thought: Make them secure. Mike Flowers, the founder of New York City’s Office of Data Analytics and a fellow at NYU’s CUSP, told me it might be time to consider establishing a quasi-governmental body that people would trust to make their personal data anonymous before they are channeled into government projects. (New York City’s Taxi and Limousine Commission did not do a very good job at this, which led to Gawker publishing a dozen celebrity cab rides.)
Another idea is to frame open data as a beneficial trade-off. “When people provide information, they want to realize the benefit of the information,” Goldstein says.
Users tell the routing company Waze where they are and get a smoother commute in return. Progressive Insurance offers drivers a “Snapshot” tracker. If it likes the way you drive, the company will lower your rates. It’s not hard to imagine that, in the long run, drivers will be penalized for refusing such a device…. (More).”