The New York Times thinks a blockchain could help stamp out fake news


MIT Technology Review: “Blockchain technology is at the core of a new research project the New York Times has launched, aimed at making “the origins of journalistic content clearer to [its] audience.”

The news: The Times has launched what it calls The News Provenance Project, which will experiment with ways to combat misinformation in the news media. The first project will focus on using a blockchain—specifically a platform designed by IBM—to prove that photos are authentic.

Blockchain? Really? Rumors and speculation swirled in March, after CoinDesk reported that the New York Times was looking for someone to help it develop a “blockchain-based proof-of-concept for news publishers.” Though the newspaper removed the job posting after the article came out, apparently it was serious. In a new blog post, project lead Sasha Koren explains that by using a blockchain, “we might in theory provide audiences with a way to determine the source of a photo, or whether it had been edited after it was published.”

Unfulfilled promise: Using a blockchain to prove the authenticity of journalistic content has long been considered a potential application of the technology, but attempts to do it so far haven’t gotten much traction. If the New York Times can develop a compelling application, it has enough influence to change that….(More)”.

“Anonymous” Data Won’t Protect Your Identity


Sophie Bushwick at Scientific American: “The world produces roughly 2.5 quintillion bytes of digital data per day, adding to a sea of information that includes intimate details about many individuals’ health and habits. To protect privacy, data brokers must anonymize such records before sharing them with researchers and marketers. But a new study finds it is relatively easy to reidentify a person from a supposedly anonymized data set—even when that set is incomplete.

Massive data repositories can reveal trends that teach medical researchers about disease, demonstrate issues such as the effects of income inequality, coach artificial intelligence into humanlike behavior and, of course, aim advertising more efficiently. To shield people who—wittingly or not—contribute personal information to these digital storehouses, most brokers send their data through a process of deidentification. This procedure involves removing obvious markers, including names and social security numbers, and sometimes taking other precautions, such as introducing random “noise” data to the collection or replacing specific details with general ones (for example, swapping a birth date of “March 7, 1990” for “January–April 1990”). The brokers then release or sell a portion of this information.

“Data anonymization is basically how, for the past 25 years, we’ve been using data for statistical purposes and research while preserving people’s privacy,” says Yves-Alexandre de Montjoye, an assistant professor of computational privacy at Imperial College London and co-author of the new study, published this week in Nature Communications.  Many commonly used anonymization techniques, however, originated in the 1990s, before the Internet’s rapid development made it possible to collect such an enormous amount of detail about things such as an individual’s health, finances, and shopping and browsing habits. This discrepancy has made it relatively easy to connect an anonymous line of data to a specific person: if a private detective is searching for someone in New York City and knows the subject is male, is 30 to 35 years old and has diabetes, the sleuth would not be able to deduce the man’s name—but could likely do so quite easily if he or she also knows the target’s birthday, number of children, zip code, employer and car model….(More)”

Battling Information Illiteracy


Article by Paul T. Jaeger and Natalie Greene Taylor on “How misinformation affects the future of policy…“California wildfires are being magnified and made so much worse by the bad environmental laws which aren’t allowing massive amounts of readily available water to be properly utilized. It is being diverted into the Pacific Ocean. Must also tree clear to stop fire from spreading!”

This tweet was a statement by a US president about a major event, suggesting changes to existing policies. It is also not true. Every element of the tweet—other than the existence of California, the Pacific Ocean, and wildfires—is false. And it was not a simple misunderstanding, because a tweet from Trump the next day reiterated these themes and blamed the state’s governor personally for holding back water to fight the fires.

So how does this pertain to information policy, since the tweet is about environmental policy issues? The answer is in the information. The use and misuse of information in governance and policymaking may be turning into the biggest information policy issue of all. And as technologies and methods of communication evolve, a large part of engaging with and advocating for information policy will consist of addressing the new challenges of teaching information literacy and behavior.

Misinformation literacy

The internet has made it easy for people to be information illiterate in new ways. Anyone can create information now—regardless of quality—and get it in front of a large number of people. The ability of social media to spread information as fast as possible, and to as many people as possible, challenges literacy, as does the ability to manipulate images, sounds, and video with ease….(More)”

The internet is rotting – let’s embrace it


Viktor Mayer-Schönberger in The Conversation: “Every year, some thousands of sites – including ones with unique information – go offline. Countless further webpages become inaccessible; instead of information, users encounter error messages.

Where some commentators may lament yet another black hole in the slowly rotting Internet, I actually feel okay. Of course, I, too, dread broken links and dead servers. But I also know: Forgetting is important.

In fact, as I argued in my book, “Delete: The Virtue of Forgetting in the Digital Age,” all through human history, humans reserved remembering for the things that really mattered to them and forgot the rest. Now the internet is making forgetting a lot harder.

Built to forget

Humans are accustomed to a world in which forgetting is the norm, and remembering is the exception.

This isn’t necessarily a bug in human evolution. The mind forgets what is no longer relevant to our present. Human memory is constantly reconstructed – it isn’t preserved in pristine condition, but becomes altered over time, helping people overcome cognitive dissonances. For example, people may see an awful past as rosier than it was, or devalue memories of past conflict with a person with whom they are close in the present.

Forgetting also helps humans to focus on current issues and to plan for the future. Research shows that those who are too tethered to their past find it difficult to live and act in the present. Forgetting creates space for something new, enabling people to go beyond what they already know.

Organizations that remember too much ossify in their processes and behavior. Learning something new requires forgetting something old – and that is hard for organizations that remember too much. There’s a growing literature on the importance of “unlearning,” or deliberately purging deeply rooted processes or practices from an organization – a fancy way to say that forgetting fulfills a valuable purpose….(More)”.

The Hidden Costs of Automated Thinking


Jonathan Zittrain in The New Yorker: “Like many medications, the wakefulness drug modafinil, which is marketed under the trade name Provigil, comes with a small, tightly folded paper pamphlet. For the most part, its contents—lists of instructions and precautions, a diagram of the drug’s molecular structure—make for anodyne reading. The subsection called “Mechanism of Action,” however, contains a sentence that might induce sleeplessness by itself: “The mechanism(s) through which modafinil promotes wakefulness is unknown.”

Provigil isn’t uniquely mysterious. Many drugs receive regulatory approval, and are widely prescribed, even though no one knows exactly how they work. This mystery is built into the process of drug discovery, which often proceeds by trial and error. Each year, any number of new substances are tested in cultured cells or animals; the best and safest of those are tried out in people. In some cases, the success of a drug promptly inspires new research that ends up explaining how it works—but not always. Aspirin was discovered in 1897, and yet no one convincingly explained how it worked until 1995. The same phenomenon exists elsewhere in medicine. Deep-brain stimulation involves the implantation of electrodes in the brains of people who suffer from specific movement disorders, such as Parkinson’s disease; it’s been in widespread use for more than twenty years, and some think it should be employed for other purposes, including general cognitive enhancement. No one can say how it works.

This approach to discovery—answers first, explanations later—accrues what I call intellectual debt. It’s possible to discover what works without knowing why it works, and then to put that insight to use immediately, assuming that the underlying mechanism will be figured out later. In some cases, we pay off this intellectual debt quickly. But, in others, we let it compound, relying, for decades, on knowledge that’s not fully known.

In the past, intellectual debt has been confined to a few areas amenable to trial-and-error discovery, such as medicine. But that may be changing, as new techniques in artificial intelligence—specifically, machine learning—increase our collective intellectual credit line. Machine-learning systems work by identifying patterns in oceans of data. Using those patterns, they hazard answers to fuzzy, open-ended questions. Provide a neural network with labelled pictures of cats and other, non-feline objects, and it will learn to distinguish cats from everything else; give it access to medical records, and it can attempt to predict a new hospital patient’s likelihood of dying. And yet, most machine-learning systems don’t uncover causal mechanisms. They are statistical-correlation engines. They can’t explain why they think some patients are more likely to die, because they don’t “think” in any colloquial sense of the word—they only answer. As we begin to integrate their insights into our lives, we will, collectively, begin to rack up more and more intellectual debt….(More)”.

The plan to mine the world’s research papers


Priyanka Pulla in Nature: “Carl Malamud is on a crusade to liberate information locked up behind paywalls — and his campaigns have scored many victories. He has spent decades publishing copyrighted legal documents, from building codes to court records, and then arguing that such texts represent public-domain law that ought to be available to any citizen online. Sometimes, he has won those arguments in court. Now, the 60-year-old American technologist is turning his sights on a new objective: freeing paywalled scientific literature. And he thinks he has a legal way to do it.

Over the past year, Malamud has — without asking publishers — teamed up with Indian researchers to build a gigantic store of text and images extracted from 73 million journal articles dating from 1847 up to the present day. The cache, which is still being created, will be kept on a 576-terabyte storage facility at Jawaharlal Nehru University (JNU) in New Delhi. “This is not every journal article ever written, but it’s a lot,” Malamud says. It’s comparable to the size of the core collection in the Web of Science database, for instance. Malamud and his JNU collaborator, bioinformatician Andrew Lynn, call their facility the JNU data depot.

No one will be allowed to read or download work from the repository, because that would breach publishers’ copyright. Instead, Malamud envisages, researchers could crawl over its text and data with computer software, scanning through the world’s scientific literature to pull out insights without actually reading the text.

The unprecedented project is generating much excitement because it could, for the first time, open up vast swathes of the paywalled literature for easy computerized analysis. Dozens of research groups already mine papers to build databases of genes and chemicals, map associations between proteins and diseases, and generate useful scientific hypotheses. But publishers control — and often limit — the speed and scope of such projects, which typically confine themselves to abstracts, not full text. Researchers in India, the United States and the United Kingdom are already making plans to use the JNU store instead. Malamud and Lynn have held workshops at Indian government laboratories and universities to explain the idea. “We bring in professors and explain what we are doing. They get all excited and they say, ‘Oh gosh, this is wonderful’,” says Malamud.

But the depot’s legal status isn’t yet clear. Malamud, who contacted several intellectual-property (IP) lawyers before starting work on the depot, hopes to avoid a lawsuit. “Our position is that what we are doing is perfectly legal,” he says. For the moment, he is proceeding with caution: the JNU data depot is air-gapped, meaning that no one can access it from the Internet. Users have to physically visit the facility, and only researchers who want to mine for non-commercial purposes are currently allowed in. Malamud says his team does plan to allow remote access in the future. “The hope is to do this slowly and deliberately. We are not throwing this open right away,” he says….(More)”.

Stop Surveillance Humanitarianism


Mark Latonero at The New York Times: “A standoff between the United Nations World Food Program and Houthi rebels in control of the capital region is threatening the lives of hundreds of thousands of civilians in Yemen.

Alarmed by reports that food is being diverted to support the rebels, the aid program is demanding that Houthi officials allow them to deploy biometric technologies like iris scans and digital fingerprints to monitor suspected fraud during food distribution.

The Houthis have reportedly blocked food delivery, painting the biometric effort as an intelligence operation, and have demanded access to the personal data on beneficiaries of the aid. The impasse led the aid organization to the decision last month to suspend food aid to parts of the starving population — once thought of as a last resort — unless the Houthis allow biometrics.

With program officials saying their staff is prevented from doing its essential jobs, turning to a technological solution is tempting. But biometrics deployed in crises can lead to a form of surveillance humanitarianism that can exacerbate risks to privacy and security.

By surveillance humanitarianism, I mean the enormous data collection systems deployed by aid organizations that inadvertently increase the vulnerability of people in urgent need….(More)”.

Betting on biometrics to boost child vaccination rates


Ben Parker at The New Humanitarian: “Thousands of children between the ages of one and five are due to be fingerprinted in Bangladesh and Tanzania in the largest biometric scheme of its kind ever attempted, the Geneva-based vaccine agency, Gavi, announced recently.

Although the scheme includes data protection safeguards – and its sponsors are cautious not to promise immediate benefits – it is emerging during a widening debate on data protection, technology ethics, and the risks and benefits of biometric ID in development and humanitarian aid.

Gavi, a global vaccine provider, is teaming up with Japanese and British partners in the venture. It is the first time such a trial has been done on this scale, according to Gavi spokesperson James Fulker.

Being able to track a child’s attendance at vaccination centres, and replace “very unreliable” paper-based records, can help target the 20 million children who are estimated to miss key vaccinations, most in poor or remote communities, Fulker said.

Up to 20,000 children will have their fingerprints taken and linked to their records in existing health projects. That collection effort will be managed by Simprints, a UK-based not-for-profit enterprise specialising in biometric technology in international development, according to Christine Kim, the company’s head of strategic partnerships….

Ethics and legal safeguards

Kim said Simprints would apply data protection standards equivalent to the EU’s General Directive on Privacy Regulation (GDPR), even if national legislation did not demand it. Families could opt out without any penalties, and informed consent would apply to any data gathering. She added that the fieldwork would be approved by national governments, and oversight would also come from institutional review boards at universities in the two countries.

Fulker said Gavi had also commissioned a third-party review to verify Simprints’ data protection and security methods.

For critics of biometrics use in humanitarian settings, however, any such plan raises red flags….

Data protection analysts have long been arguing that gathering digital ID and biometric data carries particular risks for vulnerable groups who face conflict or oppression: their data could be shared or leaked to hostile parties who could use it to target them.

In a recent commentary on biometrics and aid, Linda Raftree told The New Humanitarian that “the greatest burden and risk lies with the most vulnerable, whereas the benefits accrue to [aid] agencies.”

And during a panel discussion on “Digital Do No Harm” held last year in Berlin, humanitarian professionals and data experts discussed a range of threats and unintended consequences of new technologies, noting that they are as yet hard to predict….(More)”.

Politicians Don’t Actually Care What Voters Want


Joshua Kalla and Ethan Porter at the New York Times: We like to think that politicians care about what their constituents want. If voters in a legislative district have certain views about, say, the legality of abortion, we assume that their representative’s decisions will be shaped, or at least influenced, by those views. To a large extent, democracy depends on this assumption: The beliefs of voters should be reflected, however imperfectly, in the leaders they elect.

But there is reason to question this assumption. It is easy to think of issues, climate change and gun control chief among them, where the consensus of public opinion has provoked little legislative action. How much do legislators really care about the views of their constituents?

Over the past two years, we conducted a study to find out. We provided state legislators in the United States with access to highly detailed public opinion survey data — more detailed than almost all available opinion polls — about their constituents’ attitudes on gun control, infrastructure spending, abortion and many other policy issues. Afterward, we gauged the willingness of representatives to look at the data as well as how the data affected their perceptions of their constituents’ opinions.

What we found should alarm all Americans. An overwhelming majority of legislators were uninterested in learning about their constituents’ views. Perhaps more worrisome, however, was that when the legislators who did view the data were surveyed afterward, they were no better at understanding what their constituents wanted than legislators who had not looked at the data. For most politicians, voters’ views seemed almost irrelevant….(More)”.

Truth and Consequences


Sophia Rosenfeld at The Hedgehog Review: “Conventional wisdom has it that for democracy to work, it is essential that we—the citizens—agree in some minimal way about what reality looks like. We are not, of course, all required to think the same way about big questions, or believe the same things, or hold the same values; in fact, it is expected that we won’t. But somehow or other, we need to have acquired some very basic, shared understanding about what causes what, what’s broadly desirable, what’s dangerous, and how to characterize what’s already happened.

Some social scientists call this “public knowledge.” Some, more cynically, call it “serviceable truth” to emphasize its contingent, socially constructed quality. Either way, it is the foundation on which democratic politics—in which no one person or institution has sole authority to determine what’s what and all claims are ultimately revisable—is supposed to rest. It is also imagined to be one of the most exalted products of the democratic process. And to a certain degree, this peculiar, messy version of truth has held its own in modern liberal democracies, including the United States, for most of their history.

Lately, though, even this low-level kind of consensus has come to seem elusive. The issue is not just professional spinners talking about “alternative facts” or the current US president bending the truth and spreading conspiracy theories at every turn, from mass rallies to Twitter rants. The deeper problem stems from the growing sense we all have that, today, even hard evidence of the kind that used to settle arguments about factual questions won’t persuade people whose political commitments have already led them to the opposite conclusion. Rather, citizens now belong to “epistemic tribes”: One person’s truth is another’s hoax or lie. Just look at how differently those of different political leanings interpret the evidence of global warming or the conclusions of the Mueller Report on Russian involvement in the 2016 Trump presidential campaign. Moreover, many of those same people are also now convinced that the boundaries between truth and untruth are, in the end, as subjective as everything else. It is all a matter of perception and spin; nothing is immune, and it doesn’t really matter.

Headed for a Cliff

So what’s happened? Why has assent on even basic factual claims (beyond logically demonstrable ones, like 2 + 2 = 4) become so hard to achieve? Or, to put it slightly differently, why are we—meaning people of varied political persuasions—having so much trouble lately arriving at any broadly shared sense of the world beyond ourselves, and, even more, any consensus on which institutions, methods, or people to trust to get us there? And why, ultimately, do so many of us seem simply to have given up on the possibility of finding some truths in common?

These are questions that seem especially loaded precisely because of the traditionally close conceptual and historical relationship between truth and democracy as social values….(More)”.