How mobile text reminders earned Madagascar a 32,900% ROI in collecting unpaid taxes


Paper by Tiago Peixoto et al : “Benjamin Franklin famously once said that “nothing can be said to be certain, except death and taxes.” In developing countries, however, tax revenues are anything but certain. Madagascar is a prime example, with tax collection as a share of GDP at just under 11 percent. This is low even compared with countries of similar levels of economic development, and well below what the government could reasonably collect to fund much-needed public services, such as education, health and infrastructure. 

Poor compliance by citizens who owe taxes remains a major reason for Madagascar’s low tax collection. Madagascar’s government has therefore made increasing tax revenue collection a high priority in its strategy for promoting sustainable economic growth and addressing poverty.

Reforming a tax system can take decades. But small measures, implemented with the help of technology, can help tax authorities improve compliance.  Our team at the World Bank jointly conducted a field experiment with the Madagascar’s Directorate General for Taxation, to test whether simple text message reminders via mobile phones could increase compliance among late-tax filers.

We took a group of 15,885 late-income-tax filers and randomly assigned some of them to receive a series of messages reminding them to file a tax declaration and emphasizing various reasons to pay taxes. Late tax filers were told that they could avoid a late penalty by meeting an extended deadline and were given the link to the tax filing website. 

The results of the experiment were significant. In the control group, only 7.2% of late filers filed a tax return by the extended deadline cited in the SMS messages. This increased to 9.8% in the treatment groups who received SMS reminders. This might not sound like much, but for every dollar spent sending text messages, the tax authority collected an additional 329 dollars in revenues, making the intervention highly cost-effective.

In fact, the return on this particular investment was 32,900 percent! Although this increase in revenue is relatively small in absolute terms—around $375,000—it could be automatically integrated into the tax system. It also suggests that messaging may hold a lot of promise for cost-effectively increasing tax receipts even in developing country contexts….(More)”.

Law as Data: Computation, Text, and the Future of Legal Analysis


Book edited by Michael A. Livermore and Daniel N. Rockmore: “In recent years, the digitization of legal texts, combined with developments in the fields of statistics, computer science, and data analytics, have opened entirely new approaches to the study of law. This volume explores the new field of computational legal analysis, an approach marked by its use of legal texts as data. The emphasis herein is work that pushes methodological boundaries, either by using new tools to study longstanding questions within legal studies or by identifying new questions in response to developments in data availability and analysis.

By using the text and underlying data of legal documents as the direct objects of quantitative statistical analysis, Law as Data introduces the legal world to the broad range of computational tools already proving themselves relevant to law scholarship and practice, and highlights the early steps in what promises to be an exciting new approach to studying the law….(More)”.

The Psychology of Prediction


Blog post by Morgan Housel: “During the Vietnam War Secretary of Defense Robert McNamara tracked every combat statistic he could, creating a mountain of analytics and predictions to guide the war’s strategy.

Edward Lansdale, head of special operations at the Pentagon, once looked at McNamara’s statistics and told him something was missing.

“What?” McNamara asked.

“The feelings of the Vietnamese people,” Landsdale said.

That’s not the kind of thing a statistician pays attention to. But, boy, did it matter.

I believe in prediction. I think you have to in order to get out of bed in the morning.

But prediction is hard. Either you know that or you’re in denial about it.

A lot of the reason it’s hard is because the visible stuff that happens in the world is a small fraction of the hidden stuff that goes on inside people’s heads. The former is easy to overanalyze; the latter is easy to ignore.

This report describes 12 common flaws, errors, and misadventures that occur in people’s heads when predictions are made….(More)”.

The plan to mine the world’s research papers


Priyanka Pulla in Nature: “Carl Malamud is on a crusade to liberate information locked up behind paywalls — and his campaigns have scored many victories. He has spent decades publishing copyrighted legal documents, from building codes to court records, and then arguing that such texts represent public-domain law that ought to be available to any citizen online. Sometimes, he has won those arguments in court. Now, the 60-year-old American technologist is turning his sights on a new objective: freeing paywalled scientific literature. And he thinks he has a legal way to do it.

Over the past year, Malamud has — without asking publishers — teamed up with Indian researchers to build a gigantic store of text and images extracted from 73 million journal articles dating from 1847 up to the present day. The cache, which is still being created, will be kept on a 576-terabyte storage facility at Jawaharlal Nehru University (JNU) in New Delhi. “This is not every journal article ever written, but it’s a lot,” Malamud says. It’s comparable to the size of the core collection in the Web of Science database, for instance. Malamud and his JNU collaborator, bioinformatician Andrew Lynn, call their facility the JNU data depot.

No one will be allowed to read or download work from the repository, because that would breach publishers’ copyright. Instead, Malamud envisages, researchers could crawl over its text and data with computer software, scanning through the world’s scientific literature to pull out insights without actually reading the text.

The unprecedented project is generating much excitement because it could, for the first time, open up vast swathes of the paywalled literature for easy computerized analysis. Dozens of research groups already mine papers to build databases of genes and chemicals, map associations between proteins and diseases, and generate useful scientific hypotheses. But publishers control — and often limit — the speed and scope of such projects, which typically confine themselves to abstracts, not full text. Researchers in India, the United States and the United Kingdom are already making plans to use the JNU store instead. Malamud and Lynn have held workshops at Indian government laboratories and universities to explain the idea. “We bring in professors and explain what we are doing. They get all excited and they say, ‘Oh gosh, this is wonderful’,” says Malamud.

But the depot’s legal status isn’t yet clear. Malamud, who contacted several intellectual-property (IP) lawyers before starting work on the depot, hopes to avoid a lawsuit. “Our position is that what we are doing is perfectly legal,” he says. For the moment, he is proceeding with caution: the JNU data depot is air-gapped, meaning that no one can access it from the Internet. Users have to physically visit the facility, and only researchers who want to mine for non-commercial purposes are currently allowed in. Malamud says his team does plan to allow remote access in the future. “The hope is to do this slowly and deliberately. We are not throwing this open right away,” he says….(More)”.

What Restaurant Reviews Reveal About Cities


Linda Poon at CityLab: “Online review sites can tell you a lot about a city’s restaurant scene, and they can reveal a lot about the city itself, too.

Researchers at MIT recently found that information about restaurants gathered from popular review sites can be used to uncover a number of socioeconomic factors of a neighborhood, including its employment rates and demographic profiles of the people who live, work, and travel there.

A report published last week in the Proceedings of the National Academy of Sciences explains how the researchers used information found on Dianping—a Yelp-like site in China—to find information that might usually be gleaned from an official government census. The model could prove especially useful for gathering information about cities that don’t have that kind of reliable or up-to-date government data, especially in developing countries with limited resources to conduct regular surveys….

Zheng and her colleagues tested out their machine-learning model using restaurant data from nine Chinese cities of various sizes—from crowded ones like Beijing, with a population of more than 10 million, to smaller ones like Baoding, a city of fewer than 3 million people.

They pulled data from 630,000 restaurants listed on Dianping, including each business’s location, menu prices, opening day, and customer ratings. Then they ran it through a machine-learning model with official census data and with anonymous location and spending data gathered from cell phones and bank cards. By comparing the information, they were able to determine where the restaurant data reflected the other data they had about neighborhoods’ characteristics.

They found that the local restaurant scene can predict, with 95 percent accuracy, variations in a neighborhood’s daytime and nighttime populations, which are measured using mobile phone data. They can also predict, with 90 and 93 percent accuracy, respectively, the number of businesses and the volume of consumer consumption. The type of cuisines offered and kind of eateries available (coffeeshop vs. traditional teahouses, for example), can also predict the proportion of immigrants or age and income breakdown of residents. The predictions are more accurate for neighborhoods near urban centers as opposed to those near suburbs, and for smaller cities, where neighborhoods don’t vary as widely as those in bigger metropolises….(More)”.

From Hippocrates to Artificial Intelligence: Moving Towards a Collective Intelligence


Carlos María Galmarini at Open Mind: “Modern medicine is based upon the work of Hippocrates and his disciples and is compiled in more than 70 books comprising the Hippocratic body of work. In its essence, these writings declare that any illness originates with natural causes. Therefore, medicine must be based on detailed observation, reason, and experience in order to establish a diagnosis, prognosis, and treatment. The Hippocratic tradition stresses the importance of the symptoms and the clinical exam. As a result, medicine abandoned superstition and the magic performed by priest-doctors, and it was transformed into a real, experience-based science….

A complementary combination of both intelligences (human and artificial) could help overcome the other’s shortcomings and limitations. As we incorporate intelligent technologies into medical processes, a new, more powerful form of collaboration will emerge. Analogous to the past when the automation of human tasks completely changed the known world and ignited an evolution in the offering of products and services, the combination of human and artificial intelligence will create a new type of collective intelligence capable of building more efficient organizations, and in the healthcare industry, it will be able to solve problems that until now have been unfathomable to the human mind alone.

Finally, it is worth remembering that fact based sciences are divided into natural and human disciplines. Medicine occupies a special place, straddling both. It can be difficult to establish the similarities between a doctor who works, for example, with rules defined by specific clinical trials and a traditional family practitioner. The former would be more related to a natural science, and the latter with a more human science – “the art of medicine.”

The combination of human and artificial intelligence in a new type of collective intelligence will enable doctors themselves to be a combination of the two. In other words, the art of medicine – human science – based on the analysis of big data – natural science. A new collective intelligence working on behalf of a wiser medicine….(More)”.

Smart Cities in Application: Healthcare, Policy, and Innovation


Book edited by Stan McClellan: “This book explores categories of applications and driving factors surrounding the Smart City phenomenon. The contributing authors provide perspective on the Smart Cities, covering numerous applications and classes of applications. The book uses a top-down exploration of the driving factors in Smart Cities, by including focal areas including “Smart Healthcare,” “Public Safety & Policy Issues,” and “Science, Technology, & Innovation.”  Contributors have direct and substantive experience with important aspects of Smart Cities and discuss issues with technologies & standards, roadblocks to implementation, innovations that create new opportunities, and other factors relevant to emerging Smart City infrastructures….(More)”.

Applying design science in public policy and administration research


Paper by Sjoerd Romme and Albert Meijer: “There is increasing debate about the role that public policy research can play in identifying solutions to complex policy challenges. Most studies focus on describing and explaining how governance systems operate. However, some scholars argue that because current institutions are often not up to the task, researchers need to rethink this ‘bystander’ approach and engage in experimentation and interventions that can help to change and improve governance systems.

This paper contributes to this discourse by developing a design science framework that integrates retrospective research (scientific validation) and prospective research (creative design). It illustrates the merits and challenges of doing this through two case studies in the Netherlands and concludes that a design science framework provides a way of integrating traditional validation-oriented research with intervention-oriented design approaches. We argue that working at the interface between them will create new opportunities for these complementary modes of public policy research to achieve impact….(More)”

Review into bias in algorithmic decision-making


Interim Report by the Centre for Data Ethics and Innovation (UK): The use of algorithms has the potential to improve the quality of decision- making by increasing the speed and accuracy with which decisions are made. If designed well, they can reduce human bias in decision-making processes. However, as the volume and variety of data used to inform decisions increases, and the algorithms used to interpret the data become more complex, concerns are growing that without proper oversight, algorithms risk entrenching and potentially worsening bias.

The way in which decisions are made, the potential biases which they are subject to and the impact these decisions have on individuals are highly context dependent. Our Review focuses on exploring bias in four key sectors: policing, financial services, recruitment and local government. These have been selected because they all involve significant decisions being made about individuals, there is evidence of the growing uptake of machine learning algorithms in the sectors and there is evidence of historic bias in decision-making within these sectors. This Review seeks to answer three sets of questions:

  1. Data: Do organisations and regulators have access to the data they require to adequately identify and mitigate bias?
  2. Tools and techniques: What statistical and technical solutions are available now or will be required in future to identify and mitigate bias and which represent best practice?
  3. Governance: Who should be responsible for governing, auditing and assuring these algorithmic decision-making systems?

Our work to date has led to some emerging insights that respond to these three sets of questions and will guide our subsequent work….(More)”.

Participatory Citizenship and Crisis in Contemporary Brazil


Book by Valesca Lima: “This book discusses the issues of citizen rights, governance and political crisis in Brazil. The project has a focus on “citizenship in times of crisis,” i.e., seeking to understand how citizenship rights have changed since the Brazilian political and economic crisis that started in 2014. Building on theories of citizenship and governance, the author examines policy-based evidence on the retractions of participatory rights, which are consequence of a stagnant economic scenario and the re-organization of conservative sectors. This work will appeal to scholarly audiences interested in citizenship, Brazilian politics, and Latin American policy and governance….(More)”.