Innovation for the Sustainable Development Goals


UNDP: “In 2014, UNDP, with the generous support of the Government of Denmark, established an Innovation Facility to improve service delivery and support national governments and citizens to tackle complex challenges.

The report ‘Spark, Scale, Sustain’ shares UNDP’s approach to innovation, over 40 case studies of innovation for the Sustainable Development Goals in practice and Features on Alternative Finance, Behavioral Insights, Data Innovation and Public Policy Labs.

Download the report to find out more about the innovation initiatives that are testing and scaling solutions to address challenges across five areas:

  • Eradicate Poverty, Leave No One Behind
  • Protect the Planet
  • Build Peaceful Societies, Prevent Violent Conflict
  • Manage Risk, Improve Disaster Response
  • Advance Gender Equality & Women’s Empowerment….(More)”.

NIH-funded team uses smartphone data in global study of physical activity


National Institutes of Health: “Using a larger dataset than for any previous human movement study, National Institutes of Health-funded researchers at Stanford University in Palo Alto, California, have tracked physical activity by population for more than 100 countries. Their research follows on a recent estimate that more than 5 million people die each year from causes associated with inactivity.

The large-scale study of daily step data from anonymous smartphone users dials in on how countries, genders, and community types fare in terms of physical activity and what results may mean for intervention efforts around physical activity and obesity. The study was published July 10, 2017, in the advance online edition of Nature.

“Big data is not just about big numbers, but also the patterns that can explain important health trends,” said Grace Peng, Ph.D., director of the National Institute of Biomedical Imaging and Bioengineering (NIBIB) program in Computational Modeling, Simulation and Analysis.

“Data science and modeling can be immensely powerful tools. They can aid in harnessing and analyzing all the personalized data that we get from our phones and wearable devices.”

Almost three quarters of adults in developed countries and half of adults in developing economies carry a smartphone. The devices are equipped with tiny accelerometers, computer chip that maintains the orientation of the screen, and can also automatically record stepping motions. The users whose data contributed to this study subscribed to the Azumio Argus app, a free application for tracking physical activity and other health behaviors….

In addition to the step records, the researchers accessed age, gender, and height and weight status of users who registered the smartphone app. They used the same calculation that economists use for income inequality — called the Gini index — to calculate activity inequality by country.

“These results reveal how much of a population is activity-rich, and how much of a population is activity-poor,” Delp said. “In regions with high activity inequality there are many people who are activity poor, and activity inequality is a strong predictor of health outcomes.”…

The researchers investigated the idea that making improvements in a city’s walkability — creating an environment that is safe and enjoyable to walk — could reduce activity inequality and the activity gender gap.

“If you must cross major highways to get from point A to point B in a city, the walkability is low; people rely on cars,” Delp said. “In cities like New York and San Francisco, where you can get across town on foot safely, the city has high walkability.”

Data from 69 U.S. cities showed that higher walkability scores are associated with lower activity inequality. Higher walkability is associated with significantly more daily steps across all age, gender, and body-mass-index categories.  However, the researchers found that women recorded comparatively less activity than men in places that are less walkable.

The study exemplifies how smartphones can deliver new insights about key health behaviors, including what the authors categorize as the global pandemic of physical inactivity….(More)”.

Principles and Practices for a Federal Statistical Agency


National Academies of Sciences Report: “Publicly available statistics from government agencies that are credible, relevant, accurate, and timely are essential for policy makers, individuals, households, businesses, academic institutions, and other organizations to make informed decisions. Even more, the effective operation of a democratic system of government depends on the unhindered flow of statistical information to its citizens.

In the United States, federal statistical agencies in cabinet departments and independent agencies are the governmental units whose principal function is to compile, analyze, and disseminate information for such statistical purposes as describing population characteristics and trends, planning and monitoring programs, and conducting research and evaluation. The work of these agencies is coordinated by the U.S. Office of Management and Budget. Statistical agencies may acquire information not only from surveys or censuses of people and organizations, but also from such sources as government administrative records, private-sector datasets, and Internet sources that are judged of suitable quality and relevance for statistical use. They may conduct analyses, but they do not advocate policies or take partisan positions. Statistical purposes for which they provide information relate to descriptions of groups and exclude any interest in or identification of an individual person, institution, or economic unit.

Four principles are fundamental for a federal statistical agency: relevance to policy issues, credibility among data users, trust among data providers, and independence from political and other undue external influence.� Principles and Practices for a Federal Statistical Agency: Sixth Edition presents and comments on these principles as they’ve been impacted by changes in laws, regulations, and other aspects of the environment of federal statistical agencies over the past 4 years….(More)”.

The ethics issue: Should we abandon privacy online?


Special issue of the New Scientist: “Those who would give up essential Liberty to purchase a little temporary Safety,” Benjamin Franklin once said, “deserve neither Liberty nor Safety.” But if Franklin were alive today, where would he draw the line? Is the freedom to send an encrypted text message essential? How about the right to keep our browsing history private? What is the sweet spot between our need to be left alone and our desire to keep potential criminals from communicating in secret?

In an age where fear of terrorism is high in the public consciousness, governments are likely to err on the side of safety. Over the past decade, the authorities have been pushing for – and getting – greater powers of surveillance than they have ever had, all in the name of national security.

The downsides are not immediately obvious. After all, you might think you have nothing to hide. But most of us have perfectly legal secrets we’d rather someone else didn’t see. And although the chances of the authorities turning up to take you away in a black SUV on the basis of your WhatsApp messages are small in free societies, the chances of insurance companies raising your premiums are not….(More)”.

Open data: Accountability and transparency


 at Big Data and Society: “The movements by national governments, funding agencies, universities, and research communities toward “open data” face many difficult challenges. In high-level visions of open data, researchers’ data and metadata practices are expected to be robust and structured. The integration of the internet into scientific institutions amplifies these expectations. When examined critically, however, the data and metadata practices of scholarly researchers often appear incomplete or deficient. The concepts of “accountability” and “transparency” provide insight in understanding these perceived gaps. Researchers’ primary accountabilities are related to meeting the expectations of research competency, not to external standards of data deposition or metadata creation. Likewise, making data open in a transparent way can involve a significant investment of time and resources with no obvious benefits. This paper uses differing notions of accountability and transparency to conceptualize “open data” as the result of ongoing achievements, not one-time acts….(More)”.

Avoiding Garbage In – Garbage Out: Improving Administrative Data Quality for Research


Blog by : “In June, I presented the webinar, “Improving Administrative Data Quality for Research and Analysis”, for members of the Association of Public Data Users (APDU). APDU is a national network that provides a venue to promote education, share news, and advocate on behalf of public data users.

The webinar served as a primer to help smaller organizations begin to use their data for research. Participants were given the tools to transform their administrative data into “research-ready” datasets.

I first reviewed seven major issues for administrative data quality and discussed how these issues can affect research and analysis. For instance, issues with incorrect value formats, unit of analysis, and duplicate records can make the data difficult to use. Invalid or inconsistent values lead to inaccurate analysis results. Missing or outlier values can produce inaccurate and biased analysis results. All these issues make the data less useful for research.

Next, I presented concrete strategies for reviewing the data to identify each of these quality issues. I also discussed several tips to make the data review process easier, faster, and easy to replicate. Most importantly among these tips are: (1) reviewing everyvariable in the data set, whether you expect problems or not, and (2) relying on data documentation to understand how the data should look….(More)”.

Data for Development: The Case for Information, Not Just Data


Daniela Ligiero at the Council on Foreign Relations: “When it comes to development, more data is often better—but in the quest for more data, we can often forget about ensuring we have information, which is even more valuable. Information is data that have been recorded, classified, organized, analyzed, interpreted, and translated within a framework so that meaning emerges. At the end of the day, information is what guides action and change.

The need for more data

In 2015, world leaders came together to adopt a new global agenda to guide efforts over the next fifteen years, the Sustainable Development Goals. The High-level Political Forum (HLPF), to be held this year at the United Nations on July 10-19, is an opportunity for review of the 2030 Agenda, and will include an in-depth analysis of seven of the seventeen goals—including those focused on poverty, health, and gender equality. As part of the HLPF, member states are encouraged to undergo voluntary national reviews of progress across goals to facilitate the sharing of experiences, including successes, challenges, and lessons learned; to strengthen policies and institutions; and to mobilize multi-stakeholder support and partnerships for the implementation of the agenda.

A significant challenge that countries continue to face in this process, and one that becomes painfully evident during the HLPF, is the lack of data to establish baselines and track progress. Fortunately, new initiatives aligned with the 2030 Agenda are working to focus on data, such as the Global Partnership for Sustainable Development Data. There are also initiatives focus on collecting more and better data in particular areas, like gender data (e.g., Data2X; UN Women’s Making Every Girl and Woman Count). This work is important and urgently needed.

Data to monitor global progress on the goals is critical to keeping countries accountable to their commitments and allows countries to examine how they are doing across multiple, ambitious goals. However, equally important is the rich, granular national and sub-national level data that can guide the development and implementation of evidence-based, effective programs and policies. These kinds of data are also often lacking or of poor quality, in which case more data and better data is essential. But a frequently-ignored piece of the puzzle at the national level is improved use of the data we already have.

Making the most of the data we have

To illustrate this point, consider the Together for Girls partnership, which was built on obtaining new data where it was lacking and effectively translating it into information to change policies and programs. We are a partnership between national governments, UN agencies and private sector organizations working to break cycles of violence, with special attention to sexual violence against girls. …The first pillar of our work is focused on understanding violence against children within a country, always at the request of the national government. We do this through a national household survey – the Violence Against Children Survey (VACS), led by national governments, CDC, and UNICEF as part of the Together for Girls Partnership….

The truth is there is a plethora of data at the country level, generated by surveys, special studies, administrative systems, private sector, and citizens that can provide meaningful insights across all the development goals.

Connecting the dots

But data—like our programs’—often remain in silos. For example, data focused on violence against children is typically not top of mind for those working on women’s empowerment or adolescent health. Yet, as an example, the VACS can offer valuable information about how sexual violence against girls, as young as 13,is connected to adolescent pregnancy—or how one of the most common perpetrators of sexual violence against girls is a partner, a pattern that starts early and is a predictor for victimization and perpetration later in life.  However, these data are not consistently used across actors working on programs related to adolescent pregnancy and violence against women….(More)”.

Research data infrastructures in the UK


The Open Research Data Task Force : “This report is intended to inform the work of the Open Research Data Task Force, which has been established with the aim of building on the principles set out in Open Research Data Concordat (published in July 2016) to co-ordinate creation of a roadmap to develop the infrastructure for open research data across the UK. As an initial contribution to that work, the report provides an outline of the policy and service infrastructure in the UK as it stands in the first half of 2017, including some comparisons with other countries; and it points to some key areas and issues which require attention. It does not seek to identify possible courses of action, nor even to suggest priorities the Task Force might consider in creating its final report to be published in 2018. That will be the focus of work for the Task Force over the next few months.

Why is this important?

The digital revolution continues to bring fundamental changes to all aspects of research: how it is conducted, the findings that are produced, and how they are interrogated and transmitted not only within the research community but more widely. We are as yet still in the early stages of a transformation in which progress is patchy across the research community, but which has already posed significant challenges for research funders and institutions, as well as for researchers themselves. Research data is at the heart of those challenges: not simply the datasets that provide the core of the evidence analysed in scholarly publications, but all the data created and collected throughout the research process. Such data represents a potentially-valuable resource for people and organisations in the commercial, public and voluntary sectors, as well as for researchers. Access to such data, and more general moves towards open science, are also critically-important in ensuring that research is reproducible, and thus in sustaining public confidence in the work of the research community. But effective use of research data depends on an infrastructure – of hardware, software and services, but also of policies, organisations and individuals operating at various levels – that is as yet far from fully-formed. The exponential increases in volumes of data being generated by researchers create in themselves new demands for storage and computing power. But since the data is characterised more by heterogeneity then by uniformity, development of the infrastructure to manage it involves a complex set of requirements in preparing, collecting, selecting, analysing, processing, storing and preserving that data throughout its life cycle.

Over the past decade and more, there have been many initiatives on the part of research institutions, funders, and members of the research community at local, national and international levels to address some of these issues. Diversity is a key feature of the landscape, in terms of institutional types and locations, funding regimes, and nature and scope of partnerships, as well as differences between disciplines and subject areas. Hence decision-makers at various levels have fostered via their policies and strategies many community-organised developments, as well as their own initiatives and services. Significant progress has been achieved as a result, through the enthusiasm and commitment of key organisations and individuals. The less positive features have been a relative lack of harmonisation or consolidation, and there is an increasing awareness of patchiness in provision, with gaps, overlaps and inconsistencies. This is not surprising, since policies, strategies and services relating to research data necessarily affect all aspects of support for the diverse processes of research itself. Developing new policies and infrastructure for research data implies significant re-thinking of structures and regimes for supporting, fostering and promoting research itself. That in turn implies taking full account of widely-varying characteristics and needs of research of different kinds, while also keeping in clear view the benefits to be gained from better management of research data, and from greater openness in making data accessible for others to re-use for a wide range of different purposes….(More)”.

The State of Open Data Portals in Latin America


Michael Steinberg at Center for Data Innovation: “Many Latin American countries publish open data—government data made freely available online in machine-readable formats and without license restrictions. However, there is a tremendous amount of variation in the quantity and type of datasets governments publish on national open data portals—central online repositories for open data that make it easier for users to find data. Despite the wide variation among the countries, the most popular datasets tend to be those that either provide transparency into government operations or offer information that citizens can use directly. As governments continue to update and improve their open data portals, they should take steps to ensure that they are publishing the datasets most valuable to their citizens.

To better understand this variation, we collected information about open data portals in 20 Latin American countries including Argentina, Bolivia, Brazil, Chile, Colombia, Costa Rica, Ecuador, Mexico, Panama, Paraguay, Peru, and Uruguay. Not all Latin American countries have an open data portal, but even if they do not operate a unified portal, some governments may still have open data. Four Latin American countries—Belize, Guatemala, Honduras, and Nicaragua—do not have open data portals. One country— El Salvador—does not have a government-run open data portal, but does have a national open data portal (datoselsalvador.org) run by volunteers….

There are many steps Latin American governments can take to improve open data in their country. Those nations without open data portals should create them, and those who already have them should continue to update them and publish more datasets to better serve their constituents. One way to do this is to monitor the popular datasets on other countries’ open data portals, and where applicable, ensure the government produces similar datasets. Those running open data portals should also routinely monitor search queries to see what users are looking for, and if they are looking for datasets that have not yet been posted, work with the relevant government agencies to make these datasets available.

In summary, there are stark differences in the amount of data published, the format of the data, and the most popular datasets in open data portals in Latin America. However, in every country there is an appetite for data that either provides public accountability for government functions or supplies helpful information to citizens…(More)”.

The Right of Access to Public Information


Book by Hermann-Josef Blanke and Ricardo Perlingeiro: “This book presents a comparative study on access to public information in the context of the main legal orders worldwide. The international team of authors analyzes the Transparency- and Freedom-to-Information legislation with regard to the scope of the right to access, limitations of this right inherent in the respective national laws, the procedure, the relationship with domestic legislation on administrative procedure, as well as judicial protection. It particularly focuses on the Brazilian law of access to information, which is interpreted as a benchmark for regulations in other Latin-American states….(More)”