Where next for open government?


Blog Post by Natalia Domagala: “…We can all agree that open government is a necessary and valuable concept. 

Nevertheless, eight years since the Open Government Partnership (OGP) was founded — the leading intergovernmental forum moving the agenda of open government forward — the challenge is now how to adapt their processes to reflect the dynamic and often unstable realm of global politics. 

For open government to be truly impactful, policies should account for the reality of government work. If we get this wrong, there is a risk of open government becoming a token of participation without any meaning. 

The collective goal of open government practitioners/community should be to strive for open government to become the new normal — an aim that requires looking at the cracks in the current process and thinking of what can be done to address them. 

As an example, there have been an increasing number of letters sent by the OGP in the past few years as a reaction to national action plans being published too or as notifications of late self-assessment returns. 

If a large number of countries across the geographical spectrum continuously miss these deadlines, this would indicate that a change of approach may be needed. Perhaps it’s time to move away from the two year cycles of national action plans that seemingly haven’t been working for an increasing number of countries, and experiment with the length and format of open government plans. 

Changing the policy rhythm

Longer, 4 or 6 year strategic commitments could lead to structural changes in how governments approach open dataparticipatory policymaking, and other principles of open government. 

Two years is a short time in the cycle of government, and offers insufficient time to deliver desirable results. The pressure to start thinking about the next plan half way through implementing the first one can negatively impact the quality of commitments and their impact. 

Having a rolling NAP that is updated with very specific actions for every two years could be another alternative. Open government is a vibrant and fast-growing movement, therefore action plans should reflect it through being living and interactive documents. Perhaps after two or three national action plans countries should be allowed to adjust the cycle to their needs and domestic government planning timescales. 

There is an opportunity for open government as a movement in going beyond the national action plan commitments. Open government teams within governments should scrutinise existing policies and advise their colleagues on how to align their policymaking process with the principles of participation, accountability, and inclusion, to eventually embed the open government approach across all policy projects. 

Appetite for new strategies 

The rise of “open”, “agile”, and “participatory” attitudes to policy indicate that there is an appetite for more responsive and better-tailored strategies, an appetite that the global open government movement could look to satisfy. 

The next steps could be focused on raising awareness of open ways of working within governments, and developing the policymaker’s capacity to deploy them through workshops and guidance….(More)”.

Understanding our Political Nature: How to put knowledge and reason at the heart of political decision-making


EU report by Rene Van Bavel et al: “Recognising that advances in behavioural, decision and social sciences demonstrate that we are not purely rational beings, this report brings new insights into our political behaviour and this understanding have the potential to address some of the current crises in our democracies. Sixty experts from across the globe working in the fields of behavioural and social sciences as well as the humanities, have contributed to the research that underpins this JRC report that calls upon evidence-informed policymaking not to be taken for granted. There is a chapter dedicated to each key finding which outlines the latest scientific thinking as well as an overview of the possible implications for policymaking. The key findings are:

  • Misperception and Disinformation: Our thinking skills are challenged by today’s information environment and make us vulnerable to disinformation. We need to think more about how we think.
  • Collective Intelligence: Science can help us re-design the way policymakers work together to take better decisions and prevent policy mistakes.
  • Emotions: We can’t separate emotion from reason. Better information about citizens’ emotions and greater emotional literacy could improve policymaking.
  • Values and Identities drive political behaviour but are not properly understood or debated.
  • Framing, Metaphor and Narrative: Facts don’t speak for themselves. Framing, metaphors and narratives need to be used responsibly if evidence is to be heard and understood.
  • Trust and Openness: The erosion of trust in experts and in government can only be addressed by greater honesty and public deliberation about interests and values.
  • Evidence-informed policymaking: The principle that policy should be informed by evidence is under attack. Politicians, scientists and civil society need to defend this cornerstone of liberal democracy….(More)”

The personification of big data


Paper by Stevenson, Phillip Douglas and Mattson, Christopher Andrew: “Organizations all over the world, both national and international, gather demographic data so that the progress of nations and peoples can be tracked. This data is often made available to the public in the form of aggregated national level data or individual responses (microdata). Product designers likewise conduct surveys to better understand their customer and create personas. Personas are archetypes of the individuals who will use, maintain, sell or otherwise be affected by the products created by designers. Personas help designers better understand the person the product is designed for. Unfortunately, the process of collecting customer information and creating personas is often a slow and expensive process.

In this paper, we introduce a new method of creating personas, leveraging publicly available databanks of both aggregated national level and information on individuals in the population. A computational persona generator is introduced that creates a population of personas that mirrors a real population in terms of size and statistics. Realistic individual personas are filtered from this population for use in product development…(More)”.

Responding to Some Challenges Posed by the Reidentification of Anonymized Personal Data


Paper by Herman T. Tavani and Frances S. Grodzinsky: “In this paper, we examine a cluster of ethical controversies generated by the reidentification of anonymized personal data in the context of big data analytics, with particular attention to the implications for personal privacy. Our paper is organized into two main parts. Part One examines some ethical problems involving re-identification of personally identifiable information (PII) in large data sets. Part Two begins with a brief description of Moor and Weckert’s Dynamic Ethics (DE) and Nissenbaum’s Contextual Integrity (CI) Frameworks. We then investigate whether these frameworks, used together, can provide us with a more robust scheme for analyzing privacy concerns that arise in the re-identification process (as well as within the larger context of big data analytics). This paper does not specifically address re-identification-related privacy concerns that arise in the context of the European Union’s General Data Protection Regulation (GDPR). Instead, we examine those issues in a separate work….(More)”.

Scientists use phone movement to predict personality types


RMIT: “Researchers used data from mobile phone accelerometers – the tiny sensors tracking phone movement for step-counting and other apps – to predict people’s personalities.

RMIT University computer scientist Associate Professor Flora Salim said previous studies had predicted personality types using phone call and messaging activity logs, but this study showed adding accelerometer data improved accuracy

“Activity like how quickly or how far we walk, or when we pick up our phones up during the night, often follows patterns and these patterns say a lot about our personality type,” said Salim, a leading expert in human mobility data.

Physical activity is proven to have a strong correlation with human personality. Therefore, researchers analysed physical activity features from different dimensions like dispersion, diversity, and regularity.

Key findings from the study:

  • People with consistent movements on weekday evenings were generally more introverted, while extroverts displayed more random patterns, perhaps meeting up with different people and taking up unplanned options.
  • Agreeable people had more random activity patterns and were busier on weekends and weekday evenings than others.
  • Friendly and compassionate females made more outgoing calls than anyone else.
  • Conscientious, organized people didn’t tend to contact the same person often in a short space of time.
  • Sensitive or neurotic females often checked their phones or moved with their phones regularly well into the night, past midnight. Sensitive or neurotic males did the opposite.
  • More inventive and curious people tended to make and receive fewer phone calls compared to others….(More)”

This Is Not an Atlas.


Book by kollektiv orangotango: “This Is Not an Atlas gathers more than 40 counter-cartographies from all over the world. This collection shows how maps are created and transformed as a part of political struggle, for critical research or in art and education: from indigenous territories in the Amazon to the anti-eviction movement in San Francisco; from defending commons in Mexico to mapping refugee camps with balloons in Lebanon; from slums in Nairobi to squats in Berlin; from supporting communities in the Philippines to reporting sexual harassment in Cairo. This Is Not an Atlas seeks to inspire, to document the underrepresented, and to be a useful companion when becoming a counter-cartographer yourself….(More)”.

“Anonymous” Data Won’t Protect Your Identity


Sophie Bushwick at Scientific American: “The world produces roughly 2.5 quintillion bytes of digital data per day, adding to a sea of information that includes intimate details about many individuals’ health and habits. To protect privacy, data brokers must anonymize such records before sharing them with researchers and marketers. But a new study finds it is relatively easy to reidentify a person from a supposedly anonymized data set—even when that set is incomplete.

Massive data repositories can reveal trends that teach medical researchers about disease, demonstrate issues such as the effects of income inequality, coach artificial intelligence into humanlike behavior and, of course, aim advertising more efficiently. To shield people who—wittingly or not—contribute personal information to these digital storehouses, most brokers send their data through a process of deidentification. This procedure involves removing obvious markers, including names and social security numbers, and sometimes taking other precautions, such as introducing random “noise” data to the collection or replacing specific details with general ones (for example, swapping a birth date of “March 7, 1990” for “January–April 1990”). The brokers then release or sell a portion of this information.

“Data anonymization is basically how, for the past 25 years, we’ve been using data for statistical purposes and research while preserving people’s privacy,” says Yves-Alexandre de Montjoye, an assistant professor of computational privacy at Imperial College London and co-author of the new study, published this week in Nature Communications.  Many commonly used anonymization techniques, however, originated in the 1990s, before the Internet’s rapid development made it possible to collect such an enormous amount of detail about things such as an individual’s health, finances, and shopping and browsing habits. This discrepancy has made it relatively easy to connect an anonymous line of data to a specific person: if a private detective is searching for someone in New York City and knows the subject is male, is 30 to 35 years old and has diabetes, the sleuth would not be able to deduce the man’s name—but could likely do so quite easily if he or she also knows the target’s birthday, number of children, zip code, employer and car model….(More)”

Battling Information Illiteracy


Article by Paul T. Jaeger and Natalie Greene Taylor on “How misinformation affects the future of policy…“California wildfires are being magnified and made so much worse by the bad environmental laws which aren’t allowing massive amounts of readily available water to be properly utilized. It is being diverted into the Pacific Ocean. Must also tree clear to stop fire from spreading!”

This tweet was a statement by a US president about a major event, suggesting changes to existing policies. It is also not true. Every element of the tweet—other than the existence of California, the Pacific Ocean, and wildfires—is false. And it was not a simple misunderstanding, because a tweet from Trump the next day reiterated these themes and blamed the state’s governor personally for holding back water to fight the fires.

So how does this pertain to information policy, since the tweet is about environmental policy issues? The answer is in the information. The use and misuse of information in governance and policymaking may be turning into the biggest information policy issue of all. And as technologies and methods of communication evolve, a large part of engaging with and advocating for information policy will consist of addressing the new challenges of teaching information literacy and behavior.

Misinformation literacy

The internet has made it easy for people to be information illiterate in new ways. Anyone can create information now—regardless of quality—and get it in front of a large number of people. The ability of social media to spread information as fast as possible, and to as many people as possible, challenges literacy, as does the ability to manipulate images, sounds, and video with ease….(More)”

The internet is rotting – let’s embrace it


Viktor Mayer-Schönberger in The Conversation: “Every year, some thousands of sites – including ones with unique information – go offline. Countless further webpages become inaccessible; instead of information, users encounter error messages.

Where some commentators may lament yet another black hole in the slowly rotting Internet, I actually feel okay. Of course, I, too, dread broken links and dead servers. But I also know: Forgetting is important.

In fact, as I argued in my book, “Delete: The Virtue of Forgetting in the Digital Age,” all through human history, humans reserved remembering for the things that really mattered to them and forgot the rest. Now the internet is making forgetting a lot harder.

Built to forget

Humans are accustomed to a world in which forgetting is the norm, and remembering is the exception.

This isn’t necessarily a bug in human evolution. The mind forgets what is no longer relevant to our present. Human memory is constantly reconstructed – it isn’t preserved in pristine condition, but becomes altered over time, helping people overcome cognitive dissonances. For example, people may see an awful past as rosier than it was, or devalue memories of past conflict with a person with whom they are close in the present.

Forgetting also helps humans to focus on current issues and to plan for the future. Research shows that those who are too tethered to their past find it difficult to live and act in the present. Forgetting creates space for something new, enabling people to go beyond what they already know.

Organizations that remember too much ossify in their processes and behavior. Learning something new requires forgetting something old – and that is hard for organizations that remember too much. There’s a growing literature on the importance of “unlearning,” or deliberately purging deeply rooted processes or practices from an organization – a fancy way to say that forgetting fulfills a valuable purpose….(More)”.

The value of data in Canada: Experimental estimates


Statistics Canada: “As data and information take on a far more prominent role in Canada and, indeed, all over the world, data, databases and data science have become a staple of modern life. When the electricity goes out, Canadians are as much in search of their data feed as they are food and heat. Consumers are using more and more data that is embodied in the products they buy, whether those products are music, reading material, cars and other appliances, or a wide range of other goods and services. Manufacturers, merchants and other businesses depend increasingly on the collection, processing and analysis of data to make their production processes more efficient and to drive their marketing strategies.

The increasing use of and investment in all things data is driving economic growth, changing the employment landscape and reshaping how and from where we buy and sell goods. Yet the rapid rise in the use and importance of data is not well measured in the existing statistical system. Given the ‘lack of data on data’, Statistics Canada has initiated new research to produce a first set of estimates of the value of data, databases and data science. The development of these estimates benefited from collaboration with the Bureau of Economic Analysis in the United States and the Organisation for Economic Co-operation and Development.

In 2018, Canadian investment in data, databases and data science was estimated to be as high as $40 billion. This was greater than the annual investment in industrial machinery, transportation equipment, and research and development and represented approximately 12% of total non-residential investment in 2018….

Statistics Canada recently released a conceptual framework outlining how one might measure the economic value of data, databases and data science. Thanks to this new framework, the growing role of data in Canada can be measured through time. This framework is described in a paper that was released in The Daily on June 24, 2019 entitled “Measuring investments in data, databases and data science: Conceptual framework.” That paper describes the concept of an ‘information chain’ in which data are derived from everyday observations, databases are constructed from data, and data science creates new knowledge by analyzing the contents of databases….(More)”.