The human rights impacts of migration control technologies


Petra Molnar at EDRI: “At the start of this new decade, over 70 million people have been forced to move due to conflict, instability, environmental factors, and economic reasons. As a response to the increased migration into the European Union, many states are looking into various technological experiments to strengthen border enforcement and manage migration. These experiments range from Big Data predictions about population movements in the Mediterranean to automated decision-making in immigration applications and Artificial Intelligence (AI) lie detectors at European borders. However, often these technological experiments do not consider the profound human rights ramifications and real impacts on human lives

A human laboratory of high risk experiments

Technologies of migration management operate in a global context. They reinforce institutions, cultures, policies and laws, and exacerbate the gap between the public and the private sector, where the power to design and deploy innovation comes at the expense of oversight and accountability. Technologies have the power to shape democracy and influence elections, through which they can reinforce the politics of exclusion. The development of technology also reinforces power asymmetries between countries and influence our thinking around which countries can push for innovation, while other spaces like conflict zones and refugee camps become sites of experimentation. The development of technology is not inherently democratic and issues of informed consent and right of refusal are particularly important to think about in humanitarian and forced migration contexts. For example, under the justification of efficiency, refugees in Jordan have their irises scanned in order to receive their weekly rations. Some refugees in the Azraq camp have reported feeling like they did not have the option to refuse to have their irises scanned, because if they did not participate, they would not get food. This is not free and informed consent….(More)”.

How to use evidence in policymaking


Inês Prates at apolitical: “…Evidence should feed into policymaking; there is no doubt about that. However, the truth is that using evidence in policy is often a very complex process and the stumbling blocks along the way are numerous.

The world has never had a larger wealth of data and information, and that is a great opportunity to open up public debate and democratise access to knowledge. At the same time, however, we are currently living in a “post-truth” era, where personal beliefs can trump scientific knowledge.

Technology and digital platforms have given room for populists to question well-established facts and evidence, and dangerously spread misinformation, while accusing scientists and policymakers of elitism for their own political gain.

Another challenge is that political interests can strategically manipulate or select (“cherry-pick”) evidence that justifies prearranged positions. A stark example of this is the evidence “cherry-picking” done by climate change sceptics who choose restricted time periods (for example of 8 to 12 years) that may not show a global temperature increase.

In addition, to unlock the benefits of evidence informed policy, we need to bridge the “policy-research gap”. Policymakers are not always aware of the latest evidence on an issue. Very often, critical decisions are made under a lot of pressure and the very nature of democracy makes policy complex and messy, making it hard to systematically integrate evidence into the process.

At the same time, researchers may be oblivious to what the most pressing policy challenges are, or how to communicate actionable insights to a non-expert audience. This constructive guide provides tips on how scientists can handle the most challenging aspects of engaging with policymakers.

Institutions like the European Commission’s in-house science service, the Joint Research Centre (JRC) sit precisely at the intersection between science and policy. Researchers from the JRC work together with policymakers on several key policy challenges. A nice example is their work on the scarcity of critical raw materials needed for the EU’s energy transition, using a storytelling tool to raise the awareness of non-experts on an extremely complex issue.

Lastly, we cannot forget about the importance of the buy-in from the public. Although policymakers can willingly ignore or manipulate evidence, they have very little incentives to ignore the will of a critical mass. Let us go back to the climate movement; it is hard to dismiss the influence of the youth-led worldwide protests on world leaders and their climate policy efforts.

Using evidence in policymaking is key to solving the world’s most pressing climate and environmental challenges. To do so effectively, we need to connect and establish trust between government, researchers and the public…(More)”.

Are these the 20 top multi-stakeholder processes in 2020 to advance a digital ecosystem for the planet?


Paper by David Jensen, Karen Bakker and Christopher Reimer: “As outlined in our recent article, The promise and peril of a digital ecosystem for the planet, we propose that the ongoing digital revolution needs to be harnessed to drive a transformation towards global sustainability, environmental stewardship, and human well-being. Public, private and civil society actors must take deliberate action and collaborate to build a global digital ecosystem for the planet. A digital ecosystem that mobilizes hardware, software and digital infrastructures together with data analytics to generate dynamic, real-time insights that can power various structural transformations are needed to achieve collective sustainability.

The digital revolution must also be used to abolish extreme poverty and reduce inequalities that jeopardize social cohesion and stability. Often, these social inequalities are tied to and overlap with ecological challenges. Ultimately, then, we must do nothing less than direct the digital revolution for planet, people, prosperity and peace.

To achieve this goal, we must embed the vision of a fair digital ecosystem for the planet into all of the key multi-stakeholder processes that are currently unfolding. We aim to do this through two new articles on Medium: a companion article on Building a digital ecosystem for the planet: 20 substantive priorities for 2020, and this one. In the companion article, we identify three primary engagement tracks: system architecture, applications, and governance. Within these three tracks, we outline 20 priorities for the new decade. Building from these priorities, our focus for this article is to identify a preliminary list of the top 20 most important multi-stakeholder processes that we must engage and influence in 2020….(More).

People learn in different ways. The way we teach should reflect that


Article by Jason Williams-Bellamy and Beth Simone Noveck: “There’s never been more hybrid learning in the public sector than today…

There are pros and cons in online and in-person training. But some governments are combining both in a hybrid (also known as blended) learning program. According to the Online Learning Consortium, hybrid courses can be either:

  • A classroom course in which online activity is mixed with classroom meetings, replacing a significant portion, but not all face-to-face activity
  • An online course that is supplemented by required face-to-face instruction such as lectures, discussions, or labs.

A hybrid course can effectively combine the short-term activity of an in-person workshop with the longevity and scale of an online course.

The Digital Leaders program in Israel is a good example of hybrid training. Digital Leaders is a nine-month program designed to train two cohorts of 40 leaders each in digital innovation by means of a regular series of online courses, shared between Israel and a similar program in the UK, interspersed with live workshops. This style of blended learning makes optimal use of participants’ time while also establishing a digital environment and culture among the cohort not seen in traditional programs.

The State government in New Jersey, where I serve as the Chief Innovation Officer, offers a free and publicly accessible online introduction to innovation skills for public servants called the Innovation Skills Accelerator. Those who complete the course become eligible for face-to-face project coaching and we are launching our first skills “bootcamp,” blending online and the face-to-face in Q1 2020.

Blended classrooms have been linked to greater engagement and increased collaboration among participating students. Blended courses allow learners to customise their learning experience in a way that is uniquely best suited for them. One study even found that blended learning improves student engagement and learning even if they only take advantage of the traditional in-classroom resources. While the added complexity of designing for online and off may be off-putting to some, the benefits are clear.

The best way to teach public servants is to give them multiple ways to learn….(More)”.

Human-centred policy? Blending ‘big data’ and ‘thick data’ in national policy


Policy Lab (UK): “….Compared with quantitative data, ethnography creates different forms of data – what anthropologists call ‘thick data’. Complex social problems benefit from insights beyond linear, standardised evidence and this is where thick data shows its worth. In Policy Lab we have generated ethnographic films and analysis to sit alongside quantitative data, helping policy-makers to build a rich picture of current circumstances. 

On the other hand, much has been written about big data – data generated through digital interactions – whether it be traditional ledgers and spreadsheets or emerging use of artificial intelligence and the internet of things.  The ever-growing zettabytes of data can reveal a lot, providing a (sometimes real time) digital trail capturing and aggregating our individual choices, preferences, behaviours and actions.  

Much hyped, this quantitative data has great potential to inform future policy, but must be handled ethically, and also requires careful preparation and analysis to avoid biases and false assumptions creeping in. Three issues we have seen in our projects relate to:

  • partial data, for example not having data on people who are not digitally active, biasing the sample
  • the time-consuming challenge of cleaning up data, in a political context where time is often of the essence
  • the lack of data interoperability, where different localities/organisations capture different metrics

Through a number of Policy Lab projects we have used big data to see the big picture before then using thick data to zoom in to the detail of people’s lived experience.  Whereas big data can give us cumulative evidence at a macro, often systemic level, thick data provides insights at an individual or group level.  We have found the blending of ‘big data’ and ‘thick data’ – to be the sweet spot. 

This is a diagram of Policy Lab's model for combining big data and thick data.
Policy Lab’s model for combining big data and thick data (2020)

Policy Lab’s work develops data and insights into ideas for potential policy intervention which we can start to test as prototypes with real people. These operate at the ‘meso’ level (in the middle of the diagram above), informed by both the thick data from individual experiences and the big data at a population or national level. We have written a lot about prototyping for policy and are continuing to explore how you prototype a policy compared to say a digital service….(More)”.

The Wild Wild West of Data Hoarding in the Federal Government


ActiveNavigation: “There is a strong belief, both in the public and private sector, that the worst thing you can do with a piece of data is to delete it. The government stores all sorts of data, from traffic logs to home ownership statistics. Data is obviously incredibly important to the Federal Government – but storing large amounts of it poses significant compliance and security risks – especially with the rise of Nation State hackers. As the risk of being breached continues to rise, why is the government not tackling their data storage problem head on?

The Myth of “Free” Storage

Storage is cheap, especially compared to 10-15 years ago. Cloud storage has made it easier than ever to store swaths of information, creating what some call “digital landfills.” However, the true cost of storage isn’t in the ones and zeros sitting on the server somewhere. It’s the business cost.

As information stores continue to grow, the Federal Government’s ability to execute moving information to the correct place gets harder and harder, not to mention more expensive. The U.S. Government has a duty to provide accurate, up-to-date information to its taxpayers – meaning that sharing “bad data” is not an option.

The Association of Information and Image Management (AIIM) reports that half of an organization’s retained data has no value. So far, in 2019, through our work with Federal Agencies, we have discovered that this number, is in fact, low. Over 66% of data we’ve indexed, by the client’s definition, has fallen into that “junk” category. Eliminating junk data paves the way for greater accessibility, transparency and major financial savings. But what is “junk” data?

Redundant, Obsolete and Trivial (ROT) Data

Data is important – but if you can’t assign a value to it, it can become impossible to manage. Simply put, ROT data is digital information that an organization retains, that has no business or legal value. To be efficient from both a cyber hygiene and business perspective, the government needs to get better at purging their ROT data.

Again, purging data doesn’t just help with the hard cost of storage and backups, etc. For example, think about what needs to be done to answer a Freedom of Information Act (FOIA) request. You have a petabyte of data. You have at least a billion documents you need to funnel through to be able to respond to that FOIA request. By eliminating 50% of your ROT data, you probably have also reduced your FOIA response time by 50%.

Records and information governance, taken at face value, might seem fairly esoteric. It may not be as fun or as sexy as the new Space Force, but the reality is, the only way to know if the government is doing what it says it’s through records and information. You can’t answer an FOIA request if there’s no material. You can’t answer Congress if the material isn’t accurate. Being able to access timely, accurate information is critical. That’s why NARA is advocating a move to electronic records.…(More)”.

The future is intelligent: Harnessing the potential of artificial intelligence in Africa


Youssef Travaly and Kevin Muvunyi at Brookings: “…AI in particular presents countless avenues for both the public and private sectors to optimize solutions to the most crucial problems facing the continent today, especially for struggling industries. For example, in health care, AI solutions can help scarce personnel and facilities do more with less by speeding initial processing, triage, diagnosis, and post-care follow up. Furthermore, AI-based pharmacogenomics applications, which focus on the likely response of an individual to therapeutic drugs based on certain genetic markers, can be used to tailor treatments. Considering the genetic diversity found on the African continent, it is highly likely that the application of these technologies in Africa will result in considerable advancement in medical treatment on a global level.

In agricultureAbdoulaye Baniré Diallo, co-founder and chief scientific officer of the AI startup My Intelligent Machines, is working with advanced algorithms and machine learning methods to leverage genomic precision in livestock production models. With genomic precision, it is possible to build intelligent breeding programs that minimize the ecological footprint, address changing consumer demands, and contribute to the well-being of people and animals alike through the selection of good genetic characteristics at an early stage of the livestock production process. These are just a few examples that illustrate the transformative potential of AI technology in Africa.

However, a number of structural challenges undermine rapid adoption and implementation of AI on the continent. Inadequate basic and digital infrastructure seriously erodes efforts to activate AI-powered solutions as it reduces crucial connectivity. (For more on strategies to improve Africa’s digital infrastructure, see the viewpoint on page 67 of the full report). A lack of flexible and dynamic regulatory systems also frustrates the growth of a digital ecosystem that favors AI technology, especially as tech leaders want to scale across borders. Furthermore, lack of relevant technical skills, particularly for young people, is a growing threat. This skills gap means that those who would have otherwise been at the forefront of building AI are left out, preventing the continent from harnessing the full potential of transformative technologies and industries.

Similarly, the lack of adequate investments in research and development is an important obstacle. Africa must develop innovative financial instruments and public-private partnerships to fund human capital development, including a focus on industrial research and innovation hubs that bridge the gap between higher education institutions and the private sector to ensure the transition of AI products from lab to market….(More)”.

On Digital Disinformation and Democratic Myths


 David Karpf at MediaWell: “…How many votes did Cambridge Analytica affect in the 2016 presidential election? How much of a difference did the company actually make?

Cambridge Analytica has become something of a Rorschach test among those who pay attention to digital disinformation and microtargeted propaganda. Some hail the company as a digital Svengali, harnessing the power of big data to reshape the behavior of the American electorate. Others suggest the company was peddling digital snake oil, with outlandish marketing claims that bore little resemblance to their mundane product.

One thing is certain: the company has become a household name, practically synonymous with disinformation and digital propaganda in the aftermath of the 2016 election. It has claimed credit for the surprising success of the Brexit referendum and for the Trump digital strategy. Journalists such as Carole Cadwalladr and Hannes Grasseger and Mikael Krogerus have published longform articles that dive into the “psychographic” breakthroughs that the company claims to have made. Cadwalladr also exposed the links between the company and a network of influential conservative donors and political operatives. Whistleblower Chris Wylie, who worked for a time as the company’s head of research, further detailed how it obtained a massive trove of Facebook data on tens of millions of American citizens, in violation of Facebook’s terms of service. The Cambridge Analytica scandal has been a driving force in the current “techlash,” and has been the topic of congressional hearings, documentaries, mass-market books, and scholarly articles.

The reasons for concern are numerous. The company’s own marketing materials boasted about radical breakthroughs in psychographic targeting—developing psychological profiles of every US voter so that political campaigns could tailor messages to exploit psychological vulnerabilities. Those marketing claims were paired with disturbing revelations about the company violating Facebook’s terms of service to scrape tens of millions of user profiles, which were then compiled into a broader database of US voters. Cambridge Analytica behaved unethically. It either broke a lot of laws or demonstrated that old laws needed updating. When the company shut down, no one seemed to shed a tear.

But what is less clear is just how different Cambridge Analytica’s product actually was from the type of microtargeted digital advertisements that every other US electoral campaign uses. Many of the most prominent researchers warning the public about how Cambridge Analytica uses our digital exhaust to “hack our brains” are marketing professors, more accustomed to studying the impact of advertising in commerce than in elections. The political science research community has been far more skeptical. An investigation from Nature magazine documented that the evidence of Cambridge Analytica’s independent impact on voter behavior is basically nonexistent (Gibney 2018). There is no evidence that psychographic targeting actually works at the scale of the American electorate, and there is also no evidence that Cambridge Analytica in fact deployed psychographic models while working for the Trump campaign. The company clearly broke Facebook’s terms of service in acquiring its massive Facebook dataset. But it is not clear that the massive dataset made much of a difference.

At issue in the Cambridge Analytica case are two baseline assumptions about political persuasion in elections. First, what should be our point of comparison for digital propaganda in elections? Second, how does political persuasion in elections compare to persuasion in commercial arenas and marketing in general?…(More)”.

Three Examples of Data Empowerment


Blog by Michael Cañares: “It was a humid December afternoon in Banda Aceh, a bustling city in north Indonesia. Two women members of an education reform advocacy group were busy preparing infographics on how the city government was spending its education budget and its impact on service delivery quality in schools. The room was abuzz with questions and apprehension because the next day, the group would present its analysis on the data that they were able to access for the first time to education department officials. The analyses uncovered inefficiencies, poor school performance, ineffective allocation of resources, among others.

While worried about how the officials would react, almost everyone in the room was cheerful. One advocate told me she found the whole process liberating. She found it exhilarating to use government-published data to ask civil servants why the state of education in some schools was disappointing. “Armed with data, I am no longer afraid to speak my mind,” she said.

This was five years ago, but the memory has stuck with me. It was one of many experiences that inspired me to continue advocating for governments to publish data proactively, and searching for ways to use data to strengthen people’s voice on matters that are important to them.

Globally, there are many examples of how data has enabled people to advocate for their rights, demand better public services or hold governments to account. This blog post shares a few examples, focusing largely on how people are able to access and use data that shape their lives — the first dimension of how we characterize data empowerment….

Poverty Stoplight: People use their own data to improve their lives

Data Zetu: Giving borrowed data back to citizens

Check My School: Data-based community action to improve school performance…(More)”.

What is My Data Worth?


Ruoxi Jia at Berkeley artificial intelligence research: “People give massive amounts of their personal data to companies every day and these data are used to generate tremendous business values. Some economists and politicians argue that people should be paid for their contributions—but the million-dollar question is: by how much?

This article discusses methods proposed in our recent AISTATS and VLDB papers that attempt to answer this question in the machine learning context. This is joint work with David Dao, Boxin Wang, Frances Ann Hubis, Nezihe Merve Gurel, Nick Hynes, Bo Li, Ce Zhang, Costas J. Spanos, and Dawn Song, as well as a collaborative effort between UC Berkeley, ETH Zurich, and UIUC. More information about the work in our group can be found here.

What are the existing approaches to data valuation?

Various ad-hoc data valuation schemes have been studied in the literature and some of them have been deployed in the existing data marketplaces. From a practitioner’s point of view, they can be grouped into three categories:

  • Query-based pricing attaches values to user-initiated queries. One simple example is to set the price based on the number of queries allowed during a time window. Other more sophisticated examples attempt to adjust the price to some specific criteria, such as arbitrage avoidance.
  • Data attribute-based pricing constructs a price model that takes into account various parameters, such as data age, credibility, potential benefits, etc. The model is trained to match market prices released in public registries.
  • Auction-based pricing designs auctions that dynamically set the price based on bids offered by buyers and sellers.

However, existing data valuation schemes do not take into account the following important desiderata:

  • Task-specificness: The value of data depends on the task it helps to fulfill. For instance, if Alice’s medical record indicates that she has disease A, then her data will be more useful to predict disease A as opposed to other diseases.
  • Fairness: The quality of data from different sources varies dramatically. In the worst-case scenario, adversarial data sources may even degrade model performance via data poisoning attacks. Hence, the data value should reflect the efficacy of data by assigning high values to data which can notably improve the model’s performance.
  • Efficiency: Practical machine learning tasks may involve thousands or billions of data contributors; thus, data valuation techniques should be capable of scaling up.

With the desiderata above, we now discuss a principled notion of data value and computationally efficient algorithms for data valuation….(More)”.