When Data Science Destabilizes Democracy and Facilitates Genocide


Rachel Thomas in Fast.AI onWhat is the ethical responsibility of data scientists?”…What we’re talking about is a cataclysmic change… What we’re talking about is a major foreign power with sophistication and ability to involve themselves in a presidential election and sow conflict and discontent all over this country… You bear this responsibility. You’ve created these platforms. And now they are being misusedSenator Feinstein said this week in a senate hearing. Who has created a cataclysmic change? Who bears this large responsibility? She was talking to executives at tech companies and referring to the work of data scientists.

Data science can have a devastating impact on our world, as illustrated by inflammatory Russian propaganda being shown on Facebook to 126 million Americans leading up to the 2016 election (and the subject of the senate hearing described above) or by lies spread via Facebook that are fueling ethnic cleansing in Myanmar. Over half a million Rohinyga have been driven from their homes due to systematic murder, rape, and burning. Data science is foundational to Facebook’s newsfeed, in determining what content is prioritized and who sees what….

The examples of bias in data science are myriad and include:

You can do awesome and meaningful things with data science (such as diagnosing cancer, stopping deforestation, increasing farm yields, and helping patients with Parkinson’s disease), and you can (often unintentionally) enable terrible things with data science, as the examples in this post illustrate. Being a data scientist entails both great opportunity, as well as great responsibility, to use our skills to not make the world a worse place. Ultimately, doing data science is about humans, not just the users of our products, but everyone who will be impacted by our work. (More)”.

The Challenge of VR to the Liberal Democratic Order


Paper by Edward Castronova: “The rapid expansion of virtual reality (VR) technology in the years 2016-2021 awakens a significant constitutional issue. In a liberal democratic order, rule is by consent of the governed. In the medium-term future, many of the governed will be immersed fully within VR environments, environments which, we are told, will provide entertainment of extraordinary power. These people will be happy. Happy people do not demand change. Yet there surely will be a change as VR takes hold: The quality of life will erode. People fully immersed in VR will come to be isolated, sedentary, and unhealthy. Objectively speaking, this is nothing to be desired. Subjectively, however, it will seem to be wonderful. The people themselves will be happy, and they will resist interference. At the moment this matter concerns only a few thousand nerds, but trends in technology and entertainment point to a future in which many people will be happily living awful, VR-dominated lives. How then will the liberal democratic order promote human well-being while remaining a liberal and democratic order?…(More)”

Understanding Corporate Data Sharing Decisions: Practices, Challenges, and Opportunities for Sharing Corporate Data with Researchers


Leslie Harris at the Future of Privacy Forum: “Data has become the currency of the modern economy. A recent study projects the global volume of data to grow from about 0.8 zettabytes (ZB) in 2009 to more than 35 ZB in 2020, most of it generated within the last two years and held by the corporate sector.

As the cost of data collection and storage becomes cheaper and computing power increases, so does the value of data to the corporate bottom line. Powerful data science techniques, including machine learning and deep learning, make it possible to search, extract and analyze enormous sets of data from many sources in order to uncover novel insights and engage in predictive analysis. Breakthrough computational techniques allow complex analysis of encrypted data, making it possible for researchers to protect individual privacy, while extracting valuable insights.

At the same time, these newfound data sources hold significant promise for advancing scholarship and shaping more impactful social policies, supporting evidence-based policymaking and more robust government statistics, and shaping more impactful social interventions. But because most of this data is held by the private sector, it is rarely available for these purposes, posing what many have argued is a serious impediment to scientific progress.

A variety of reasons have been posited for the reluctance of the corporate sector to share data for academic research. Some have suggested that the private sector doesn’t realize the value of their data for broader social and scientific advancement. Others suggest that companies have no “chief mission” or public obligation to share. But most observers describe the challenge as complex and multifaceted. Companies face a variety of commercial, legal, ethical, and reputational risks that serve as disincentives to sharing data for academic research, with privacy – particularly the risk of reidentification – an intractable concern. For companies, striking the right balance between the commercial and societal value of their data, the privacy interests of their customers, and the interests of academics presents a formidable dilemma.

To be sure, there is evidence that some companies are beginning to share for academic research. For example, a number of pharmaceutical companies are now sharing clinical trial data with researchers, and a number of individual companies have taken steps to make data available as well. What is more, companies are also increasingly providing open or shared data for other important “public good” activities, including international development, humanitarian assistance and better public decision-making. Some are contributing to data collaboratives that pool data from different sources to address societal concerns. Yet, it is still not clear whether and to what extent this “new era of data openness” will accelerate data sharing for academic research.

Today, the Future of Privacy Forum released a new study, Understanding Corporate Data Sharing Decisions: Practices, Challenges, and Opportunities for Sharing Corporate Data with ResearchersIn this report, we aim to contribute to the literature by seeking the “ground truth” from the corporate sector about the challenges they encounter when they consider making data available for academic research. We hope that the impressions and insights gained from this first look at the issue will help formulate further research questions, inform the dialogue between key stakeholders, and identify constructive next steps and areas for further action and investment….(More)”.

Bot.Me: A revolutionary partnership


PWC Consumer Intelligence Series: “The modern world has been shaped by the technological revolutions of the past, like the Industrial Revolution and the Information Revolution. The former redefined the way the world values both human and material resources; the latter redefined value in terms of resources while democratizing information. Today, as technology progresses even further, value is certain to shift again, with a focus on sentiments more intrinsic to the human experience: thinking, creativity, and problem-solving. AI, shorthand for artificial intelligence, defines technologies emerging today that can understand, learn, and then act based on that information. Forms of AI in use today include digital assistants, chatbots, and machine learning.

Today, AI works in three ways:

  • Assisted intelligence, widely available today, improves what people and organizations are already doing. A simple example, prevalent in cars today, is the GPS navigation program that offers directions to drivers and adjusts to road conditions.
  • Augmented intelligence, emerging today, enables people and organizations to do things they couldn’t otherwise do. For example, the combination of programs that organize cars in ride-sharing services enables businesses that could not otherwise exist.
  • Autonomous intelligence, being developed for the future, establishes machines that act on their own. An example of this will be self-driving vehicles, when they come into widespread use.

With a market projected to reach $70 billion by 2020, AI is poised to have a transformative effect on consumer, enterprise, and government markets around the world. While there are certainly obstacles to overcome, consumers believe that AI has the potential to assist in medical breakthroughs, democratize costly services, elevate poor customer service, and even free up an overburdened workforce. Some tech optimists believe AI could create a world where human abilities are amplified as machines help mankind process, analyze, and evaluate the abundance of data that creates today’s world, allowing humans to spend more time engaged in high-level thinking, creativity, and decision-making. Technological revolutions, like the Industrial Revolution and the Information Revolution, didn’t happen overnight. In fact, people in the midst of those revolutions often didn’t even realize they were happening, until history was recorded later.

That is where we find ourselves today, in the very beginning of what some are calling the “augmented age.” Just like humans in the past, it is up to mankind to find the best ways to leverage these machine revolutions to help the world evolve. As Isaac Asimov, the prolific science fiction writer with many works on AI mused, “No sensible decision can be made any longer without taking into account not only the world as it is, but the world as it will be.” As a future with AI approaches, it’s important to understand how people think of it today, how it will amplify the world tomorrow, and what guiding principles will be required to navigate this monumental change….(More)”.

Augmented CI and Human-Driven AI: How the Intersection of Artificial Intelligence and Collective Intelligence Could Enhance Their Impact on Society


Blog by Stefaan Verhulst: “As the technology, research and policy communities continue to seek new ways to improve governance and solve public problems, two new types of assets are occupying increasing importance: data and people. Leveraging data and people’s expertise in new ways offers a path forward for smarter decisions, more innovative policymaking, and more accountability in governance. Yet, unlocking the value of these two assets not only requires increased availability and accessibility (through, for instance, open data or open innovation), it also requires innovation in methodology and technology.

The first of these innovations involves Artificial Intelligence (AI). AI offers unprecedented abilities to quickly process vast quantities of data that can provide data-driven insights to address public needs. This is the role it has for example played in New York City, where FireCast, leverages data from across the city government to help the Fire Department identify buildings with the highest fire risks. AI is also considered to improve education, urban transportation,  humanitarian aid and combat corruption, among other sectors and challenges.

The second area is Collective Intelligence (CI). Although it receives less attention than AI, CI offers similar potential breakthroughs in changing how we govern, primarily by creating a means for tapping into the “wisdom of the crowd” and allowing groups to create better solutions than even the smartest experts working in isolation could ever hope to achieve. For example, in several countries patients’ groups are coming together to create new knowledge and health treatments based on their experiences and accumulated expertise. Similarly, scientists are engaging citizens in new ways to tap into their expertise or skills, generating citizen science – ranging from mapping our solar system to manipulating enzyme models in a game-like fashion.

Neither AI nor CI offer panaceas for all our ills; they each pose certain challenges, and even risks.  The effectiveness and accuracy of AI relies substantially on the quality of the underlying data as well as the human-designed algorithms used to analyse that data. Among other challenges, it is becoming increasingly clear how biases against minorities and other vulnerable populations can be built into these algorithms. For instance, some AI-driven platforms for predicting criminal recidivism significantly over-estimate the likelihood that black defendants will commit additional crimes in comparison to white counterparts. (for more examples, see our reading list on algorithmic scrutiny).

In theory, CI avoids some of the risks of bias and exclusion because it is specifically designed to bring more voices into a conversation. But ensuring that that multiplicity of voices adds value, not just noise, can be an operational and ethical challenge. As it stands, identifying the signal in the noise in CI initiatives can be time-consuming and resource intensive, especially for smaller organizations or groups lacking resources or technical skills.

Despite these challenges, however, there exists a significant degree of optimism  surrounding both these new approaches to problem solving. Some of this is hype, but some of it is merited—CI and AI do offer very real potential, and the task facing both policymakers, practitioners and researchers is to find ways of harnessing that potential in a way that maximizes benefits while limiting possible harms.

In what follows, I argue that the solution to the challenge described above may involve a greater interaction between AI and CI. These two areas of innovation have largely evolved and been researched separately until now. However, I believe that there is substantial scope for integration, and mutual reinforcement. It is when harnessed together, as complementary methods and approaches, that AI and CI can bring the full weight of technological progress and modern data analytics to bear on our most complex, pressing problems.

To deconstruct that statement, I propose three premises (and subsequent set of research questions) toward establishing a necessary research agenda on the intersection of AI and CI that can build more inclusive and effective approaches to governance innovation.

Premise I: Toward Augmented Collective Intelligence: AI will enable CI to scale

Premise II: Toward Human-Driven Artificial Intelligence: CI will humanize AI

Premise III: Open Governance will drive a blurring between AI and CI

…(More)”.

Democracy Needs a Reboot for the Age of Artificial Intelligence


Katharine Dempsey at The Nation: “…A healthy modern democracy requires ordinary citizens to participate in public discussions about rapidly advancing technologies. We desperately need new policies, regulations, and safety nets for those displaced by machines. With computing power accelerating exponentially, the scale of AI’s significance is still not being fully internalized. The 2017 McKinsey Global Initiative report “A Future that Works” predicts that AI and advanced robotics could automate roughly half of all work globally by 2055, but, McKinsey notes, “this could happen up to 20 years earlier or later depending on the various factors, in addition to other wider economic conditions.”

Granted, the media are producing more articles focused on artificial intelligence, but too often these pieces veer into hysterics. Wired magazine labeled this year’s coverage “The Great Tech Panic of 2017.” We need less fear-mongering and more rational conversation. Dystopian narratives, while entertaining, can also be disorienting. Skynet from the Terminatormovies is not imminent. But that doesn’t mean there aren’t hazards ahead….

Increasingly, to thoughtfully discuss ethics, politics, or business, the general population needs to pay attention to AI. In 1989, Ursula Franklin, the distinguished German-Canadian experimental physicist, delivered a series of lectures titled “The Real World of Technology.” Franklin opened her lectures with an important observation: “The viability of technology, like democracy, depends in the end on the practice of justice and on the enforcements of limits to power.”

For Franklin, technology is not a neutral set of tools; it can’t be divorced from society or values. Franklin further warned that “prescriptive technologies”—ones that isolate tasks, such as factory-style work—find their way into our social infrastructures and create modes of compliance and orthodoxy. These technologies facilitate top-down control….(More)”.

The Human Strategy


A Video Conversation with Sandy Pentland at Edge: “People have lots of capabilities; they know lots of things about the world; they can perceive things in a human way. What would happen if you had a network of people where you could reinforce the ones that were helping and maybe discourage the ones that weren’t?

That begins to sound like a society or a company. We all live in a human social network. We’re reinforced for things that seem to help everybody and discouraged from things that are not appreciated. Culture is something that comes from a sort of human AI, the function of reinforcing the good and penalizing the bad, but applied to humans and human problems. Once you realize that you can take this general framework of AI and create a human AI, the question becomes, what’s the right way to do that? Is it a safe idea? Is it completely crazy?…(More)”.

Growing the artificial intelligence industry in the UK


Summary from an independent review, carried out by Professor Dame Wendy Hall and Jérôme Pesenti: “Increased use of Artificial Intelligence (AI) can bring major social and economic benefits to the UK. With AI, computers can analyse and learn from information at higher accuracy and speed than humans can. AI offers massive gains in efficiency and performance to most or all industry sectors, from drug discovery to logistics. AI is software that can be integrated into existing processes, improving them, scaling them, and reducing their costs, by making or suggesting more accurate decisions through better use of information.

It has been estimated that AI could add an additional USD $814 billion (£630bn) to the UK economy by 2035, increasing the annual growth rate of GVA from 2.5 to 3.9%.

Our vision is for the UK to become the best place in the world for businesses developing and deploying AI to start, grow and thrive, to realise all the benefits the technology offers….

Key factors have combined to increase the capability of AI in recent years, in particular:

  • New and larger volumes of data
  • Supply of experts with the specific high level skills
  • Availability of increasingly powerful computing capacity. The barriers to achieving performance have fallen significantly, and continue to fall.

To continue developing and applying AI, the UK will need to increase ease of access to data in a wider range of sectors. This Review recommends:

  • Development of data trusts, to improve trust and ease around sharing data
  • Making more research data machine readable
  • Supporting text and data mining as a standard and essential tool for research.

Skilled experts are needed to develop AI, and they are in short supply. To develop more AI, the UK will need a larger workforce with deep AI expertise, and more development of lower level skills to work with AI. …

Increasing uptake of AI means increasing demand as well as supply through a better understanding of what AI can do and where it could be applied. This review recommends:

  • An AI Council to promote growth and coordination in the sector
  • Guidance on how to explain decisions and processes enabled by AI
  • Support for export and inward investment
  • Guidance on successfully applying AI to drive improvements in industry
  • A programme to support public sector use of AI
  • Funded challenges around data held by public organisations.

Our work has indicated that action in these areas could deliver a step-change improvement in growth of UK AI. This report makes the 18 recommendations listed in full below, which describe how Government, industry and academia should work together to keep the UK among the world leaders in AI…(More)”

Linux Foundation Debuts Community Data License Agreement


Press Release: “The Linux Foundation, the nonprofit advancing professional open source management for mass collaboration, today announced the Community Data License Agreement(CDLA) family of open data agreements. In an era of expansive and often underused data, the CDLA licenses are an effort to define a licensing framework to support collaborative communities built around curating and sharing “open” data.

Inspired by the collaborative software development models of open source software, the CDLA licenses are designed to enable individuals and organizations of all types to share data as easily as they currently share open source software code. Soundly drafted licensing models can help people form communities to assemble, curate and maintain vast amounts of data, measured in petabytes and exabytes, to bring new value to communities of all types, to build new business opportunities and to power new applications that promise to enhance safety and services.

The growth of big data analytics, machine learning and artificial intelligence (AI) technologies has allowed people to extract unprecedented levels of insight from data. Now the challenge is to assemble the critical mass of data for those tools to analyze. The CDLA licenses are designed to help governments, academic institutions, businesses and other organizations open up and share data, with the goal of creating communities that curate and share data openly.

For instance, if automakers, suppliers and civil infrastructure services can share data, they may be able to improve safety, decrease energy consumption and improve predictive maintenance. Self-driving cars are heavily dependent on AI systems for navigation, and need massive volumes of data to function properly. Once on the road, they can generate nearly a gigabyte of data every second. For the average car, that means two petabytes of sensor, audio, video and other data each year.

Similarly, climate modeling can integrate measurements captured by government agencies with simulation data from other organizations and then use machine learning systems to look for patterns in the information. It’s estimated that a single model can yield a petabyte of data, a volume that challenges standard computer algorithms, but is useful for machine learning systems. This knowledge may help improve agriculture or aid in studying extreme weather patterns.

And if government agencies share aggregated data on building permits, school enrollment figures, sewer and water usage, their citizens benefit from the ability of commercial entities to anticipate their future needs and respond with infrastructure and facilities that arrive in anticipation of citizens’ demands.

“An open data license is essential for the frictionless sharing of the data that powers both critical technologies and societal benefits,” said Jim Zemlin, Executive Director of The Linux Foundation. “The success of open source software provides a powerful example of what can be accomplished when people come together around a resource and advance it for the common good. The CDLA licenses are a key step in that direction and will encourage the continued growth of applications and infrastructure.”…(More)”.