Linux Foundation Debuts Community Data License Agreement


Press Release: “The Linux Foundation, the nonprofit advancing professional open source management for mass collaboration, today announced the Community Data License Agreement(CDLA) family of open data agreements. In an era of expansive and often underused data, the CDLA licenses are an effort to define a licensing framework to support collaborative communities built around curating and sharing “open” data.

Inspired by the collaborative software development models of open source software, the CDLA licenses are designed to enable individuals and organizations of all types to share data as easily as they currently share open source software code. Soundly drafted licensing models can help people form communities to assemble, curate and maintain vast amounts of data, measured in petabytes and exabytes, to bring new value to communities of all types, to build new business opportunities and to power new applications that promise to enhance safety and services.

The growth of big data analytics, machine learning and artificial intelligence (AI) technologies has allowed people to extract unprecedented levels of insight from data. Now the challenge is to assemble the critical mass of data for those tools to analyze. The CDLA licenses are designed to help governments, academic institutions, businesses and other organizations open up and share data, with the goal of creating communities that curate and share data openly.

For instance, if automakers, suppliers and civil infrastructure services can share data, they may be able to improve safety, decrease energy consumption and improve predictive maintenance. Self-driving cars are heavily dependent on AI systems for navigation, and need massive volumes of data to function properly. Once on the road, they can generate nearly a gigabyte of data every second. For the average car, that means two petabytes of sensor, audio, video and other data each year.

Similarly, climate modeling can integrate measurements captured by government agencies with simulation data from other organizations and then use machine learning systems to look for patterns in the information. It’s estimated that a single model can yield a petabyte of data, a volume that challenges standard computer algorithms, but is useful for machine learning systems. This knowledge may help improve agriculture or aid in studying extreme weather patterns.

And if government agencies share aggregated data on building permits, school enrollment figures, sewer and water usage, their citizens benefit from the ability of commercial entities to anticipate their future needs and respond with infrastructure and facilities that arrive in anticipation of citizens’ demands.

“An open data license is essential for the frictionless sharing of the data that powers both critical technologies and societal benefits,” said Jim Zemlin, Executive Director of The Linux Foundation. “The success of open source software provides a powerful example of what can be accomplished when people come together around a resource and advance it for the common good. The CDLA licenses are a key step in that direction and will encourage the continued growth of applications and infrastructure.”…(More)”.

Are you doing what’s needed to get the state to respond to its citizens? Or are you part of the problem?


Vanessa Herringshaw at Making All Voices Count: ” …I’ve been reading over the incredibly rich and practically-orientated research and practice papers already on the Making All Voices Count website, and some of those coming out soon. There’s a huge amount of useful and challenging learning, and I’ll be putting out a couple of papers summarising some important threads later this year.

But as different civic tech and accountability actors prepare to come together in Brighton for Making All Voices Count’s final learning event later this week, I’m going to focus here on three things that really stood out and would benefit from the kind of ‘group grappling’ that such a gathering can best support. And I aim to be provocative!

  1. Improving state responsiveness to citizens is a complex business – even more than we realised – and a lot more complex than most interventions are designed to address. If we don’t address this complexity, our interventions won’t work. And that risks making things worse, not better.
  2. It’s clear we need to make more of a shift from narrow, ‘tactical’ approaches to more complex, systems-based ‘strategic’ approaches. Thinking is developing on how to do this in theory. But it’s not clear that our current institutional approaches will support, or even allow, a major shift in practice.
  3. So when we each look at our individual roles and those of our organisations, are we part of the solution, or part of the problem?

Let’s look at each of these in turn….(More)”

Laboratories for news? Experimenting with journalism hackathons


Jan Lauren Boyles in Journalism: “Journalism hackathons are computationally based events in which participants create news product prototypes. In the ideal case, the gatherings are rooted in local community, enabling a wide set of institutional stakeholders (legacy journalists, hacker journalists, civic hackers, and the general public) to gather in conversation around key civic issues. This study explores how and to what extent journalism hackathons operate as a community-based laboratory for translating open data from practitioners to the public. Surfaced from in-depth interviews with event organizers encompassing nine countries, the findings illustrate that journalism hackathons are most successful when collaboration integrates civic organizations and community leaders….(More)”.

When Cartography Meets Disaster Relief


Mimi Kirk at CityLab: “Almost three weeks after Hurricane Maria hit Puerto Rico, the island is in a grim state. Fewer than 15 percent of residents have power, and much of the island has no clean drinking water. Delivery of food and other necessities, especially to remote areas, has been hampered by a variety of ills, including a lack of cellular service, washed-out roads, additional rainfall, and what analysts and Puerto Ricans say is a slow and insufficient response from the U.S. government.

Another issue slowing recovery? Maps—or lack of them. While pre-Maria maps of Puerto Rico were fairly complete, their level of detail was nowhere near that of other parts of the United States. Platforms such as Google Maps are more comprehensive on the mainland than on the island, explains Juan Saldarriaga, a research scholar at the Center for Spatial Research at Columbia University. This is because companies like Google often create maps for financial reasons, selling them to advertisers or as navigation devices, so areas that have less economic activity are given less attention.

This lack of detail impedes recovery efforts: Without basic information on the location of buildings, for instance, rescue workers don’t know how many people were living in an area before the hurricane struck—and thus how much aid is needed.

Crowdsourced mapping can help. Saldarriaga recently organized a “mapathon” at Columbia, in which volunteers examined satellite imagery of Puerto Rico and added missing buildings, roads, bridges, and other landmarks in the open-source platform OpenStreetMap. While some universities and other groups are hosting similar events, anyone with an internet connection and computer can participate.

Saldarriaga and his co-organizers collaborated with Humanitarian OpenStreetMap Team (HOT), a nonprofit that works to create crowdsourced maps for aid and development work. Volunteers like Saldarriaga largely drive HOT’s “crisis mapping” projects, the first of which occurred in 2010 after Haiti’s earthquake…(More)”.

Can Blockchain Bring Voting Online?


Ben Miller at Government Technology: “Hash chains are not a new concept in cryptography. They are, essentially, a long chain of data connected by values called hashes that prove the connection of each part to the next. By stringing all these pieces together and representing them in small values, then, one can represent a large amount of information without doing much. Josh Benaloh, a senior cryptographer for Microsoft Research and director of the International Association for Cryptologic Research, gives the rough analogy of taking a picture of a person, then taking another picture of that person holding the first picture, and so on. Loss of resolution aside, each picture would contain all the images from the previous pictures.

It’s only recently that people have found a way to extend the idea to commonplace applications. That happened with the advent of bitcoin, a digital “cryptocurrency” that has attained real-world value and become a popular exchange medium for ransomware attacks. The bitcoin community operates using a specific type of hash chain called a blockchain. It works by asking a group of users to solve complex problems as a sort of proof that bitcoin transactions took place, in exchange for a reward.

“Academics who have been looking at this for years, when they saw bitcoin, they said, ‘This can’t work, this has too many problems,’” Benaloh said. “It surprised everybody that this seems to work and to hold.”

But the blockchain concept is by no means limited to money. It’s simply a public ledger, a bulletin board meant to ensure accuracy based on the fact that everyone can see it — and what’s been done to it — at all times. It could be used to keep property records, or to provide an audit trail for how a product got from factory to buyer.

Or perhaps it could be used to prove the veracity and accuracy of digital votes in an election.

It is a potential solution to the problem of cybersecurity in online elections because the foundation of blockchain is the audit trail: If anybody tampered with votes, it would be easy to see and prove.

And in fact, blockchain elections have already been run in the U.S. — just not in the big leagues. Voatz, a Massachusetts-based startup that has struck up a partnership with one of the few companies in the country that actually builds voting systems, has used a blockchain paradigm to run elections for colleges, school boards, unions and other nonprofit and quasi-governmental groups. Perhaps its most high-profile endeavor was authenticating delegate badges at the 2016 Massachusetts Democratic Convention….

Rivest and Benaloh both talk about another online voting solution with much more enthusiasm. And much in the spirit of academia, the technology’s name is pragmatic rather than sleek and buzzworthy: end-to-end verifiable Internet voting (E2E-VIV).

It’s not too far off from blockchain in spirit, but it relies on a centralized approach instead of a decentralized one. Votes are sent from remote electronic devices to the election authority, most likely the secretary of state for the state the person is voting in, and posted online in an encrypted format. The person voting can use her decryption key to check that her vote was recorded accurately.

But there are no validating peers, no chain of blocks stretching back to the first vote….(More)”.

Paraguay’s transparency alchemists


Story by the Open Contracting Partnership: “….The “Cocido de oro” scandal is seen as part of a well-organized and well-informed youth movement that has sprung up in Paraguay in recent years. An equally dramatic controversyinvolving alleged corruption and unfair staff appointments at one of the country’s top public universities led to the resignation of the Chancellor and other senior staff in September 2015. Mostly high school and university students, they are no longer willing to tolerate the waste and corruption in public spending — a hangover from 35 years of authoritarian rule. They expect their government to be more open and accountable, and public decision-making processes to be more inclusive and democratic.

Thanks to government initiatives that have sought to give citizens greater access to information about public institutions, these students, along with investigative journalists and other civil society groups, are starting to engage actively in civic affairs. And they are data-savvy, basing recommendations on empirical evidence about government policies and processes, how they are implemented, and whether they are working.

Leading the pack is the country’s public procurement office, which runs a portal that ranks among the most open government data sources in the world. Together with information about budgets, public bodies’ payrolls, and other government data, this is helping Paraguayans to tackle some of the biggest long-standing problems faced by the government, like graft, overpricing, nepotism and influence-peddling….

The government recognizes there’s still a long way to go in their quest to open up public data. Few institutions have opened their databases or publish their data on an open data portal, and use of the data that has been published is still limited, according to a report on the country’s third OGP Action Plan. Priority data sets aren’t accessible in ways that meet the needs of civil society, the report adds.

And yet, the tremors of a tectonic shift in transparency and accountability in Paraguay are already being felt. In a short time, armed with access to information, citizens have started engaging with how public money is and should be spent.

The government is now doubling down on its strategy of fostering public participation, using cutting-edge technology to increase citizens’ access to data about their state institutions. Health, education, and municipal-level government, and procurement spending across these areas are being prioritized….(More).

Cross-sector Collaboration in Data Science for Social Good: Opportunities, Challenges, and Open Questions Raised by Working with Academic Researchers


Paper by presented by Anissa Tanweer and Brittany Fiore-Gartland at the Data Science for Social Good Conference: “Recent years have seen growing support for attempts to solve complex social problems through the use of increasingly available, increasingly combinable, and increasingly computable digital data. Sometimes referred to as “data science for social good” (DSSG), these efforts are not concentrated in the hands of any one sector of society. Rather, we see DSSG emerging as an inherently multi-sector and collaborative phenomenon, with key participants hailing from governments, nonprofit organizations, technology companies, and institutions of higher education. Based on three years of participant observation in a university-hosted DSSG program, in this paper we highlight academic contributions to multi-sector DSSG collaborations, including expertise, labor, ethics, experimentation, and neutrality. After articulating both the opportunities and challenges that accompany those contributions, we pose some key open questions that demand attention from participants in DSSG programs and projects. Given the emergent nature of the DSSG phenomenon, it is our contention that how these questions come to be answered will have profound implications for the way society is organized and governed….(More)”.

The Digital Social Innovation Manifesto


ChiC: “The unprecedented hyper connectivity enabled by digital technologies and the Internet are rapidly changing the opportunities we have to address some of the society’s biggest challenges: environmental preservation, reducing inequalities, fostering inclusion and putting in place sustainable economic models.
However, to make the most of these opportunities we need to move away from the current centralization of power by a small number of large tech companies and enable a much broader group of people and organisations to develop and share innovative digital solutions.

Across Europe, a growing movement of people is exploring opportunities for Digital Social Innovation (DSI), developing bottom-up solutions leveraging on participation, collaboration, decentralization, openness, and multi-disciplinarity. However, it is still at a relatively small scale, because of the little public and private investment in DSI, the limited experience in large-scale take-up of collective solutions, and the relative lack of skills of DSI actors (civil society) compared to commercial companies.

This Manifesto aims at fostering civic participation into democratic and social processes, increasing societal resilience and mutual trust as core element of the Digital Society. It provides recommendations for policy makers, to drive the development of the European Digital Single Market to fulfill first and foremost societal and sustainability challenges (rather than short-lived economic interests), with the help and engagement of all citizens.

This Manifesto reflects the views of a broad community of innovators, catalyzed by the coordination action ChiC, which is funded by the European Commission, within the context of the CAPS initiative. As such, it is open to incorporating incoming views and opinions from other stakeholders and it does not intend to promote the specific commercial interests of actors of any kind….(More)”

UN Opens New Office to Monitor AI Development and Predict Possible Threats


Interesting Engineering: “The United Nations has created a new office in the Netherlands dedicated to the monitoring and research of Artificial Intelligence (AI) technologies. The new office will collect information about the way in which AI is impacting the world. Researchers will have a particular focus on the way AI relates to global security but will also monitor the effects of job loss from AI and automation.

Irakli Beridze, a UN senior strategic adviser will head the office. They have described the new office saying, “A number of UN organisations operate projects involving robots and AI, such as the group of experts studying the role of autonomous military robots in the realm of conventional weapons. These are temporary measures. Ours is the first permanent UN office on this subject. We are looking at the risks as well as the advantages.”….He suggests that the speed of AI technology development is of primary concern. He explains, “This can make for instability if society does not adapt quickly enough. One of our most important tasks is to set up a network of experts from business, knowledge institutes, civil society organisations and governments. We certainly do not want to plead for a ban or a brake on technologies. We will also explore how new technology can contribute to the sustainable development goals of the UN. For this, we want to start concrete projects. We will not be a talking club.”…(More).

Ethical Guidelines for Applying Predictive Tools Within Human Services


MetroLab Network: “Predictive analytical tools are already being put to work within human service agencies to help make vital decisions about when and how to intervene in the lives of families and communities. The sector may not be entirely comfortable with this trend, but it should not be surprised. Predictive models are in wide use within the justice and education sectors and, more to the point, they work: risk assessment is fundamental to what social services do, and these tools can help agencies respond more quickly to prevent harm, to create more personalized interventions, and allocate scarce public resources to where they can do the most good.

There is also a strong case that predictive risk models (PRM) can reduce bias in decision-making. Designing a predictive model forces more explicit conversations about how agencies think about different risk factors and how they propose to guard against disadvantaging certain demographic or socioeconomic groups. And the standard that agencies are trying to improve upon is not perfect equity—it is the status quo, which is neither transparent nor uniformly fair. Risk scores do not eliminate the possibility of personal or institutional prejudice but they can make it more apparent by providing a common reference point.

That the use of predictive analytics in social services can reduce bias is not to say that it will. Careless or unskilled development of these predictive tools could worsen disparities among clients receiving social services. Child and civil rights advocates rightly worry about the potential for “net widening”—drawing more people in for unnecessary scrutiny by the government. They worry that rather than improving services for vulnerable clients, these models will replicate the biases in existing public data sources and expose them to greater trauma. Bad models scale just as quickly as good ones, and even the best of them can be misused.

The stakes here are real: for children and families that interact with these social systems and for the reputation of the agencies that turn to these tools. What, then, should a public leader know about risk modeling, and what lessons does it offer about how to think about data science, data stewardship, and the public interest?…(More)”.