Selected Readings on Crowdsourcing Data


The Living Library’s Selected Readings series seeks to build a knowledge base on innovative approaches for improving the effectiveness and legitimacy of governance. This curated and annotated collection of recommended works on the topic of crowdsourcing data was originally published in 2013.

As institutions seek to improve decision-making through data and put public data to use to improve the lives of citizens, new tools and projects are allowing citizens to play a role in both the collection and utilization of data. Participatory sensing and other citizen data collection initiatives, notably in the realm of disaster response, are allowing citizens to crowdsource important data, often using smartphones, that would be either impossible or burdensomely time-consuming for institutions to collect themselves. Civic hacking, often performed in hackathon events, on the other hand, is a growing trend in which governments encourage citizens to transform data from government and other sources into useful tools to benefit the public good.

Selected Reading List (in alphabetical order)

Annotated Selected Reading List (in alphabetical order)

Baraniuk, Chris. “Power Politechs.” New Scientist 218, no. 2923 (June 29, 2013): 36–39. http://bit.ly/167ul3J.

  • In this article, Baraniuk discusses civic hackers, “an army of volunteer coders who are challenging preconceptions about hacking and changing the way your government operates. In a time of plummeting budgets and efficiency drives, those in power have realised they needn’t always rely on slow-moving, expensive outsourcing and development to improve public services. Instead, they can consider running a hackathon, at which tech-savvy members of the public come together to create apps and other digital tools that promise to enhace the provision of healthcare, schools or policing.”
  • While recognizing that “civic hacking has established a pedigree that demonstrates its potential for positive impact,” Baraniuk argues that a “more rigorous debate over how this activity should evolve, or how authorities ought to engage in it” is needed.

Barnett, Brandon, Muki Hansteen Izora, and Jose Sia. “Civic Hackathon Challenges Design Principles: Making Data Relevant and Useful for Individuals and Communities.” Hack for Change, https://bit.ly/2Ge6z09.

  • In this paper, researchers from Intel Labs offer “guiding principles to support the efforts of local civic hackathon organizers and participants as they seek to design actionable challenges and build useful solutions that will positively benefit their communities.”
  • The authors proposed design principles are:
    • Focus on the specific needs and concerns of people or institutions in the local community. Solve their problems and challenges by combining different kinds of data.
    • Seek out data far and wide (local, municipal, state, institutional, non-profits, companies) that is relevant to the concern or problem you are trying to solve.
    • Keep it simple! This can’t be overstated. Focus [on] making data easily understood and useful to those who will use your application or service.
    • Enable users to collaborate and form new communities and alliances around data.

Buhrmester, Michael, Tracy Kwang, and Samuel D. Gosling. “Amazon’s Mechanical Turk A New Source of Inexpensive, Yet High-Quality, Data?” Perspectives on Psychological Science 6, no. 1 (January 1, 2011): 3–5. http://bit.ly/H56lER.

  • This article examines the capability of Amazon’s Mechanical Turk to act a source of data for researchers, in addition to its traditional role as a microtasking platform.
  • The authors examine the demographics of MTurkers and find that “MTurk participants are slightly more demographically diverse than are standard Internet samples and are significantly more diverse than typical American college samples; (b) participation is affected by compensation rate and task length, but participants can still be recruited rapidly and inexpensively; (c) realistic compensation rates do not affect data quality; and (d) the data obtained are at least as reliable as those obtained via traditional methods.”
  • The paper concludes that, just as MTurk can be a strong tool for crowdsourcing tasks, data derived from MTurk can be high quality while also being inexpensive and obtained rapidly.

Goodchild, Michael F., and J. Alan Glennon. “Crowdsourcing Geographic Information for Disaster Response: a Research Frontier.” International Journal of Digital Earth 3, no. 3 (2010): 231–241. http://bit.ly/17MBFPs.

  • This article examines issues of data quality in the face of the new phenomenon of geographic information being generated by citizens, in order to examine whether this data can play a role in emergency management.
  • The authors argue that “[d]ata quality is a major concern, since volunteered information is asserted and carries none of the assurances that lead to trust in officially created data.”
  • Due to the fact that time is crucial during emergencies, the authors argue that, “the risks associated with volunteered information are often outweighed by the benefits of its use.”
  • The paper examines four wildfires in Santa Barbara in 2007-2009 to discuss current challenges with volunteered geographical data, and concludes that further research is required to answer how volunteer citizens can be used to provide effective assistance to emergency managers and responders.

Hudson-Smith, Andrew, Michael Batty, Andrew Crooks, and Richard Milton. “Mapping for the Masses Accessing Web 2.0 Through Crowdsourcing.” Social Science Computer Review 27, no. 4 (November 1, 2009): 524–538. http://bit.ly/1c1eFQb.

  • This article describes the way in which “we are harnessing the power of web 2.0 technologies to create new approaches to collecting, mapping, and sharing geocoded data.”
  • The authors examine GMapCreator and MapTube, which allow users to do a range of map-related functions such as create new maps, archive existing maps, and share or produce bottom-up maps through crowdsourcing.
  • They conclude that “these tools are helping to define a neogeography that is essentially ‘mapping for the masses,’ while noting that there are many issues of quality, accuracy, copyright, and trust that will influence the impact of these tools on map-based communication.”

Kanhere, Salil S. “Participatory Sensing: Crowdsourcing Data from Mobile Smartphones in Urban Spaces.” In Distributed Computing and Internet Technology, edited by Chittaranjan Hota and Pradip K. Srimani, 19–26. Lecture Notes in Computer Science 7753. Springer Berlin Heidelberg. 2013. https://bit.ly/2zX8Szj.

  • This paper provides a comprehensive overview of participatory sensing — a “new paradigm for monitoring the urban landscape” in which “ordinary citizens can collect multi-modal data streams from the surrounding environment using their mobile devices and share the same using existing communications infrastructure.”
  • In addition to examining a number of innovative applications of participatory sensing, Kanhere outlines the following key research challenges:
    • Dealing with incomplete samples
    •  Inferring user context
    • Protecting user privacy
    • Evaluating data trustworthiness
    • Conserving energy

The Decay of American Political Institutions


in the American Interest: “Many political institutions in the United States are decaying. This is not the same thing as the broader phenomenon of societal or civilization decline, which has become a highly politicized topic in the discourse about America. Political decay in this instance simply means that a specific political process—sometimes an individual government agency—has become dysfunctional. This is the result of intellectual rigidity and the growing power of entrenched political actors that prevent reform and rebalancing. This doesn’t mean that America is set on a permanent course of decline, or that its power relative to other countries will necessarily diminish. Institutional reform is, however, an extremely difficult thing to bring about, and there is no guarantee that it can be accomplished without a major disruption of the political order. So while decay is not the same as decline, neither are the two discussions unrelated.
There are many diagnoses of America’s current woes. In my view, there is no single “silver bullet” cause of institutional decay, or of the more expansive notion of decline. In general, however, the historical context of American political development is all too often given short shrift in much analysis. If we look more closely at American history as compared to that of other liberal democracies, we notice three key structural characteristics of American political culture that, however they developed and however effective they have been in the past, have become problematic in the present.
The first is that, relative to other liberal democracies, the judiciary and the legislature (including the roles played by the two major political parties) continue to play outsized roles in American government at the expense of Executive Branch bureaucracies. Americans’ traditional distrust of government thus leads to judicial solutions for administrative problems. Over time this has become a very expensive and inefficient way to manage administrative requirements.
The second is that the accretion of interest group and lobbying influences has distorted democratic processes and eroded the ability of the government to operate effectively. What biologists label kin selection and reciprocal altruism (the favoring of family and friends with whom one has exchanged favors) are the two natural modes of human sociability. It is to these types of relationships that people revert when modern, impersonal government breaks down.
The third is that under conditions of ideological polarization in a federal governance structure, the American system of checks and balances, originally designed to prevent the emergence of too strong an executive authority, has become a vetocracy. The decision system has become too porous—too democratic—for its own good, giving too many actors the means to stifle adjustments in public policy. We need stronger mechanisms to force collective decisions but, because of the judicialization of government and the outsized role of interest groups, we are unlikely to acquire such mechanisms short of a systemic crisis. In that sense these three structural characteristics have become intertwined….
In short, the problems of American government flow from a structural imbalance between the strength and competence of the state, on the one hand, and the institutions that were originally designed to constrain the state, on the other. There is too much law and too much “democracy”, in the form of legislative intervention, relative to American state capacity. Some history can make this assertion clearer….
In well-functioning governance systems, moreover, a great deal of deliberation occurs not just in legislatures but within bureaucracies. This is not a matter of bureaucrats simply talking to one another, but rather a complex series of consultations between government officials and businesses, outside implementers and service providers, civil society groups, the media and other sources of information about societal interests and opinions. The Congress wisely mandated consultation in the landmark 1946 Administrative Procedures Act, which requires regulatory agencies to publicly post proposed rule changes and to solicit comment about them. But these consultative procedures have become highly routinized and pro forma, with actual decisions being the outcome not of genuine deliberation, but of political confrontations between well organized interest groups….”

The Age of Democracy


Xavier Marquez at Abandoned Footnotes: “This is the age of democracy, ideologically speaking. As I noted in an earlier post, almost every state in the world mentions the word “democracy” or “democratic” in its constitutional documents today. But the public acknowledgment of the idea of democracy is not something that began just a few years ago; in fact, it goes back much further, all the way back to the nineteenth century in a surprising number of cases.
Here is a figure I’ve been wanting to make for a while that makes this point nicely (based on data graciously made available by the Comparative Constitutions Project). The figure shows all countries that have ever had some kind of identifiable constitutional document (broadly defined) that mentions the word “democracy” or “democratic” (in any context – new constitution, amendment, interim constitution, bill of rights, etc.), arranged from earliest to latest mention. Each symbol represents a “constitutional event” – a new constitution adopted, an amendment passed, a constitution suspended, etc. – and colored symbols indicate that the text associated with the constitutional event in question mentions the word “democracy” or “democratic”…
The earliest mentions of the word “democracy” or “democratic” in a constitutional document occurred in Switzerland and France in 1848, as far as I can tell.[1] Participatory Switzerland and revolutionary France look like obvious candidates for being the first countries to embrace the “democratic” self-description; yet the next set of countries to embrace this self-description (until the outbreak of WWI) might seem more surprising: they are all Latin American or Caribbean (Haiti), followed by countries in Eastern Europe (various bits and pieces of the Austro-Hungarian empire), Southern Europe (Portugal, Spain), Russia, and Cuba. Indeed, most “core” countries in the global system did not mention democracy in their constitutions until much later, if at all, despite many of them having long constitutional histories; even French constitutions after the fall of the Second Republic in 1851 did not mention “democracy” until after WWII. In other words, the idea of democracy as a value to be publicly affirmed seems to have caught on first not in the metropolis but in the periphery. Democracy is the post-imperial and post-revolutionary public value par excellence, asserted after national liberation (as in most of the countries that became independent after WWII) or revolutions against hated monarchs (e.g., Egypt 1956, Iran 1979, both of them the first mentions of democracy in these countries but not their first constitutions).

Today only 16 countries have ever failed to mention their “democratic” character in their constitutional documents (Australia, Brunei, Denmark, Japan, Jordan, Malaysia, Monaco, Nauru, Oman, Samoa, Saudi Arabia, Singapore, Tonga, the United Kingdom, the USA, and Vatican City).[2] And no country that has ever mentioned “democracy” in an earlier constitutional document fails to mention it in its current constitutional documents (though some countries in the 19th and early 20th centuries went back and forth – mentioning democracy in one constitution, not mentioning it in the next). Indeed, after WWII the first mention of democracy in constitutions tended to be contemporaneous with the first post-independence constitution of the country; and with time, even countries with old and settled constitutional traditions seem to be more and more likely to mention “democracy” or “democratic” in some form as amendments or bills of rights accumulate (e.g., Belgium in 2013, New Zealand in 1990, Canada in 1982, Finland in 1995). The probability of a new constitution mentioning “democracy” appears to be asymptotically approaching 1. To use the language of biology, the democratic “meme” has nearly achieved “fixation” in the population, despite short-term fluctuations, and despite the fact that there appears to be no particular correlation between a state calling itself democratic and actually being democratic, either today or in the past.[3]
Though the actual measured level of democracy around the world has trended upwards (with some ups and downs) over the last two centuries, I don’t think this is the reason why the idea of democracy has achieved near-universal recognition in public documents. Countries do not first become democratic and then call themselves democracies; if anything, most mentions of democracy seem to be rather aspirational, if not entirely cynical. (Though many constitutions that mention democracy were also produced by people who seem to have been genuinely committed to some such ideal, even if the regimes that eventually developed under these constitutions were not particularly democratic). What we see, instead, is a broad process in which earlier normative claims about the basis of authority – monarchical, imperial, etc. – get almost completely replaced, regardless of the country’s cultural context, by democratic claims, regardless of the latter’s effectiveness as an actual basis for authority or the existence of working mechanisms for participation or vertical accountability. (These democratic claims to authority also sometimes coexist in uneasy tension with other claims to authority based on divine revelation, ideological knowledge, or tradition, invented or otherwise; consider the Chinese constitution‘s claims about the “people’s democratic dictatorship” led by the CCP).
I thus suspect the conquest of ideological space by “democratic” language did not happen just because democratic claims to authority (especially in the absence of actual democracy) have proved more persuasive than other claims to authority. Rather, I think the same processes that resulted in the emergence of modern national communities – e.g. the rituals associated with nationalism, which tended to “sacralize” a particular kind of imagined community – led to the symbolic production of the nation not only as the proper object of government but also as its proper active agent (the people, actively ruling itself), regardless of whether or not “the people” had any ability to rule or even to exercise minimal control over the rulers.[4] There thus seems to have been a kind of co-evolution of symbols of nationality and symbols of democracy, helped along by the practice/ritual of drafting constitutions and approving them through plebiscites or other forms of mass politics, a ritual that already makes democratic assumptions about “social contracts.” The question is whether the symbolic politics of democracy eventually has any sort of impact on actual institutions. But more on this later….”

Index: Measuring Impact with Evidence


The Living Library Index – inspired by the Harper’s Index – provides important statistics and highlights global trends in governance innovation. This installment focuses on measuring impact with evidence and was originally published in 2013.

United States

  • Amount per $100 of government spending that is backed by evidence that the money is being spent wisely: less than $1
  • Number of healthcare treatments delivered in the U.S. that lack evidence of effectiveness: more than half
  • How much of total U.S. healthcare expenditure is spent to determine what works: less than 0.1 percent
  • Number of major U.S. federal social programs evaluated since 1990 using randomized experiments and found to have “weak or no positive effects”: 9 out of 10
  • Year the Coalition for Evidence-Based Policy was set up to work with federal policymakers to advance evidence-based reforms in major U.S. social programs: 2001
  • Year the Program Assessment Rating Tool (PART) was introduced by President Bush’s Office of Management and Budget (OMB): 2002
    • Out of about 1,000 programs assessed, number found to be effective in 2008: 19%
    • Percentage of programs that could not be assessed due to insufficient data: 17%
    • Amount spent on the Even Start Family Literacy Program, rated ineffective by PART, over the life of the Bush administration: more than $1 billion
  •  Year Washington State legislature began using Washington State Institute for Public Policy’s estimates on how “a portfolio of evidence-based and economically sound programs . . . could affect the state’s crime rate, the need to build more prisons, and total criminal-justice spending”: 2007
    • Amount invested by legislature in these programs: $48 million
    • Amount saved by the legislature: $250 million
  • Number of U.S. States in a pilot group working to adapt The Pew-MacArthur Results First Initiative, based on the Washington State model, to make performance-based policy decisions: 14
  • Net savings in health care expenditure by using the Transitional Care Model, which meets the Congressionally-based Top Tier Evidence Standard: $4,000 per patient
  • Number of states that conducted “at least some studies that evaluated multiple program or policy options for making smarter investments of public dollars” between 2008-2011: 29
  • Number of states that reported that their cost-benefit analysis influenced policy decisions or debate: 36
  • Date the Office of Management and Budget issued a memorandum proposing new evaluations and advising agencies to include details on determining effectiveness of their programs, link disbursement to evidence, and support evidence-based initiatives: 2007
  • Percentage increase in resources for innovation funds that use a tiered model for evidence, according to the President’s FY14 budget: 44% increase
  • Amount President Obama proposed in his FY 2013 budget to allocate in existing funding to Performance Partnerships “in which states and localities would be given the flexibility to propose better ways to combine federal resources in exchange for greater accountability for results”:  $200 million
  • Amount of U.S. federal program funding that Harvard economist Jeffrey Liebman suggests be directed towards evaluations of outcomes: 1%
  • Amount of funding the City of New York has committed for evidence-based research and development initiatives through its Center for Economic Opportunity: $100 million a year

Internationally

  • How many of the 30 OECD countries in 2005-6 have a formal requirement by law that the benefits of regulation justify the costs: half
    • Number of 30 OECD member countries in 2008 that reported quantifying benefits to regulations: 16
    • Those who reported quantifying costs: 24
  • How many members make up the Alliance for Useful Evidence, a network that “champion[s]  evidence, the opening up of government data for interrogation and use, alongside the sophistication in research methods and their applications”: over 1,000
  • Date the UK government, the ESRC and the Big Lottery Fund announced plans to create a network of ‘What Works’ evidence centres: March 2013
  • Core funding for the What Works Centre for Local Economic Growth: £1m p.a. over an initial three year term
  • How many SOLACE Summit members in 2012 were “very satisfied” with how Research and Intelligence resources support evidence-based decision-making: 4%
    • Number of areas they identified for improving evidence-based decision-making: 5
    • Evaluation of the impact of past decisions: 46% of respondents
    • Benchmarking data with other areas: 39%
    • assessment of options available: 33% 
    • how evidence is presented: 29% 
    • Feedback on public engagement and consultation: 25%
  •  Number of areas for improvement for Research and Intelligence staff development identified at the SOLACE Summit: 6
    • Strengthening customer insight and data analysis: 49%
    • Impact evaluation: 48%
    • Strategic/corporate thinking/awareness: 48%
    • Political acumen: 46%
    • Raising profile/reputation of the council for evidence-based decisions: 37%
    • Guidance/mentoring on use of research for other officers: 25%

Sources

Selected Readings on Smart Disclosure


The Living Library’s Selected Readings series seeks to build a knowledge base on innovative approaches for improving the effectiveness and legitimacy of governance. This curated and annotated collection of recommended works on the topic of smart disclosure was originally published in 2013.

While much attention is paid to open data, data transparency need not be managed by a simple On/Off switch: It’s often desirable to make specific data available to the public or individuals in targeted ways. A prime example is the use of government data in Smart Disclosure, which provides consumers with data they need to make difficult marketplace choices in health care, financial services, and other important areas. Governments collect two kinds of data that can be used for Smart Disclosure: First, governments collect information on services of high interest to consumers, and are increasingly releasing this kind of data to the public. In the United States, for example, the Department of Health and Human Services collects and releases online data on health insurance options, while the Department of Education helps consumers understand the true cost (after financial aid) of different colleges. Second, state, local, or national governments hold information on consumers themselves that can be useful to them. In the U.S., for example, the Blue Button program was launched to help veterans easily access their own medical records.

Selected Reading List (in alphabetical order)

Annotated Selected Reading List (in alphabetical order)

Better Choices: Better Deals Report on Progress in the Consumer Empowerment Strategy. Progress Report. Consumer Empowerment Strategy. United Kingdom: Department for Business Innovation & Skills, December 2012. http://bit.ly/17MqnL3.

  • The report details the progress made through the United Kingdom’s consumer empowerment strategy, Better Choices: Better Deals. The plan seeks to mitigate knowledge imbalances through information disclosure programs and targeted nudges.
  • The empowerment strategy’s four sections demonstrate the potential benefits of Smart Disclosure: 1. The power of information; 2. The power of the crowd; 3. Helping the vulnerable; and 4. A new approach to Government working with business.
Braunstein, Mark L.,. “Empowering the Patient.” In Health Informatics in the Cloud, 67–79. Springer Briefs in Computer Science. Springer New York Heidelberg Dordrecht London, 2013. https://bit.ly/2UB4jTU.
  • This book discusses the application of computing to healthcare delivery, public health and community based clinical research.
  • Braunstein asks and seeks to answer critical questions such as: Who should make the case for smart disclosure when the needs of consumers are not being met? What role do non-profits play in the conversation on smart disclosure especially when existing systems (or lack thereof) of information provision do not work or are unsafe?

Brodi, Elisa. “Product-Attribute Information” and “Product-Use Information”: Smart Disclosure and New Policy Implications for Consumers’ Protection. SSRN Scholarly Paper. Rochester, NY: Social Science Research Network, September 4, 2012. http://bit.ly/17hssEK.

  • This paper from the Research Area of the Bank of Italy’s Law and Economics Department “surveys the literature on product use information and analyzes whether and to what extent Italian regulator is trying to ensure consumers’ awareness as to their use pattern.” Rather than focusing on the type of information governments can release to citizens, Brodi proposes that governments require private companies to provide valuable use pattern information to citizens to inform decision-making.
  • The form of regulation proposed by Brodi and other proponents “is based on a basic concept: consumers can be protected if companies are forced to disclose data on the customers’ consumption history through electronic files.”
National Science and Technology Council. Smart Disclosure and Consumer Decision Making: Report of the Task Force on Smart Disclosure. Task Force on Smart Disclosure: Information and Efficiency in Consumer Markets. Washington, DC: United States Government: Executive Office of the President, May 30, 2013. http://1.usa.gov/1aamyoT.
    • This inter-agency report is a comprehensive description of smart disclosure approaches being used across the Federal Government. The report not only highlights the importance of making data available to consumers but also to innovators to build better options for consumers.
  • In addition to providing context about government policies that guide smart disclosure initiatives, the report raises questions about what parties have influence in this space.

“Policies in Practice: The Download Capability.” Markle Connecting for Health Work Group on Consumer Engagement, August 2010. http://bit.ly/HhMJyc.

  • This report from the Markle Connecting for Health Work Group on Consumer Engagement — the creator of the Blue Button system for downloading personal health records — features a “set of privacy and security practices to help people download their electronic health records.”
  • To help make health information easily accessible for all citizens, the report lists a number of important steps:
    • Make the download capability a common practice
    • Implement sound policies and practices to protect individuals and their information
    • Collaborate on sample data sets
    • Support the download capability as part of Meaningful Use and qualified or certified health IT
    • Include the download capability in procurement requirements.
  • The report also describes the rationale for the development of the Blue Button — perhaps the best known example of Smart Disclosure currently in existence — and the targeted release of health information in general:
    • Individual access to information is rooted in fair information principles and law
    • Patients need and want the information
    • The download capability would encourage innovation
    • A download capability frees data sources from having to make many decisions about the user interface
    • A download capability would hasten the path to standards and interoperability.
Sayogo, Djoko Sigit, and Theresa A. Pardo. “Understanding Smart Data Disclosure Policy Success: The Case of Green Button.” In Proceedings of the 14th Annual International Conference on Digital Government Research, 72–81. New York: ACM New York, NY, USA, 2013. http://bit.ly/1aanf1A.
  • This paper from the Proceedings of the 14th Annual International Conference on Digital Government Research explores the implementation of the Green Button Initiative, analyzing qualitative data from interviews with experts involved in Green Button development and implementation.
  • Moving beyond the specifics of the Green Button initiative, the authors raise questions on the motivations and success factors facilitating successful collaboration between public and private organizations to support smart disclosure policy.

Thaler, Richard H., and Will Tucker. “Smarter Information, Smarter Consumers.” Harvard Business Review January – February 2013. The Big Idea. http://bit.ly/18gimxw.

  • In this article, Thaler and Tucker make three key observations regarding the challenges related to smart disclosure:
    • “We are constantly confronted with information that is highly important but extremely hard to navigate or understand.”
    • “Repeated attempts to improve disclosure, including efforts to translate complex contracts into “plain English,” have met with only modest success.”
    • “There is a fundamental difficulty of explaining anything complex in simple terms. Most people find it difficult to write instructions explaining how to tie a pair of shoelaces.

Social movements and their technologies. wiring social change


New book by Stefania Milan: “Social Movements and Their Technologies. Wiring Social Change explores the interplay between social movements and their “liberated technologies”. It analyzes the rise of low-power radio stations and radical internet projects (“emancipatory communication practices”) as a political subject, focusing on the sociological and cultural processes at play. It provides an overview of the relationship between social movements and technology and investigates what is behind the communication infrastructure that made possible the main protest events of the past 15 years. In doing so, Stefania Milan illustrates how contemporary social movements organize in order to create autonomous alternatives to communication systems and networks and how they contribute to change the way people communicate in daily life, as well as try to change communication policy from the grassroots….
Read an excerpt from the book
“The question of infrastructure might sound trivial in times of abundance of “free” social media, microblogging platforms and apps allowing people to voice their opinions and share pictures and videos at will, and at virtually no cost. But we often forget that these platforms are owned and controlled by media and telecoms corporations whose agenda focuses on profit and corporate interests rather than participation, empowerment, and social justice. With this in mind, in recent decades activist groups have increasingly challenged media corporations and state-owned broadcasters on their own terrain. They have created alternatives to existing communication infrastructure by setting up community radio and television stations, and alternative websites for self-produced information. Such grassroots media have allowed broader swathes of the citizenry to access media production and secure communication channels. They have become what DeeDee Halleck calls “infrastructures of resistance” (2002, p. 191) to the neoliberal order in the media realm.”

DataViva: a Big Data Engine for the Brazilian Economy


Piece by André Victor dos Santos Barrence and Cesar A. Hidalgo: “The current Internet paradigm in which one can search about anything and retrieve information is absolutely empowering. We can browse files, websites and indexes and effortlessly reach good amount of information. Google, for instance, was explicitly built on a library analogy available to everyone. However, it is a world where information that should be easily accessible is still hidden in unfriendly databases, and that the best-case scenario is finding few snippets of information embedded within the paragraphs of a report. But is this the way it should be? Or is this just the world we are presently stuck with?
The last decade has been particularly marked by an increasing hype on big data and analytics, mainly fueled by those who are interested in writing narratives on the topic but not necessarily coding about it, even when data itself is not the problem.
Let’s take the case of governments. Governments have plenty of data and in many cases it is actually public (at least in principle). Governments “know” how many people work in every occupation, in every industry and in every location; they know their salaries, previous employers and education history. From a pure data perspective all that is embedded in tax, social security records or annual registrations. From a more pragmatic perspective, it is still inaccessible and hidden even when it is legally open the public. We live in a world where the data is there, but where the statistics and information are not.
The state government of Minas Gerais in Brazil (3rd economy the country, territory larger than France and 20 millions inhabitants) made an important step in that direction by releasing DataViva.info, a platform that opens data for exports and occupations for the entire formal sector of the Brazilian economy through more than 700 million interactive visualizations. Instead of poorly designed tables and interfaces, it guides users to answer questions or freely discover locations, industries and occupations in Brazil that are of interest to them. DataViva allows users to explore simple questions such as the evolution of exports in the last decade for each of the 5,567 municipalities in the country, or highly specific queries, for instance, the average salaries paid to computer scientists working in the software development industry in Belo Horizonte, the state capital of Minas.
DataViva’s visualizations are built on the idea that the industrial and economic activity development of locations is highly path dependent. This means that locations are more likely to be successful at developing industries and activities that are related to the ones already existing, since it indicates the existence of labor inputs, and other capabilities, that are specific and that can often be redeployed to a few related industries and activities. Thus, it informs the processes by which opportunities can be explored and prospective pathways for greater prosperity.
The idea that information is key for the functioning of economies is at least as old as Friedrich Hayek’s seminal paper The Use of Knowledge in Society from 1945. According to Hayek, prices help coordinate economic activities by providing information about the wants and needs of goods and services. Yet, the price information can only serve as a signal as long as people know those prices. Maybe the salaries for engineers in the municipality of Betim (Minas Gerais) are excellent and indicate a strong need for them? But who would have known how many engineers are there in Betim and what are their average salaries?
But the remaining question is: why is Minas Gerais making all of this public data easily available? More than resorting to the contemporary argument of open government Minas understands this is extremely valuable information for investors searching for business opportunities, entrepreneurs pursuing new ventures or workers looking for better career prospects. Lastly, the ultimate goal of DataViva is to provide a common ground for open discussions, moving away from the information deprived idea of central planning and into a future where collaborative planning might become the norm. It is a highly creative attempt to renew public governance for the 21st century.
Despite being a relatively unknown state outside of Brazil, by releasing a platform as DataViva, Minas is providing a strong signal about where in world governments are really pushing forward innovation rather than simply admiring and copying solutions that used to come from trendsetters in the developed world. It seems like real innovation isn’t necessarily taking place in Washington, Paris or London anymore.”
 

Data Mining Reveals the Secret to Getting Good Answers


arXiv: “If you want a good answer, ask a decent question. That’s the startling conclusion to a study of online Q&As.

If you spend any time programming, you’ll probably have come across the question and answer site Stack Overflow. The site allows anybody to post a question related to programing and receive answers from the community. And it has been hugely successful. According to Alexa, the site is the 3rd most popular Q&A site in the world and 79th most popular website overall.
But this success has naturally led to a problem–the sheer number of questions and answers the site has to deal with. To help filter this information, users can rank both the questions and the answers, gaining a reputation for themselves as they contribute.
Nevertheless, Stack Overflow still struggles to weed out off topic and irrelevant questions and answers. This requires considerable input from experienced moderators. So an interesting question is whether it is possible to automate the process of weeding out the less useful question and answers as they are posted.
Today we get an answer of sorts thanks to the work of Yuan Yao at the State Key Laboratory for Novel Software Technology in China and a team of buddies who say they’ve developed an algorithm that does the job.
And they say their work reveals an interesting insight: if you want good answers, ask a decent question. That may sound like a truism, but these guys point out that there has been no evidence to support this insight, until now.
…But Yuan and co digged a little deeper. They looked at the correlation between well received questions and answers. And they discovered that these are strongly correlated.
A number of factors turn out to be important. These include the reputation of the person asking the question or answering it, the number of previous questions or answers they have posted, the popularity of their input in the recent past along with measurements like the length of the question and its title.
Put all this into a number cruncher and the system is able to predict the quality score of the question and its expected answers. That allows it to find the best questions and answers and indirectly the worst ones.
There are limitations to this approach, of course…In the meantime, users of Q&A sites can learn a significant lesson from this work. If you want good answers, first formulate a good question. That’s something that can take time and experience.
Perhaps the most interesting immediate application of this new work might be as a teaching tool to help with this learning process and to boost the quality of questions and answers in general.
See also: http://arxiv.org/abs/1311.6876: Want a Good Answer? Ask a Good Question First!”

El Hacker Cívico: Civic-Minded Techies Gain Sway with Government in Mexico and Beyond


in the Huffington Post: “A handful of young hackers looked up from their laptops when Jorge Soto burst into the upstairs office they shared in an old Mexico City house one morning last spring. Soto wanted to be sure they’d seen the front-page headline then flying across Twitter: Mexico’s congress was set to spend 115 million pesos (then US $9.3 million) on a mobile app that would let 500 lawmakers track legislative affairs from their cellphones — more than a hundred times what such software could cost.
To many in Mexico, what became known as the “millionaire’s app” was just the latest in an old story of bloated state spending; but Soto and his colleagues saw a chance to push a new approach instead. In two days, they’d covered their white office walls with ideas for a cheaper alternative and launched an online contest that drew input from more than 150 software developers and designers, producing five open-source options in two weeks.
Lawmakers soon insisted they’d never known about the original app, which had been quietly approved by a legislative administrative board; and a congressional spokesman rushed to clarify that the project had been suspended. Invited to pitch their alternatives to Congress, a half-dozen young coders took the podium in a sloping auditorium at the legislature. The only cost for their work: a 11,500-peso (then US $930) prize for the winner.
“We didn’t just ‘angry tweet,’ we actually did something,” Soto, a 28-year-old IT engineer and social entrepreneur, said at the time. “Citizens need to understand democracy beyond voting every few years, and government needs to understand that we’re willing to participate.”
Seven months later, Mexico’s president appears to have heard them, hiring Soto and nine others to launch one of the world’s first federal civic innovation offices, part of a broader national digital agenda to be formally unveiled today. Building on a model pioneered in a handful of U.S. cities since 2010, Mexico’s civic innovation team aims to integrate so-called “civic hackers” with policy experts already inside government — to not only build better technology, but to seed a more tech-minded approach to problem-solving across federal processes and policy. What began as outside activism is slowly starting to creep into government….Mexico’s app incident reflects a common problem in that process: wasteful spending by non-techie bureaucrats who don’t seem to know what they’re buying — at best, out-of-touch; at worst, party to crony contracting; and overseen, if at all, by officials even less tech-savvy than themselves. Citizens, in contrast, are adopting new technologies faster than much of the public sector, growing the gap between the efficiency and accountability that they expect as private consumers, and the bureaucratic, buggy experience that government still provides.
To break that cycle, a movement of community-minded “civic hackers” like Soto has stepped into the void, offering their own low-cost tools to make government more efficient, collaborative and transparent….The platform, named Codeando Mexico, has since hosted more than 30 civic-themed challenges.
With Soto as an advisor, the team seized on the scandal surrounding the “millionare’s app” to formally launch in March, calling for help “taking down the Mexican tech mafia” – a play on the big, slow software makers that dominate public contracting around the world. In that, Codeando Mexico tackled a central civic-tech target: procurement, widely considered one of the public spheres ripest for reform. Its goal, according to Wilhelmy, was to replace clueless or “compadrismo” crony contracting with micro-procurement, swapping traditional suppliers for leaner teams of open-source coders who can release and revise what they build in near real-time. “It’s like the Robin Hood of procurement: You take money that’s being spent on big projects and bring it to the developer community, giving them an opportunity to work on stuff that matters,” he said. “There’s a whole taboo around software: government thinks it has to be expensive. We’re sending a message that there are different ways to do this; it shouldn’t cost so much.”
The maker of the costly congressional app in question, Mexico City consultancy Pulso Legislativo, insisted last spring that its hefty price tag didn’t reflect its software as much as the aggregated data and analysis behind it. But critics were quick to note that Mexican lawmakers already had access to similar data compiled by at least five publicly-funded research centers – not to mention from INFOPAL, a congressional stats system with its own mobile application. With Mexico then in the midst of a contentious telecom reform, the public may’ve been especially primed to pounce on any hint of corruption or wasteful IT spending. Codeando Mexico saw an opening.
So it was that a crew of young coders, almost all under 23-years-old, traipsed into the legislature, a motley mix of suits and skinny jeans, one-by-one pitching a panel of judges that included the head of the congressional Science and Technology Committee. Drawing on public data culled by local transparency groups, their Android and iOS apps – including the winner, “Diputados” — allowed citizens to track and opine on pending bills and to map and contact their representatives — still a relatively new concept in Mexico’s young democracy.”

Open Data: From ‘Platform’ to ‘Program’


Engaging Cities: “A few months ago, Dutch designer Mark van der Net launched OSCity.nl, a highly interesting example of what can be done with open data. At first, it looks like a mapping tool. The interface shows a – beautifully designed – map of The Netherlands, color coded according to whatever open data set the user selects, varying from geographical height to the location of empty office buildings. As such it is an example of a broader current in which artists, citizens, ngos and business actors have build online tools to visualize all kinds of data, varying from open government data to collaboratively produced data sets focused on issues like environmental pollution.
What makes OSCity interesting is that it allows users to intuitively map various datasets in combination with each other in so called ‘map stories’. For instance, a map of empty office space can be combined with maps of urban growth and decline, the average renting price per square meter of office space, as well as map that displays the prices of houses for sale. The intersection of those maps shows you where empty office spaces are offered at or below half the price of regular houses and apartments. The result is thus not just an aesthetically pleasing state of affairs, but an action map. Policy makers, developers and citizens can use the insights produced by the map to find empty offices that are worthwhile to turn into houses.
There are two important lessons we can learn from this project. First, it shows the importance of programs like OSCity to make open data platforms operationable for various actors. Over the last few years governments and other organizations have started to open up their datasets, often accompanied with high expectations of citizen empowerment and greater transparency of governments. However, case studies have showed that opening up data and building an open platform is only a first step. Dawes and Helbig have shown that various stakeholders have various needs in terms of standards and protocols, whereas both citizens and government officials need the relevant skills to be able to understand and operate upon the data. ‘Vast amounts of useful information are contained in government data systems’, they write, ‘but the systems themselves are seldom designed for use beyond the collecting agency’s own needs.’ In other words: what is needed to deliver on the expectations of open data, is not only a platform – a publicly available database – but also what I have called ‘programs’ – online tools with intuitive interfaces that make this data intelligible and actionable in concert with the needs of the public.
There is a second issue that OSCity raises. As Jo Bates has pointed out, the main question is: who exactly is empowered through programs like this? Will ‘programs’ that make data operationable work for citizens? Or will their procedures, standards and access be organized to benefit corporate interests? These do not have to be necessarily contradicting, but if the goal is to empower citizens, it is important to engage them as stakeholders in the design of these programs.”