Reinventing Participation: Civic Agency and the Web Environment


New paper by Peter Dahlgren: “Participation is a key concept in the vocabulary of democracy, and can encompass a variety of dimensions. Moreover, it can be shaped by a range of different factors; my emphasis here is on the significance of the web environment in this regard. I first situate participation against the backdrop of democracy’s contemporary developments, including the onslaught of neolibealism. From there I offer a set of parameters that can help us grasp participation both conceptually and empirically: trajectory, visibility, voice , and sociality, and relate these to the affordances of the digital media. Thereafter I explore the cultural resources necessary for the facilitation of participation; for this I make use of a six-dimensional model of civic cultures. My discussion focuses on two of the dimensions, practices and identities; I again relate these to the web environment. I conclude with a dilemma that online democratic participation faces, namely what I call the isolation of the solo sphere, yet affirm that we are justified in maintaining a guarded optimism about the future of participation.”

Selected Readings on Crowdsourcing Data


The Living Library’s Selected Readings series seeks to build a knowledge base on innovative approaches for improving the effectiveness and legitimacy of governance. This curated and annotated collection of recommended works on the topic of crowdsourcing data was originally published in 2013.

As institutions seek to improve decision-making through data and put public data to use to improve the lives of citizens, new tools and projects are allowing citizens to play a role in both the collection and utilization of data. Participatory sensing and other citizen data collection initiatives, notably in the realm of disaster response, are allowing citizens to crowdsource important data, often using smartphones, that would be either impossible or burdensomely time-consuming for institutions to collect themselves. Civic hacking, often performed in hackathon events, on the other hand, is a growing trend in which governments encourage citizens to transform data from government and other sources into useful tools to benefit the public good.

Selected Reading List (in alphabetical order)

Annotated Selected Reading List (in alphabetical order)

Baraniuk, Chris. “Power Politechs.” New Scientist 218, no. 2923 (June 29, 2013): 36–39. http://bit.ly/167ul3J.

  • In this article, Baraniuk discusses civic hackers, “an army of volunteer coders who are challenging preconceptions about hacking and changing the way your government operates. In a time of plummeting budgets and efficiency drives, those in power have realised they needn’t always rely on slow-moving, expensive outsourcing and development to improve public services. Instead, they can consider running a hackathon, at which tech-savvy members of the public come together to create apps and other digital tools that promise to enhace the provision of healthcare, schools or policing.”
  • While recognizing that “civic hacking has established a pedigree that demonstrates its potential for positive impact,” Baraniuk argues that a “more rigorous debate over how this activity should evolve, or how authorities ought to engage in it” is needed.

Barnett, Brandon, Muki Hansteen Izora, and Jose Sia. “Civic Hackathon Challenges Design Principles: Making Data Relevant and Useful for Individuals and Communities.” Hack for Change, https://bit.ly/2Ge6z09.

  • In this paper, researchers from Intel Labs offer “guiding principles to support the efforts of local civic hackathon organizers and participants as they seek to design actionable challenges and build useful solutions that will positively benefit their communities.”
  • The authors proposed design principles are:
    • Focus on the specific needs and concerns of people or institutions in the local community. Solve their problems and challenges by combining different kinds of data.
    • Seek out data far and wide (local, municipal, state, institutional, non-profits, companies) that is relevant to the concern or problem you are trying to solve.
    • Keep it simple! This can’t be overstated. Focus [on] making data easily understood and useful to those who will use your application or service.
    • Enable users to collaborate and form new communities and alliances around data.

Buhrmester, Michael, Tracy Kwang, and Samuel D. Gosling. “Amazon’s Mechanical Turk A New Source of Inexpensive, Yet High-Quality, Data?” Perspectives on Psychological Science 6, no. 1 (January 1, 2011): 3–5. http://bit.ly/H56lER.

  • This article examines the capability of Amazon’s Mechanical Turk to act a source of data for researchers, in addition to its traditional role as a microtasking platform.
  • The authors examine the demographics of MTurkers and find that “MTurk participants are slightly more demographically diverse than are standard Internet samples and are significantly more diverse than typical American college samples; (b) participation is affected by compensation rate and task length, but participants can still be recruited rapidly and inexpensively; (c) realistic compensation rates do not affect data quality; and (d) the data obtained are at least as reliable as those obtained via traditional methods.”
  • The paper concludes that, just as MTurk can be a strong tool for crowdsourcing tasks, data derived from MTurk can be high quality while also being inexpensive and obtained rapidly.

Goodchild, Michael F., and J. Alan Glennon. “Crowdsourcing Geographic Information for Disaster Response: a Research Frontier.” International Journal of Digital Earth 3, no. 3 (2010): 231–241. http://bit.ly/17MBFPs.

  • This article examines issues of data quality in the face of the new phenomenon of geographic information being generated by citizens, in order to examine whether this data can play a role in emergency management.
  • The authors argue that “[d]ata quality is a major concern, since volunteered information is asserted and carries none of the assurances that lead to trust in officially created data.”
  • Due to the fact that time is crucial during emergencies, the authors argue that, “the risks associated with volunteered information are often outweighed by the benefits of its use.”
  • The paper examines four wildfires in Santa Barbara in 2007-2009 to discuss current challenges with volunteered geographical data, and concludes that further research is required to answer how volunteer citizens can be used to provide effective assistance to emergency managers and responders.

Hudson-Smith, Andrew, Michael Batty, Andrew Crooks, and Richard Milton. “Mapping for the Masses Accessing Web 2.0 Through Crowdsourcing.” Social Science Computer Review 27, no. 4 (November 1, 2009): 524–538. http://bit.ly/1c1eFQb.

  • This article describes the way in which “we are harnessing the power of web 2.0 technologies to create new approaches to collecting, mapping, and sharing geocoded data.”
  • The authors examine GMapCreator and MapTube, which allow users to do a range of map-related functions such as create new maps, archive existing maps, and share or produce bottom-up maps through crowdsourcing.
  • They conclude that “these tools are helping to define a neogeography that is essentially ‘mapping for the masses,’ while noting that there are many issues of quality, accuracy, copyright, and trust that will influence the impact of these tools on map-based communication.”

Kanhere, Salil S. “Participatory Sensing: Crowdsourcing Data from Mobile Smartphones in Urban Spaces.” In Distributed Computing and Internet Technology, edited by Chittaranjan Hota and Pradip K. Srimani, 19–26. Lecture Notes in Computer Science 7753. Springer Berlin Heidelberg. 2013. https://bit.ly/2zX8Szj.

  • This paper provides a comprehensive overview of participatory sensing — a “new paradigm for monitoring the urban landscape” in which “ordinary citizens can collect multi-modal data streams from the surrounding environment using their mobile devices and share the same using existing communications infrastructure.”
  • In addition to examining a number of innovative applications of participatory sensing, Kanhere outlines the following key research challenges:
    • Dealing with incomplete samples
    •  Inferring user context
    • Protecting user privacy
    • Evaluating data trustworthiness
    • Conserving energy

Fort McMoney Online Game-Documentary Puts Fate Of Alberta Oilsands In Players' Hands


The Huffington Post: “The fate of the Alberta oilsands is now in the hands of the people. An interactive web documentary-game, titled “Fort McMoney,” launched Monday, inviting players into an immersive online experience set far north in the oil town of Fort McMurray, Alta.
Players explore the city and connect with key players in the oil industry, environmental activists as well as those living and working in the city and surrounding oil patch. Players learn the town’s environmental, cultural, political, social and economic concerns.
Every week for four weeks, players will explore different themes and issues of concern in the oil patch, virtually walking through the city to interview residents, executives and activists. At the end of each week, players vote in a referendum and try to convince other players of their opinions. The results of the referendum will decide the course of the game – for instance, users may decide to make the environment a priority over economy, or vice versa.
Votes will be tallied each Sunday evening, and Fort McMurray will change accordingly…
Fort McMoney, a joint project by the National Film Board and Montreal-based Toxa and Franco-German TV network Arte, is available in English, French and German. It can be played on a computer browser or tablet and requires players to register with Facebook or Twitter beyond the first segment.”

'Fix-It Squads': prototyping a way to better work with business in fixing problems


DesignGov (AU): “Can you really understand a problem if you have no experience with it? And can you really fix a problem if you don’t understand it?
The public sector plays a key role in the business environment, and sometimes that includes the generation and the resolution of problems faced by businesses. Our research as part of the business and government interactions project suggests that when businesses face issues that relate to multiple government agencies and/or multiple jurisdictions, it can be difficult for them to convey their experience. A problem that is very real for them can seem distributed and minor to the parts of the public sector that are connected with it. It is also difficult for public servants to get across the ‘whole’ of the problem and what can be done about it, when the solution may require coordination across agencies and where it may be hard to prioritise competing issues.
Through the ‘Fix-It Squads’ concept, we’re investigating how problem resolution processes might be improved:

  • To give public servants the opportunity to become immersed in the problem at hand and to share in the lived experience of it
  • To give businesses the opportunity to explain and show the problem as they experience it, rather than in the terms of public sector agencies (who might be contributing to the problem).

We are seeking your help and participation in prototyping. This post gives a quick overview of the ‘Fix-It Squads’ concept, asks for your help with the prototyping, and provides an expanded explanation of what Fix-It Squads might involve and why something like them are needed (in addition to the description provided in the Lost in Translation report and the associated prototyping prospectus).
FixIt Squads Synopsis
 

What do you know?


Article in the Financial Times by Eric Openshaw John Hagel: “Talent holds the key to company performance. From business units to Finance to IT, getting the right skills to the right place at the right time is a constant challenge. Now consider that workplace technologies are becoming obsolete faster, and the useful life of many skills is shorter. Workers at all levels need to be able to learn and relearn rapidly to adapt to and anticipate changing demands. Recruitment and retention initiatives can’t address the need, nor can standardised training programs and knowledge management.
New technologies offer an opportunity to rethink both talent development and traditional knowledge management and integrate learning directly into the daily work experience. Virtual platforms enable workers to connect with each other to solve problems across distributed work settings and beyond organisational boundaries.
The preponderance of sensors and advanced analytics today make it more possible than ever to collect and share individual’s real-time performance in a variety of settings. Sensors and the integration of social platforms into work allows for knowledge and experiences to be captured automatically, as they occur, rather than depending on compliance and coerced participation to populate reputation profiles and knowledge management databases. These technologies support rich, context-specific learning and participation driven by momentum as users discover and create value.
Training programs and knowledge management have a place, but they may not deliver the skilled workers needed to the right place and time in a rapidly changing environment. Pre-developed content quickly becomes obsolete or lacks the context to make it relevant to the individual. Perhaps more importantly, classroom training and knowledge databases tend to focus on the commonalities between work, the standard processes and practices, when in fact workers spend most of their days dealing with the exceptions that don’t fit into the standard processes and systems, whether it’s a one-off shipping request or a customer who can’t make your software work with their hardware. Workers typically get better at handling the non-standard aspects of their work through on-the-job experience…

In a recent paper, we detailed nine principles that help to create the type of environment that fosters learning and improvement. The following three principles demonstrate how technology can play an important role in enabling this on-the-job learning and in amplifying the learning, especially across a virtual workforce.

Real-time feedback for individuals and teams. For workers to improve performance and learn what works or doesn’t work, they need to have a context-specific understanding of what is expected and how they are doing relative to others, in the moment rather than three- or six-months down the line.
At virtual call-centre LiveOps, the independent agents see customer and program-specific metrics on an online dashboard. These metrics define the level of performance necessary for agents to remain eligible to take calls for a program and are continuously updated, providing real-time feedback after each call so that agents can see how they are doing relative to their peer group and where they can improve. This level of performance data transparency creates a meritocracy, as agents are compared to their peers and rewarded based on their relative performance.
Smart capture and share. In any work setting, a great deal of information is generated and exchanged in meetings, conversations, instant messages, and email. Easy access to that information helps foster collaboration, solve problems, and improve business processes. At SAP Community Network (SCN), intelligent cataloguing of insights from discussion forums, tagged for searchability, helps make the right information available at the right time to those who need it without requiring the burdensome documentation associated with typical knowledge management. Other users can search for solutions in the context of the original problem posed, as well as through related discussions that may have led to an ultimate solution. Instead of days of internal debate or experimentation, the typical time to receive a response is 17 minutes.
Helping workers make relevant connections. In a typical organisation, physical or virtual, it can be difficult to know who everyone is and what their experience, expertise, and interests are, and the typical knowledge and resource management tools that require individuals to maintain profiles rarely see the level of continuing compliance and participation to make them useful. Instead, workers tend to fall back on relationships. They seek help and learn from those already known to them.
Now, virtual platforms, such as the one used by Odesk, a global online workplace, automatically generate detailed, up-to-date profiles. These profiles include the contractor’s cumulative and historical ratings and hourly wages for each completed project as well as the scores for any tests or certifications. The same type of automatic, action-based reputation profiles can be used internally to facilitate assessing and connecting with the right co-workers for the job at hand…”

Behavioural Public Policy


New book by Adam Oliver (Cambridge University Press): “How can individuals best be encouraged to take more responsibility for their well-being and their environment or to behave more ethically in their business transactions? Across the world, governments are showing a growing interest in using behavioural economic research to inform the design of nudges which, some suggest, might encourage citizens to adopt beneficial patterns of behaviour. In this fascinating collection, leading academic economists, psychologists and philosophers reflect on how behavioural economic findings can be used to help inform the design of policy initiatives in the areas of health, education, the environment, personal finances and worker remuneration. Each chapter is accompanied by a shorter ‘response’ that provides critical commentary and an alternative perspective. This accessible book will interest academic researchers, graduate students and policy-makers across a range of disciplinary perspectives.”

Regulatory Democracy Reconsidered: The Policy Impact of Public Participation Requirements


Paper by Neal D. Woods in the Journal of Public Administration Research and Theory :  “A broad range of procedural mechanisms designed to promote public involvement in regulatory decision making have been instituted at all levels of government. Depending upon the literature one consults, one could conclude that these procedures (1) enhance regulatory stringency by fostering access by previously underrepresented groups, (2) reduce regulatory stringency by institutionalizing access by regulated industries, (3) could either increase or decrease stringency depending on the relative strength of organized interests in the agency’s external environment, or (4) have no effect. This study investigates whether mechanisms designed to promote public involvement in administrative rulemaking affect the stringency of US state environmental regulation. The results suggest that requirements to provide public notice of agency rulemaking do not have a significant effect on the regulatory compliance costs imposed on industry, but mechanisms that provide direct access to rulemaking processes serve to decrease these costs. This effect is evident for access both to the agencies promulgating environmental regulations and to external entities reviewing these regulations. For promulgating agencies, the effect does not appear to be conditional on the relative power of societal interests. The results provide some evidence, however, that political officials respond to the strength of environmental and industry groups when reviewing agency regulations.”

When Nudges Fail: Slippery Defaults


New paper by Lauren E. Willis “Inspired by the success of “automatic enrollment” in increasing participation in defined contribution retirement savings plans, policymakers have put similar policy defaults in place in a variety of other contexts, from checking account overdraft coverage to home-mortgage escrows. Internet privacy appears poised to be the next arena. But how broadly applicable are the results obtained in the retirement savings context? Evidence from other contexts indicates two problems with this approach: the defaults put in place by the law are not always sticky, and the people who opt out may be those who would benefit the most from the default. Examining the new default for consumer checking account overdraft coverage reveals that firms can systematically undermine each of the mechanisms that might otherwise operate to make defaults sticky. Comparing the retirement-savings default to the overdraft default, four boundary conditions on the use of defaults as a policy tool are apparent: policy defaults will not be sticky when (1) motivated firms oppose them, (2) these firms have access to the consumer, (3) consumers find the decision environment confusing, and (4) consumer preferences are uncertain. Due to constitutional and institutional constraints, government regulation of the libertarian-paternalism variety is unlikely to be capable of overcoming these bounds. Therefore, policy defaults intended to protect individuals when firms have the motivation and means to move consumers out of the default are unlikely to be effective unless accompanied by substantive regulation. Moreover, the same is likely to be true of “nudges” more generally, when motivated firms oppose them.”

What Government Can and Should Learn From Hacker Culture


in The Atlantic: “Can the open-source model work for federal government? Not in every way—for security purposes, the government’s inner workings will never be completely open to the public. Even in the inner workings of government, fears of triggering the next Wikileaks or Snowden scandal may scare officials away from being more open with one another. While not every area of government can be more open, there are a few areas ripe for change.

Perhaps the most glaring need for an open-source approach is in information sharing. Today, among and within several federal agencies, a culture of reflexive and unnecessary information withholding prevails. This knee-jerk secrecy can backfire with fatal consequences, as seen in the 1998 embassy bombings in Africa, the 9/11 attacks, and the Boston Marathon bombings. What’s most troubling is that decades after the dangers of information-sharing were identified, the problem persists.
What’s preventing reform? The answer starts with the government’s hierarchical structure—though an information-is-power mentality and “need to know” Cold War-era culture contribute too. To improve the practice of information sharing, government needs to change the structure of information sharing. Specifically, it needs to flatten the hierarchy.
Former Obama Administration regulation czar Cass Sunstein’s “nudge” approach shows how this could work. In his book Simpler: The Future of Government, he describes how making even small changes to an environment can affect significant changes in behavior. While Sunstein focuses on regulations, the broader lesson is clear: Change the environment to encourage better behavior and people tend to exhibit better behavior. Without such strict adherence to the many tiers of the hierarchy, those working within it could be nudged towards, rather than fight to, share information.
One example of where this worked is in with the State Department’s annual Religious Engagement Report (RER). In 2011, the office in charge of the RER decided that instead of having every embassy submit their data via email, they would post it on a secure wiki. On the surface, this was a decision to change an information-sharing procedure. But it also changed the information-sharing culture. Instead of sharing information only along the supervisor-subordinate axis, it created a norm of sharing laterally, among colleagues.
Another advantage to flattening information-sharing hierarchies is that it reduces the risk of creating “single points of failure,” to quote technology scholar Beth Noveck. The massive amounts of data now available to us may need massive amounts of eyeballs in order to spot patterns of problems—small pools of supervisors atop the hierarchy cannot be expected to shoulder those burdens alone. And while having the right tech tools to share information is part of the solution—as the wiki made it possible for the RER—it’s not enough. Leadership must also create a culture that nudges their staff to use these tools, even if that means relinquishing a degree of their own power.
Finally, a more open work culture would help connect interested parties across government to let them share the hard work of bringing new ideas to fruition. Government is filled with examples of interesting new projects that stall in their infancy. Creating a large pool of collaborators dedicated to a project increases the likelihood that when one torchbearer burns out, others in the agency will pick up for them.
When Linus Torvalds released Linux, it was considered, in Raymond’s words, “subversive” and “a distinct shock.” Could the federal government withstand such a shock?
Evidence suggests it can—and the transformation is already happening in small ways. One of the winners of the Harvard Kennedy School’s Innovations in Government award is State’s Consular Team India (CTI), which won for joining their embassy and four consular posts—each of which used to have its own distinct set of procedures-into a single, more effective unit who could deliver standardized services. As CTI describes it, “this is no top-down bureaucracy” but shares “a common base of information and shared responsibilities.” They flattened the hierarchy, and not only lived, but thrived.”

Residents remix their neighborhood’s streets through platform


Springwise: “City residents may not have degrees in urban planning, but their everyday use of high streets, parks and main roads means they have some valuable input into what’s best for their local environment. A new website called Streetmix is helping to empower citizens, enabling them to become architects with an easy-to-use street-building platform.
Developed by Code for America, the site greets users with a colorful cartoon representation of a typical street, split into segments of varying widths. Designers can then swap and change each piece into road, cycle paths, pedestrian areas, bus stops, bike racks and other amenities, as well as alter their dimensions. Users can create their own perfect high street or use the exact measurements of their own neighborhood to come up with new propositions for planned construction work. Indeed, Streetmix has already found use among residents and organizations to demonstrate how to better use the local space available. Kansas City’s Bike Walk KC has utilized the platform to show how new bike lanes could figure in an upcoming study of traffic flow in the region, while New Zealand’s Transport Blog has presented several alternatives to current street layouts in Auckland.
Streetmix is an easy-to-use visualization tool that can help amateurs present their ideas to local authorities in a more coherent way, potentially increasing the chances of politicians hearing calls for change. Are there other ways to help laymen express complex ideas more eloquently?”
Spotted by Murtaza Patel, written by Springwise