Information for the People: Tunisia Embraces Open Government, 2011–2016


Case study by Tristan Dreisback at Innovations for Successful Societies: “In January 2011, mass demonstrations in Tunisia ousted a regime that had tolerated little popular participation, opening the door to a new era of transparency. The protesters demanded an end to the secrecy that had protected elite privilege. Five months later, the president issued a decree that increased citizen access to government data and formed a steering committee to guide changes in information practices, building on small projects already in development. Advocates in the legislature and the public service joined with civil society leaders to support a strong access-to-information policy, to change the culture of public administration, and to secure the necessary financial and technical resources to publish large quantities of data online in user-friendly formats. Several government agencies launched their own open-data websites. External pressure, coupled with growing interest from civil society and legislators, helped keep transparency reforms on the cabinet office agenda despite frequent changes in top leadership. In 2016, Tunisia adopted one of the world’s strongest laws regarding access to information. Although members of the public did not put all of the resources to use immediately, the country moved much closer to having the data needed to improve access to services, enhance government performance, and support the evidence-based deliberation on which a healthy democracy depended…(More)”

ControCurator: Understanding Controversy Using Collective Intelligence


Paper by Benjamin Timmermans et al: “There are many issues in the world that people do not agree on, such as Global Warming [Cook et al. 2013], Anti-Vaccination [Kata 2010] and Gun Control [Spitzer 2015]. Having opposing opinions on such topics can lead to heated discussions, making them appear controversial. Such opinions are often expressed through news articles and social media. There are increasing calls for methods to detect and monitor these online discussions on different topics. Existing methods focus on using sentiment analysis and Wikipedia for identifying controversy [Dori-Hacohen and Allan 2015]. The problem with this is that it relies on a well structured and existing debate, which may not always be the case. Take for instance news reporting during large disasters, in which case the structure of a discussion is not yet clear and may change rapidly. Adding to this is that there is currently no agreed upon definition as to what exactly defines controversy. It is only agreed that controversy arises when there is a large debate by people with opposing viewpoints, but we do not yet understand which are the characteristic aspects and how they can be measured. In this paper we use the collective intelligence of the crowd in order to gain a better understanding of controversy by evaluating the aspects that have impact on it….(More)”

See also http://crowdtruth.org/

 

Citizenship office wants ‘Emma’ to help you


 at FedScoop: “U.S. Citizenship and Immigration Services unveiled a new virtual assistant live-chat service, known as “Emma,” to assist customers and website visitors in finding information and answering questions in a timely and efficient fashion.

The agency told FedScoop that it built the chatbot with the help of Verizon and artificial intelligence interface company Next IT. The goal  is “to address the growing need for customers to obtain information quicker and through multiple access points, USCIS broadened the traditional call center business model to include web-based self-help tools,” the agency says.

USCIS, a component agency of the Department of Homeland Security, says it receives nearly 14 million calls relating to immigration every year. The virtual assistant and live-chat services are aimed at becoming the first line of help available to users of USCIS.gov who might have trouble finding answers by themselves.

The bot greets customers when they enter the website, answers basic questions via live chat and supplies additional information in both English and Spanish. As a result, the amount of time customers spend searching for information on the website is greatly reduced, according to USCIS. Because the virtual assistant is embedded within the website, it can rapidly provide relevant information that may have been difficult to access manually.

The nature of the bot lends itself to potential encounters with personally identifiable information (PII), or PII, of the customers it interacts with. Because of this, USCIS recently conducted a privacy impact assessment (PIA).

Much of the assessment revolved around accuracy and the security of information that Emma could potentially encounter in a customer interaction. For the most part, the chat bot doesn’t require customers to submit personal information. Instead, it draws its responses from content already available on USCIS.gov, relative to the amount of information that users choose to provide. Answers are, according to the PIA, verified by thorough and frequent examination of all content posted to the site.

According to USCIS, the Emma will delete all chat logs — and therefore all PII — immediately after the customer ends the chat session. Should a customer reach a question that it can’t answer effectively and choose to continue the session with an agent in a live chat, the bot will ask for the preferred language (English or Spanish), the general topic of conversation, short comments on why the customer wishes to speak with a live agent, and the case on file and receipt number.

This information would then be transferred to the live agent. All other sensitive information entered, such as Social Security numbers or receipt numbers, would then be automatically masked in the subsequent transfer to the live agent…(More)”

How can we study disguised propaganda on social media? Some methodological reflections


Jannick Schou and Johan Farkas at DataDrivenJournalism: ’Fake news’ has recently become a seemingly ubiquitous concept among journalists, researchers, and citizens alike. With the rise of platforms such as Facebook and Twitter, it has become possible to spread deliberate forms of misinformation in hitherto unforeseen ways. This has also spilled over into the political domain, where new forms of (disguised) propaganda and false information have recently begun to emerge. These new forms of propaganda have very real effects: they serve to obstruct political decision-making processes, instil false narratives within the general public, and add fuel to already heated sites of political conflict. They represent a genuine democratic problem.

Yet, so far, both critical researchers and journalists have faced a number of issues and challenges when attempting to understand these new forms of political propaganda. Simply put: when it comes to disguised propaganda and social media, we know very little about the actual mechanisms through which such content is produced, disseminated, and negotiated. One of the key explanations for this might be that fake profiles and disguised political agendas are incredibly difficult to study. They present a serious methodological challenge. This is not only due to their highly ephemeral nature, with Facebook pages being able to vanish after only a few days or hours, but also because of the anonymity of its producers. Often, we simply do not know who is disseminating what and with what purpose. This makes it difficult for us to understand and research exactly what is going on.

This post takes its point of departure from a new article published in the international academic journal New Media & Society. Based on the research done for this article, we want to offer some methodological reflections as to how disguised propaganda might be investigated. How can we research fake and disguised political agendas? And what methodological tools do we have at our disposal?…

two main methodological advices spring to mind. First of all: collect as much data as you can in as many ways as possible. Make screenshots, take detailed written observations, use data scraping, and (if possible) participate in citizen groups. One of the most valuable resources we had at our disposal was the set of heterogeneous data we collected from each page. Using this allowed us to carefully dissect and retrace the complex set of practices involved in each page long after they were gone. While we certainly tried to be as systematic in our data collection as possible, we also had to use every tool at our disposal. And we had to constantly be on our toes. As soon as a page emerged, we were there: ready to write down notes and collect data.

Second: be willing to participate and collaborate. Our research showcases the immense potential in researchers (and journalists) actively collaborating with citizen groups and grassroots movements. Using the collective insights and attention of this group allowed us to quickly find and track down pages. It gave us renewed methodological strength. Collaborating across otherwise closed boundaries between research and journalism opens up new avenues for deeper and more detailed insights….(More)”

A framework for analyzing digital volunteer contributions in emergent crisis response efforts


 and  in New Media and Society: “Advances in information, communication, and computational technologies allow digital volunteer networks formed by concerned publics across the globe to contribute to an effective response to disasters and crises. Digital volunteer networks are event-centric and emergent networks. Currently, the literature is sharply growing in the fields of communication, computer science, emergency management, and geography. This article aims to assess the current status of the literature and suggest a comprehensive conceptual framework of digital volunteer networks in response to disasters and crises. This framework is based on a traditional input–process–output model consisting of three dimensions: the disaster and crisis context, a voluntary response process, and outputs and outcomes. We also discuss challenges of digital volunteer networks for crisis response. This article is expected to contribute to the development of related theories and hypotheses and practical strategies for managing digital volunteer networks…(More)”,

Future Libraries


ARUP: “Libraries are going through a renaissance, both in terms of the social infrastructure they provide and in terms of a diversification of the services and experiences offered. In corporate environments they are playing an increasingly important role in the provision of collaborate workspace and innovation. In communities they are evolving into hubs for education, health, entertainment and work….

This report brings to light significant trends that will influence the future of public, academic and corporate libraries and outlines the implications on their design, operation and user experience. It is the result of a collective exploration through series of workshop events held in London, Melbourne, San Francisco and Sydney, attended by experts in the design and management of libraries. This piece of research presents a glimpse into the future. It explores what we may expect to see as the physical and the digital arena continues to evolve and aims to serve as a foundation for further discussion around the future role of libraries in the communities they serve….(More)”

Big Data: A New Empiricism and its Epistemic and Socio-Political Consequences


Chapter by Gernot Rieder and Judith Simon in by Berechenbarkeit der Welt? Philosophie und Wissenschaft im Zeitalter von Big Data: “…paper investigates the rise of Big Data in contemporary society. It examines the most prominent epistemological claims made by Big Data proponents, calls attention to the potential socio-political consequences of blind data trust, and proposes a possible way forward. The paper’s main focus is on the interplay between an emerging new empiricism and an increasingly opaque algorithmic environment that challenges democratic demands for transparency and accountability. It concludes that a responsible culture of quantification requires epistemic vigilance as well as a greater awareness of the potential dangers and pitfalls of an ever more data-driven society….(More)”.

Advocacy and Policy Change Evaluation: Theory and Practice


Book by Annette Gardner and Claire Brindis: “This is the first book-length treatment of the concepts, designs, methods, and tools needed to conduct effective advocacy and policy change evaluations. By integrating insights from different disciplines, Part I provides a conceptual foundation for navigating advocacy tactics within today’s turbulent policy landscape. Part II offers recommendations for developing appropriate evaluation designs and working with unique advocacy and policy change–oriented instruments. Part III turns toward opportunities and challenges in this growing field. In addition to describing actual designs and measures, the chapters includes suggestions for addressing the specific challenges of working in a policy setting, such as a long time horizon for achieving meaningful change.

To illuminate and advance this area of evaluation practice, the authors draw on over 30 years of evaluation experience; collective wisdom based on a new, large-scale survey of evaluators in the field; and in-depth case studies on diverse issues—from the environment, to public health, to human rights. Ideal for evaluators, change makers, and funders, this book is the definitive guide to advocacy and policy change evaluation….(More)”.

Microtasking: Redefining crowdsourcing practices in emergency management


Paper by Poblet, Marta; Fitzpatrick, Mari and Chhetri, Prem: “examines the roles, types and forms of virtual microtasking for emergency information management in order to better understand collective intelligence mechanisms and the potential for logistics response. Using three case studies this paper reviews the emerging body of knowledge in microtasking practices in emergency management to demonstrate how crowd-sourced information is captured and processed during emergency events to provide critical intelligence throughout the emergency cycle. It also considers the impact of virtual information collection, collation and management on traditional humanitarian operations and relief efforts.

Based on the case studies the emergent forms of microtasking for emergency information management were identified. Opportunities for continuities, adaptations and innovations are explained. The contribution of virtual microtasking extends to all supply chain strategic domains to help maximise resource use and optimise service delivery response….(More)”

The Transparency Fix: Secrets, Leaks, and Uncontrollable Government Information


Book by Mark Fenster:  Is the government too secret or not secret enough? Why is there simultaneously too much government secrecy and a seemingly endless procession of government leaks? The Transparency Fix asserts that we incorrectly assume that government information can be controlled. The same impulse that drives transparency movements also drives secrecy advocates. They all hold the mistaken belief that government information can either be released or kept secure on command.

The Transparency Fix argues for a reformation in our assumptions about secrecy and transparency. The world did not end because Julian Assange, WikiLeaks, and Edward Snowden released classified information. But nor was there a significant political change. “Transparency” has become a buzzword, while secrecy is anathema. Using a variety of real-life examples to examine how government information actually flows, Mark Fenster describes how the legal regime’s tenuous control over state information belies both the promise and peril of transparency. He challenges us to confront the implausibility of controlling government information and shows us how the contemporary obsession surrounding transparency and secrecy cannot radically change a state that is defined by so much more than information….(More)”.