Stefaan Verhulst
New York University Law Review Paper by Elizabeth G. Porter and Kathryn A. Watts: “Federal rulemaking has traditionally been understood as a text-bound, technocratic process. However, as this Article is the first to uncover, rulemaking stakeholders — including agencies, the President and members of the public — are now deploying politically tinged visuals to push their agendas at every stage of high-stakes, often virulently controversial, rulemakings. Rarely do these visual contributions appear in the official rulemaking record, which remains defined by dense text, lengthy cost-benefit analyses, and expert reports. Perhaps as a result, scholars have overlooked the phenomenon we identify here: the emergence of a visual rulemaking universe that is splashing images, GIFs, and videos across social media channels. While this new universe, which we call “visual rulemaking,” might appear to be wholly distinct from the textual rulemaking universe on which administrative law has long focused, the two are not in fact distinct. Visual politics are seeping into the technocracy.
This Article argues that visual rulemaking is a good thing. It furthers fundamental regulatory values, including transparency and political accountability. It may also facilitate participation by more diverse stakeholders — not merely regulatory insiders who are well-equipped to navigate dense text. Yet we recognize that visual rulemaking poses risks. Visual appeals may undermine the expert-driven foundation of the regulatory state, and some uses may threaten or outright violate key legal doctrines, including the Administrative Procedure Act and longstanding prohibitions on agency lobbying and propaganda. Nonetheless, we conclude that administrative law theory and doctrine ultimately can and should welcome this robust new visual rulemaking culture….(More)”
New report by Jos Berens, Ulrich Mans and Stefaan Verhulst: “Recent years have witnessed something of a sea-change in the way humanitarian organizations consider and use data. Growing awareness of the potential of data has led to new enthusiasm and new, innovative applications that seek to respond to and mitigate crises in fresh ways. At the same time, it has become apparent that the potential benefits are accompanied by risks. A new framework is needed that can help balance the benefits and risks, and that can aid humanitarian organizations and others (e.g., policymakers) develop a more responsible approach to data collection and use in their efforts to combat natural and man-made crises around the world. …
The report we are releasing today, “Mapping and Comparing Responsible Data Approaches”, attempts to guide the first steps toward such a framework by learning from current approaches and principles. It is the outcome of a joint research project commissioned by UNOCHA and conducted in collaboration between the GovLab at NYU and Leiden University. In an effort to better understand the landscape, we have considered existing data use policies and principles from 17 organizations. These include 7 UN agencies, 7 International Organizations, 2 government agencies and 1 research institute. Our study of these organizations’ policies allowed us to extract a number of key takeaways that, together, amount to something like a roadmap for responsible data use for any humanitarian organization considering using data in new ways.
We began our research by closely mapping the existing responsible data use policies. To do this, we developed a template with eight broad themes that determines the key ingredients of responsible data framework. This use of a consistent template across organizations permits us to study and compare the 17 data use policies in a structured and systematic manner. Based on this template, we were able to extract 7 key takeaways for what works best when using data in a humanitarian context – presented in the conclusion to the paper being released today. They are designed to be broad enough to be broadly applicable, yet specific enough to be operational and actually usable….(More)”
Essay by John Gastil: “Dozens—and possibly hundreds—of online platforms have been built in the past decade to facilitate specific forms of civic engagement. Unconnected to each other, let alone an integrated system easy for citizens to use, these platforms cannot begin to realize their full potential. The author proposes a massive collaborative project to build an integrated platform called, tongue squarely in cheek, “The Democracy Machine.” The Machine draws on public energy and ideas, mixing those into concrete policy advice, influencing government decision making, and creating a feedback loop that helps officials and citizens track progress together as they continuously turn the policymaking crank. This online system could help to harmonize civic leaders, vocal and marginalized citizens, and government. Democracy’s need for ongoing public consultation would fuel the Machine, which would, in turn, generate the empowered deliberation and public legitimacy that government needs to make tough policy decisions….(More)”
However, the rapid increase in the quantity of personal information that is being collected and retained, combined with our increased ability to analyze and combine it with other information, is creating concerns about privacy. When information about people and their activities can be collected, analyzed, and repurposed in so many ways, it can create new opportunities for crime, discrimination, inadvertent disclosure, embarrassment, and harassment.
This Administration has been a strong champion of initiatives to improve the state of privacy, such as the “Consumer Privacy Bill of Rights” proposal and the creation of the Federal Privacy Council. Similarly, the White House report Big Data: Seizing Opportunities, Preserving Values highlights the need for large-scale privacy research, stating: “We should dramatically increase investment for research and development in privacy-enhancing technologies, encouraging cross-cutting research that involves not only computer science and mathematics, but also social science, communications and legal disciplines.”
Today, we are pleased to release the National Privacy Research Strategy. Research agencies across government participated in the development of the strategy, reviewing existing Federal research activities in privacy-enhancing technologies, soliciting inputs from the private sector, and identifying priorities for privacy research funded by the Federal Government. The National Privacy Research Strategy calls for research along a continuum of challenges, from how people understand privacy in different situations and how their privacy needs can be formally specified, to how these needs can be addressed, to how to mitigate and remediate the effects when privacy expectations are violated. This strategy proposes the following priorities for privacy research:
- Foster a multidisciplinary approach to privacy research and solutions;
- Understand and measure privacy desires and impacts;
- Develop system design methods that incorporate privacy desires, requirements, and controls;
- Increase transparency of data collection, sharing, use, and retention;
- Assure that information flows and use are consistent with privacy rules;
- Develop approaches for remediation and recovery; and
- Reduce privacy risks of analytical algorithms.
With this strategy, our goal is to produce knowledge and technology that will enable individuals, commercial entities, and the Federal Government to benefit from technological advancements and data use while proactively identifying and mitigating privacy risks. Following the release of this strategy, we are also launching a Federal Privacy R&D Interagency Working Group, which will lead the coordination of the Federal Government’s privacy research efforts. Among the group’s first public activities will be to host a workshop to discuss the strategic plan and explore directions of follow-on research. It is our hope that this strategy will also inspire parallel efforts in the private sector….(More)”
Nesta: “This paper by Mor Rubinstein (Open Knowledge International) and Josh Cowls and Corinne Cath (Oxford Internet Institute) explores the methods and motivations behind innovative uses of open government data in five specific country contexts – Chile, Argentine, Uruguay, Israel, and Denmark; and considers how the insights it uncovers might be adopted in a UK context.
Through a series of interviews with ‘social hackers’ and open data practitioners and experts in countries with recognised open government data ‘hubs’, the authors encountered a diverse range of practices and approaches in how actors in different sectors of society make innovative uses of open government data. This diversity also demonstrated how contextual factors shape the opportunities and challenges for impactful open government data use.
Based on insights from these international case studies, the paper offers a number of recommendations – around community engagement, data literacy and practices of opening data – which aim to support governments and citizens unlock greater knowledge exchange and social impact through open government data….(More)”
Book by Manuel Stagars: “This book explores the power of greater openness, accountability, and transparency in digital information and government data for the nations of Southeast Asia. The author demonstrates that, although the term “open data” seems to be self-explanatory, it involves an evolving ecosystem of complex domains. Through empirical case studies, this book explains how governments in the ASEAN may harvest the benefits of open data to maximize their productivity, efficiency and innovation. The book also investigates how increasing digital divides in the population, boundaries to civil society, and shortfalls in civil and political rights threaten to arrest open data in early development, which may hamper post-2015 development agendas in the region. With robust open data policies and clear roadmaps, member states of the ASEAN can harvest the promising opportunities of open data in their particular developmental, institutional and legal settings. Governments, policy makers, entrepreneurs and academics will gain a clearer understanding of the factors that enable open data from this timely research….(More)”
Alexander B. Howard and Patrice McDermott in Science: “Five decades after the United States first enacted the Freedom of Information Act (FOIA), Congress has voted to make the first major reforms to the statute since 2007. President Lyndon Johnson signed the first FOIA on 4 July 1966, enshrining in law the public’s right to access to information from executive branch government agencies. Scientists and others around the world can use the FOIA to learn what the U.S. government has done in its policies and practices. Proposed reforms should be a net benefit to public understanding of the scientific process and knowledge, by increasing the access of scientists to archival materials and reducing the likelihood of science and scientists being suppressed by official secrecy or bureaucracy.
Although the FOIA has been important for accountability, reform is sorely needed. An analysis of the 15 federal government agencies that received the most FOIA requests found poor to abysmal compliance rates (1, 2). In 2016, the Associated Press found that the Obama Administration had set a new record for unfulfilled FOIA requests (3). Although that has to be considered in the context of a rise in request volume without commensurate increases in resources to address them, researchers have found that most agencies simply ignore routine requests for travel schedules (4). An audit of 165 federal government agencies found that only 40% complied with the E-FOIA Act of 1996; just 67 of them had online libraries that were regularly updated with a substantial number of documents released under FOIA (5).
In the face of growing concerns about compliance, FOIA reform was one of the few recent instances of bicameral bipartisanship in Congress, with both the House and Senate each passing bills this spring with broad support. Now that Congress moved to send the Senate bill on to the president to sign into law, implementation of specific provisions will bear close scrutiny, including the potential impact of disclosure upon scientists who work in or with government agencies (6). Proposed revisions to the FOIA statute would improve how government discloses information to the public, while leaving intact exemptions for privacy, proprietary information, deliberative documents, and national security.
Features of Reforms
One of the major reforms in the House and Senate bills was to codify the “presumption of openness” outlined by President Obama the day after he took office in January 2009 when he declared that FOIA should be administered with a clear presumption: In the face of doubt, “openness” would prevail. This presumption of openness was affirmed by U.S. Attorney General Holder in March 2009. Although these declarations have had limited effect in the agencies (as described above), codifying these reforms into law is crucial not only to ensure that this remains executive branch policy after this president leaves office but also to provide requesters with legal force beyond an executive order….(More)”
Liesbet van Zoonen in Government Information Quarterly: “In this paper a framework is constructed to hypothesize if and how smart city technologies and urban big data produce privacy concerns among the people in these cities (as inhabitants, workers, visitors, and otherwise). The framework is built on the basis of two recurring dimensions in research about people’s concerns about privacy: one dimensions represents that people perceive particular data as more personal and sensitive than others, the other dimension represents that people’s privacy concerns differ according to the purpose for which data is collected, with the contrast between service and surveillance purposes most paramount. These two dimensions produce a 2 × 2 framework that hypothesizes which technologies and data-applications in smart cities are likely to raise people’s privacy concerns, distinguishing between raising hardly any concern (impersonal data, service purpose), to raising controversy (personal data, surveillance purpose). Specific examples from the city of Rotterdam are used to further explore and illustrate the academic and practical usefulness of the framework. It is argued that the general hypothesis of the framework offers clear directions for further empirical research and theory building about privacy concerns in smart cities, and that it provides a sensitizing instrument for local governments to identify the absence, presence, or emergence of privacy concerns among their citizens….(More)”
Gillian Tett in the Financial Times: “Last week, I decided to take a gaggle of kids for an end-of-school-year lunch in a New York neighbourhood that I did not know well. I duly began looking for a suitable restaurant. A decade ago, I would have done that by turning to a restaurant guide. In the world I grew up in, it was normal to seek advice from the “experts”.
But in Manhattan last week, it did not occur to me to consult Fodor’s. Instead, I typed what I needed into my cellphone, scrolled through a long list of online restaurant recommendations, including comments from people who had eaten in them — and picked one.
Yes, it was a leap of faith; those restaurant reviews might have been fake. But there were enough voices for me to feel able to trust the wisdom of the cyber crowds — and, as it happened, our lunch choice was very good.
This is a trivial example of a much bigger change that is under way, and one that has some thought-provoking implications in the wake of the Brexit vote. Before the referendum, British citizens were subjected to a blitz of advice about the potential costs of Brexit from “experts”: economists, central bankers, the International Monetary Fund and world leaders, among others. Indeed, the central strategy of the government (and other “Remainers”) appeared to revolve around wheeling out these experts, with their solemn speeches and statistics….
I suspect that it indicates something else: that citizens of the cyber world no longer have much faith in anything that experts say, not just in the political sphere but in numerous others too. At a time when we increasingly rely on crowd-sourced advice rather than official experts to choose a restaurant, healthcare and holidays, it seems strange to expect voters to listen to official experts when it comes to politics.
In our everyday lives, we are moving from a system based around vertical axes of trust, where we trust people who seem to have more authority than we do, to one predicated on horizontal axes of trust: we take advice from our peer group.
You can see this clearly if you look at the surveys conducted by groups such as the Pew Research Center. These show that faith in institutions such as the government, big business and the media has crumbled in recent years; indeed, almost the only institution in the US that has bucked the trend is the military.
What is even more interesting to look at, however, are the areas where trust remains high. In an annual survey conducted by the Edelman public relations firm, people in 20 countries are asked who they trust. They show rising confidence in the “a person like me” category, and surprisingly high trust in digital technology. We live in a world where we increasingly trust our Facebook friends and the Twitter crowd more than we do the IMF or the prime minister.
In some senses, this is good news. Relying on horizontal axes of trust should mean more democracy and empowerment for ordinary citizens. But the problem of this new world is that people can fall prey to social fads and tribalism — or groupthink…..
Either way, nobody is going to put this genie back into the bottle. So we all need to think about what creates the bonds of “trust” in today’s world. And recognise that the 20th-century model of politics, with its reverence for experts and fixed parties, may eventually seem as outdated as restaurant guides. We live in volatile time…(More)”
The White House: “Every year, more than 11 million people move through America’s 3,100 local jails, many on low-level, non-violent misdemeanors, costing local governments approximately $22 billion a year. In local jails, 64 percent of people suffer from mental illness, 68 percent have a substance abuse disorder, and 44 percent suffer from chronic health problems. Communities across the country have recognized that a relatively small number of these highly vulnerable people cycle repeatedly not just through local jails, but also hospital emergency rooms, shelters, and other public systems, receiving fragmented and uncoordinated care at great cost to American taxpayers, with poor outcomes.
For example, in Miami-Dade, Florida found that 97 people with serious mental illness accounted for $13.7 million in services over four years, spending more than 39,000 days in either jail, emergency rooms, state hospitals or psychiatric facilities in their county. In response, the county provided key mental health de-escalation training to their police officers and 911 dispatchers and, over the past five years, Miami-Dade police have responded to nearly 50,000 calls for service for people in mental health crisis, but have made only 109 arrests, diverting more than 10,000 people to services or safely stabilizing situations without arrest. The jail population fell from over 7000 to just over 4700 and the county was able to close an entire jail facility, saving nearly $12 million a year.
In addition, on any given day, more than 450,000 people are held in jail before trial, nearly 63 percent of the local jail population, even though they have not been convicted of a crime. A 2014 study of New York’s Riker’s Island jail found more than 86% percent of detained individuals were held on a bond of $500 or less. To tackle the challenges of bail, in 2014 Charlotte-Mecklenburg, NC began using a data-based risk assessment tool to identify low risk people in jail and find ways to release them safely. Since they began using the tool, the jail population has gone down 20 percent, significantly more low-risk individuals have been released from jail, and there has been no increase in reported crime.
To break this cycle of incarceration, the Administration has launched the Data-Driven Justice Initiative with a bipartisan coalition of city, county, and state governments who have committed to using data-driven strategies to divert low-level offenders with mental illness out of the criminal system and to change approaches to pre-trial incarceration so that low risk offenders no longer stay in jail simply because they cannot afford a bond. These innovative strategies, which have measurably reduced jail populations in several communities, help stabilize individuals and families, better serve communities, and, often, saves money in the process. DDJ communities commit to:
- combining data from across criminal justice and health systems to identify the individuals with the highest number of contacts with police, ambulance, emergency departments, and other services, and, leverage existing resources to link them to health, behavioral health, and social services in the community;
- equipping law enforcement and first responders to enable more rapid deployment of tools, approaches, and other innovations they need to safely and more effectively respond to people in mental health crisis and divert people with high needs to identified service providers instead of arrest; and
- working towards using objective, data-driven, validated risk assessment tools to inform the safe release of low-risk defendants from jails in order to reduce the jail population held pretrial….(More: FactSheet)”