Crowdsourcing America’s cybersecurity is an idea so crazy it might just work


at the Washington Post: “One idea that’s starting to bubble up from Silicon Valley is the concept of crowdsourcing cybersecurity. As Silicon Valley venture capitalist Robert R. Ackerman, Jr. has pointed out, due to “the interconnectedness of our society in cyberspace,” cyber networks are best viewed as an asset that we all have a shared responsibility to protect. Push on that concept hard enough and you can see how many of the core ideas from Silicon Valley – crowdsourcing, open source software, social networking, and the creative commons – can all be applied to cybersecurity.

Silicon Valley venture capitalists are already starting to fund companies that describe themselves as crowdsourcing cybersecurity. For example, take Synack, a “crowd security intelligence” company that received $7.5 million in funding from Kleiner Perkins (one of Silicon Valley’s heavyweight venture capital firms), Allegis Ventures, and Google Ventures in 2014. Synack’s two founders are ex-NSA employees, and they are using that experience to inform an entirely new type of business model. Synack recruits and vets a global network of “white hat hackers,” and then offers their services to companies worried about their cyber networks. For a fee, these hackers are able to find and repair any security risks.

So how would crowdsourced national cybersecurity work in practice?

For one, there would be free and transparent sharing of computer code used to detect cyber threats between the government and private sector. In December, the U.S. Army Research Lab added a bit of free source code, a “network forensic analysis network” known as Dshell, to the mega-popular code sharing site GitHub. Already, there have been 100 downloads and more than 2,000 unique visitors. The goal, says William Glodek of the U.S. Army Research Laboratory, is for this shared code to “help facilitate the transition of knowledge and understanding to our partners in academia and industry who face the same problems.”

This open sourcing of cyber defense would be enhanced with a scaled-up program of recruiting “white hat hackers” to become officially part of the government’s cybersecurity efforts. Popular annual events such as the DEF CON hacking conference could be used to recruit talented cyber sleuths to work alongside the government.

There have already been examples of communities where people facing a common cyber threat gather together to share intelligence. Perhaps the best-known example is the Conficker Working Group, a security coalition that was formed in late 2008 to share intelligence about malicious Conficker malware. Another example is the Financial Services Information Sharing and Analysis Center, which was created by presidential mandate in 1998 to share intelligence about cyber threats to the nation’s financial system.

Of course, there are some drawbacks to this crowdsourcing idea. For one, such a collaborative approach to cybersecurity might open the door to government cyber defenses being infiltrated by the enemy. Ackerman makes the point that you never really know who’s contributing to any community. Even on a site such as Github, it’s theoretically possible that an ISIS hacker or someone like Edward Snowden could download the code, reverse engineer it, and then use it to insert “Trojan Horses” intended for military targets into the code….  (More)

If Data Sharing is the Answer, What is the Question?


Christine L. Borgman at ERCIM News: “Data sharing has become policy enforced by governments, funding agencies, journals, and other stakeholders. Arguments in favor include leveraging investments in research, reducing the need to collect new data, addressing new research questions by reusing or combining extant data, and reproducing research, which would lead to greater accountability, transparency, and less fraud. Arguments against data sharing rarely are expressed in public fora, so popular is the idea. Much of the scholarship on data practices attempts to understand the socio-technical barriers to sharing, with goals to design infrastructures, policies, and cultural interventions that will overcome these barriers.
However, data sharing and reuse are common practice in only a few fields. Astronomy and genomics in the sciences, survey research in the social sciences, and archaeology in the humanities are the typical exemplars, which remain the exceptions rather than the rule. The lack of success of data sharing policies, despite accelerating enforcement over the last decade, indicates the need not just for a much deeper understanding of the roles of data in contemporary science but also for developing new models of scientific practice. Science progressed for centuries without data sharing policies. Why is data sharing deemed so important to scientific progress now? How might scientific practice be different if these policies were in place several generations ago?
Enthusiasm for “big data” and for data sharing are obscuring the complexity of data in scholarship and the challenges for stewardship. Data practices are local, varying from field to field, individual to individual, and country to country. Studying data is a means to observe how rapidly the landscape of scholarly work in the sciences, social sciences, and the humanities is changing. Inside the black box of data is a plethora of research, technology, and policy issues. Data are best understood as representations of observations, objects, or other entities used as evidence of phenomena for the purposes of research or scholarship. Rarely do they stand alone, separable from software, protocols, lab and field conditions, and other context. The lack of agreement on what constitutes data underlies the difficulties in sharing, releasing, or reusing research data.
Concerns for data sharing and open access raise broader questions about what data to keep, what to share, when, how, and with whom. Open data is sometimes viewed simply as releasing data without payment of fees. In research contexts, open data may pose complex issues of licensing, ownership, responsibility, standards, interoperability, and legal harmonization. To scholars, data can be assets, liabilities, or both. Data have utilitarian value as evidence, but they also serve social and symbolic purposes for control, barter, credit, and prestige. Incentives for scientific advancement often run counter to those for sharing data.
….
Rather than assume that data sharing is almost always a “good thing” and that doing so will promote the progress of science, more critical questions should be asked: What are the data? What is the utility of sharing or releasing data, and to whom? Who invests the resources in releasing those data and in making them useful to others? When, how, why, and how often are those data reused? Who benefits from what kinds of data transfer, when, and how? What resources must potential re-users invest in discovering, interpreting, processing, and analyzing data to make them reusable? Which data are most important to release, when, by what criteria, to whom, and why? What investments must be made in knowledge infrastructures, including people, institutions, technologies, and repositories, to sustain access to data that are released? Who will make those investments, and for whose benefit?
Only when these questions are addressed by scientists, scholars, data professionals, librarians, archivists, funding agencies, repositories, publishers, policy makers, and other stakeholders in research will satisfactory answers arise to the problems of data sharing…(More)”.

Reclaiming Accountability: Transparency, Executive Power, and the U.S. Constitution


New book by Heidi Kitrosser: “Americans tend to believe in government that is transparent and accountable. Those who govern us work for us, and therefore they must also answer to us. But how do we reconcile calls for greater accountability with the competing need for secrecy, especially in matters of national security? Those two imperatives are usually taken to be antithetical, but Heidi Kitrosser argues convincingly that this is not the case—and that our concern ought to lie not with secrecy, but with the sort of unchecked secrecy that can result from “presidentialism,” or constitutional arguments for broad executive control of information.
In Reclaiming Accountability, Kitrosser traces presidentialism from its start as part of a decades-old legal movement through its appearance during the Bush and Obama administrations, demonstrating its effects on secrecy throughout. Taking readers through the key presidentialist arguments—including “supremacy” and “unitary executive theory”—she explains how these arguments misread the Constitution in a way that is profoundly at odds with democratic principles. Kitrosser’s own reading offers a powerful corrective, showing how the Constitution provides myriad tools, including the power of Congress and the courts to enforce checks on presidential power, through which we could reclaim government accountability….(More)”

The crowd-sourcing web project bringing amateur and professional archaeologists together


Sarah Jackson at Culture 24: “With only limited funds and time, professional archaeologists consistently struggle to protect and interpret the UK’s vast array of archaeological finds and resources. Yet there are huge pools of amateur groups and volunteers who are not only passionate but also skilled and knowledgeable about archaeology in the UK.
Now a new web platform called MicroPasts has been produced in a collaboration between University College London (UCL) and the British Museum to connect institutions and volunteers so that they can create, fund and work on archaeological projects together.
Work by UCL postdoctoral researchers Chiara Bonacchi and Adi Keinan-Schoonbaert and British Museum post doc researcher Jennifer Wexler established much of the ground work including the design, implementation and the public engagement aspects of the of the new site.
According to one of the project leaders, Professor Andrew Bevan at UCL, MicroPasts emerged from a desire to harness the expertise (and manpower) of volunteers and to “pull together crowd-sourcing and crowd-funding in a way that hadn’t been tried before”.
Although there are many crowd-sourcing portals online, they are either specific to one project (DigVentures, for example, which conducted the world’s first crowd-funded dig in 2012) or can be used to create almost anything you can imagine (such as Kickstarter).
MicroPasts was also inspired by Crowdcrafting, which encourages citizen science projects and, like MicroPasts, offers a platform for volunteers and researchers with an interest in a particular subject to come together to create and contribute to projects….(More)”

The Utopia of Rules


New book by David Graeber on “Technology, Stupidity, and the Secret Joys of Bureaucracy”: “Where does the desire for endless rules, regulations, and bureaucracy come from? How did we come to spend so much of our time filling out forms? And just how much are our lives being ruined by all this nonstop documentation?

To answer these questions, anthropologist David Graeber—one of our most important and provocative thinkers—traces the peculiar and unexpected ways we relate to bureaucracy today and reveals how it shapes our lives in ways we may not even notice. Is the inane, annoying paperwork we confront every day really a cipher for state violence? And is the capitalist promise of salvation-through-technology just a tool for the powerful to exert more control? Graeber provides a forceful, radical answer to these questions, though he also suggests that there may be something perversely appealing—even romantic—about bureaucracy.
Leaping from the ascendance of right-wing economics in the second half of the twentieth century to the hidden meanings behind James Bond, Sherlock Holmes, and Batman, The Utopia of Rules is at once a powerful work of social theory in the tradition of Foucault and Marx, and an entertaining reckoning with popular culture that calls to mind Slavoj Žižek at his most accessible.
An essential book for our times, The Utopia of Rules is sure to start a million conversations about the institutions that rule our lives—and the better, freer world we should, perhaps, begin to imagine for ourselves….(More)”

The Metrics Myth


Jed Emerson at BlendedValue: “…Simply because our present, dominant approaches to assessing metrics fall short of our task—How can one measure the full value of a life saved or possible future changed? What, ultimately, is the real impact and value created through the allocation of our capital?—we persist because we know two things:
First, we know we are on a Hero’s Journey of inquiry and innovation. Too often we forget the present system of tracking financial performance (the basis upon which trillions of dollars flow through global capital markets and the foundation upon which too many of us build our lives) is the outcome of over sixty years of development, refinement and debate. In the U.S., GAAP and FASB (the fundamental building blocks of mainstream business and finance) were not created until after World War II; and it was not until the creation of the Environmental Protection Agency in 1970 that business and many nonprofits began tracking and assessing environmental metrics on a consistent basis. And while social metrics have always been a part of the parlance of government and philanthropic funding, many foundations and social investors have not sought to weave performance assessment into their process of allocating funds until recent decades. It is for these reasons I am quite comfortable with the reality that those creating the metrics and evaluation frameworks of tomorrow will need another twenty years to build what is not yet ours, for I know it will come in good time.
Second, we are creating Total Portfolio Reporting frameworks to track the returns of unified investing strategies (capable of reflecting the aggregate performance of philanthropic, social and environmental value creation) because we know it can be done—and indeed, we see the metrics mist clearing by the year.
As initiatives such as

The Principles for Responsible Investing’s Integrated Reporting work,

the recently re-organized SROI Network,

the Sustainable Accounting Standards Board,

B-Lab’s B-Analytics framework,

CapRock’s iPar system,

the ANDE Metrics Working Group

and a variety of grassroots initiatives coming together around various sets of common reporting for assessing community impact,we find one can create a balance between our aspirations for a better world and the challenges of demarcating our progress toward that goal.
In the end, I hate the whole metrics debate.
It is repetitive, mind numbing and distracting from the critical task of fighting the forces presently destroying our societies and planet. Each time some ignorant (not stupid, mind you, and yet, not fully aware of what they do not know; they are quite rightly, ignorant) newcomer enters the discussion, we’re all expected to re-group and re-define concepts and issues well documented and explored in the past. The continual, mindless reminders that not everything that counts can be counted leave me frustrated and even angry at some who for reasons beyond me don’t seem to understand that such now trite insights were the very starting place of this journey well more than 25 years ago and that, indeed, as newcomers they are as far behind the current exploration as we are from our goal.
Yet, we make progress despite our doubts and complications.
We advance the practice of both impact investing and performance measurement one step forward and two steps back as the current “knowledge” of the crowd actually pulls us backward to previous thinking and practice. And we know the appropriate application of metrics bring meaning and insight just as they demonstrate the limitations of such efforts….(More)”

Measuring government impact in a social media world


Arthur Mickoleit & Ryan Androsoff at OECD Insights: “There is hardly a government around the world that has not yet felt the impact of social media on how it communicates and engages with citizens. And while the most prominent early adopters in the public sector have tended to be politicians (think of US President Barack Obama’s impressive use of social media during his 2008 campaign), government offices are also increasingly jumping on the bandwagon. Yes, we are talking about those – mostly bricks-and-mortar – institutions that often toil away from the public gaze, managing the public administration in our countries. As the world changes, they too are increasingly engaging in a very public way through social media.
Research from our recent OECD working paper “Social Media Use by Governments” shows that as of November 2014, out of 34 OECD countries, 28 have a Twitter account for the office representing the top executive institution (head of state, head of government, or government as a whole), and 21 have a Facebook account….
 
But what is the impact governments can or should expect from social media? Is it all just vanity and peer pressure? Surely not.
Take the Spanish national police force (e.g. on Twitter, Facebook & YouTube), a great example of using social media to build long-term engagement, trust and a better public service. The thing so many governments yearn for, in this case the Spanish police seem to have managed well.
Or take the Danish “tax daddy” on Twitter – @Skattefar. It started out as the national tax administration’s quest to make it easier for everyone to submit correct tax filings; it is now one of the best examples around of a tax agency gone social.
Government administrations can use social media for internal purposes too. The Government of Canada used public platforms like Twitter and internal platforms like GCpedia and GCconnex to conduct a major employee engagement exercise (Blueprint 2020) to develop a vision for the future of the Canadian federal public service.
And when it comes to raising efficiency in the public sector, read this account of a Dutch research facility’s Director who decided to stop email. Not reduce it, but stop it altogether and replace it with social media.
There are so many other examples that could be cited. But the major question is how can we even begin to appraise the impact of these different initiatives? Because as we’ve known since the 19th century, “if you cannot measure it, you cannot improve it” (quote usually attributed to Lord Kelvin). Some aspects of impact measurement for social media can be borrowed from the private sector with regards to presence, popularity, penetration, and perception. But it’s around purpose that impact measurement agendas will split between the private sector and government. Virtually all companies will want to calculate the return on social media investments based on whether it helps them improve their financial returns. That’s different in the public sector where purpose is rarely defined in commercial terms.
A good impact assessment for social media in the public sector therefore needs to be built around its unique purpose-orientation. This is much more difficult to measure and it will involve a mix of quantitative data (e.g. reach of target audience) and qualitative data (e.g. case studies describing tangible impact). Social Media Use by Governments proposes a framework to start looking at social media measurement in gradual steps – from measuring presence, to popularity, to penetration, to perception, and finally, to purpose-orientation. The aim of this framework is to help governments develop truly relevant metrics and start treating social media activity by governments with the same public management rigour that is applied to other government activities. You can see a table summarising the framework by clicking on the thumbnail below.
This is far from an exact science, but we are beginning the work collaborating with member and partner governments to develop a toolkit that will help decision-makers implement the OECD Recommendation on Digital Government Strategies, including on the issue of social media metrics…(More)”.

Where is Our Polis In the 21st Century?


Hollie Russon Gilman: “If you could improve the relationship between citizens and the state, how would you do it? It’s likely that your answer would be different from mine and still different from the next five people I ask. Because rules and structures of government are constantly changing and the tools people use to communicate shift with newly available technologies, this relationship must continue to evolve…

Multiple factors shape the quality of democracy, such as the safety of free speech and reliability of public transit or secure long-term planning. Democracy, at least the glorified ancient ideal some like to lay claim to as our founding heritage, also involves the creation of a polis — specifically, a place where man is freed from the burdens of household goods, most famously articulated by Plato in The Republic.

We can’t mistake an ideal for the reality — Plato’s polis was highly constrained and available only to the most privileged of Greek men within a social system that also sanctioned slavery. However, the ideal of the polis  — a place to experience democratic virtues — also holds at least theoretical promise and compelling possibilities for real change to the current state of American democracy.

We need what this ideal has to offer, because the social contract as we know it today can feel more like a series of alienating, disconnected obligations than what it could and should be: an enabler of civic creativity or power. Our current social contract does not come with a polis — or, to put it another way, room to imagine new ways of thinking.

Why is this a problem? Because in order to truly harness civic innovation, we need to embrace deeper ways of thinking about democracy.

What would a deeper democracy look like? Harvard political theorist Robert Unger describes “deepened democracy” in his recent book Democracy Realized: The Progressive Alternative as a system in which citizens “must be able to see themselves and one another as individuals capable of escaping their confined roles.” One promising way citizens can perform new roles in a “deeper democracy” is by working with public institutions, and amongst themselves, to influence policymaking.

We need tools to empower these citizens use their work to fashion a polis for the 21st century. One particularly promising innovation is Participatory Budgeting (often shortened to “PB”), which is a process whereby citizens make spending decisions on a defined public budget and operate as active participants in public decision-making like allocating local funds in their neighborhood. The Brazilian Workers’ Party first attempted PB in 1989, where its success led to the World Bank calling PB a “best practice” in democratic innovation….

Why is PB so effective as a civic engagement tool? PB is especially powerful because it engages citizens with complex political issues on the local level, where they live. PB’s strength as an intervention in our social contract lies in municipal budgets as the scale at which citizens can be experts. In other words, people who live day to day in communities know best what resources those communities need to solve problems, be successful, and thrive.

Many of our governance decisions face the dual challenges of integrating individual-level participation efforts with the scale of contemporary national U.S. politics. Part of PB’s power may be breaking down complex decisions into their manageable parts. This strategy could be applied beyond budgets to a range of decision-making such as climate adaption or addressing food deserts.

PB represents one of the best tools in a broader toolkit designed to re-engage citizens in governance, but it’s far from the only one. Look around your very block, community, and city. Examples of places that could operate as a 21st-century polis range from traditional community anchor institutions engaging in new ways to the application of digital tools for civic ends. Citizenvestor is a civic crowd-funding site that works online and with traditional brick-and-mortar organizations. In Mount Rainer, MD, Community Forklift — a “nonprofit reuse center for home improvement supplies” (or, you might say, a library for tools) — and a local bike share engage a large group of residents.

Civic and social innovation is built from the exchange of resources between government institutions and community networks. Ideally, through coming together to talk, debate, and engage in the public sphere, people can flex their civic muscle and transform their lives. The fabric of communities is woven with the threads of deeply engaged and dedicated residents. A challenge of our current moment in history is to reconcile these passions with the mechanisms, and sometimes the technologies, necessary to improve public life.

Can this all add up to a wholesale civic revolution? Time will tell. At a minimum, it suggests the potential of community networks (analog and digital) to be leveraged for a stronger, more resilience and responsive 21st century polis….(More)”

The Future of Public Space Analytics


The AGILE Landscape Project: “Public space is an essential component of any great city. It brings people together to socialize, recreate, and work. More pointedly, it attracts people to the city, builds relationships, and spurs innovation and new ideas that fuel a city’s economic growth. How we optimize the investments made in our streetscapes, plazas, parks, and greenways is important to each individual project’s success and the city as a whole. Are the places being built fulfilling their promise? If so, could they be doing more and if not, how do they need to change?…
“Places in the Making”, a publication published in 2013 by MIT’s Department of Urban Studies and Planning, underscored this problem. They cited the lack of active metrics as a common challenge to placemaking efforts. They reported that “It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” They went on to explain the existing placemaking culture focuses on fuzzy, unmeasurable goals and that this has created “an inertia in assessment efforts”. The report stressed that this lack of assessment is detrimental to the field as a whole because valuable insights are left undiscovered and the same mistakes are made repeatedly….
“Places in the Making”, a publication published in 2013 by MIT’s Department of Urban Studies and Planning, underscored this problem. They cited the lack of active metrics as a common challenge to placemaking efforts. They reported that “It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” They went on to explain the existing placemaking culture focuses on fuzzy, unmeasurable goals and that this has created “an inertia in assessment efforts”. The report stressed that this lack of assessment is detrimental to the field as a whole because valuable insights are left undiscovered and the same mistakes are made repeatedly.

“It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” – “Places in the Making”, DUSP (2013)

Both William H. Whyte’s seminal Street Life Project and resulting book and documentary, Social Life of Small Urban Spaces (1980), along with Jan Gehl’s extensive public life studies and influential book Life Between Buildings (1987) serve as the foundation of the urban design profession’s knowledge base. Their observations and insights were amassed through considerable time and manpower. We are lucky to have their contribution from which to draw upon when making future decisions. Though we shouldn’t stop observing and building upon their significant contributions. We must continually add clarity and resolution particularly in the unique and specific context of each place.
social media metrics
Unfortunately, the POE and previous research are only snapshots in time. With the rise of embedded sensor infrastructure, “smart cities”, social networks, and ubiquity of smartphones that are becoming increasingly more powerful and sophisticated every year, the ability to gain real-time continuous 24/7 quantitative monitoring of public life data is now a reality. The ability to capture so much data is unprecedented and will only continue to get easier. The amount of and type of information that can and often is being collected about the world around us and ourselves is vast. When it comes to cities, transportation, parking, energy, waste collection, and water systems has attracted the most attention. In the private sector, retail analytics have been the focus.
Collecting data for data sake should not be a goal. Rather we need be knowledgeable of what is available and possible while being discerning about the data that is useful for improving the management and design of public space. By selectively incorporating sensors into the design of physical space, the data collection system can provide real-time information about the use of the space. Like website analytic systems, the network can provide useful information about the number of people using the space, how long they are using it, as well well as other meaningful measures of public life. The data gathered can be combined with a wide array of environmental data such as temperature, humidity, sun/shade patterns, rain fall, water quality, and air quality. In addition, other available data sets such as sales tax collections, building permits, real estate transactions, and programming/event schedules can be overlaid with data to make relationships between the use of the public space and its context more visible.
Array of information that can help us understand public space
While the real-time nature of the data can be used to quickly understand the immediate impact that programming and design decisions have on the space and vice versa, it can also capture data and trends over extended durations that would not otherwise be possible without considerable resources. All of this enables us to make more informed decisions rather than anecdotal observations and assumptions. With the interest of tactical urbanism and other short-term public space investments, this need for data is greater than ever. The data can be used to justify investment in long-term improvements and/or make slight adjustments in temporary spaces to increase their effectiveness….(More)”

We Need To Innovate The Science Business Model


Greg Satell at Forbes: “In 1945, Vannevar Bush, the man that led the nation’s scientific efforts during World War II, delivered a proposal to President Truman for funding scientific research in the post-war world.  Titled Science, The Endless Frontier, it led to the formation of the NSFNIHDARPA and other agencies….
One assumption inherent in Bush’s proposal was that institutions would be at the center of scientific life.  Scientists from disparate labs could read each others papers and meet at an occasional conference, but for the most part, they would be dependent on the network of researchers within their organization and those close by.
Sometimes, the interplay between institutions had major, even historical, impacts, such as John von Neumann’s sponsorship of Alan Turing, but mostly the work you did was largely a function of where you did it.  The proximity of Watson, Crick, Rosalind Franklin and Maurice Wilkins, for example, played a major role in the discovery of the structure of DNA.
Yet today, digital technology is changing not only the speed and ease of how we communicate, but the very nature of how we are able to collaborate.  When I spoke to Jonathan Adams, Chief Scientist at Digital Science, which develops and invests in software that makes science more efficient, he noted that there is a generational shift underway and said this:

When you talk to people like me, we’re established scientists who are still stuck in the old system of institutions and conferences.  But the younger scientists are using technology to access networks and they do so on an ongoing, rather than a punctuated basis.  Today, you don’t have to go to a conference or write a paper to exchange ideas.

Evidence would seem to bear this out.  The prestigious journal Nature recently noted that the average scientific paper has four times as many authors as it did in the 1950’s, when Bush’s career was at its height.  Moreover, it’s become common for co-authors to work at far-flung institutions.  Scientific practice needs to adopt to this scientific reality.
There has been some progress in this area.  The Internet, in fact, was created for the the explicit purpose of scientific collaboration.  Yet still, the way in which scientists report and share their findings remains much the same as a century ago.
Moving From Publications To Platforms For Discovery
One especially ripe area for innovation is publishing.  Typically, a researcher with a new discovery waits six months to a year for the peer review process to run its course before the work can be published.  Even then, many of the results are questionable at best.  Nature recently reported that the overwhelming majority of studies can’t be replicated…(More)”