Crowdsourcing Expertise


Simons Foundation: “Ever wish there was a quick, easy way to connect your research to the public?

By hosting a Wikipedia ‘edit-a-thon’ at a science conference, you can instantly share your research knowledge with millions while improving the science content on the most heavily trafficked and broadly accessible resource in the world. In 2016, in partnership with the Wiki Education Foundation, we helped launched the Wikipedia Year of Science, an ambitious initiative designed to better connect the work of scientists and students to the public. Here, we share some of what we learned.

The Simons Foundation — through its Science Sandbox initiative, dedicated to public engagement — co-hosted a series of Wikipedia edit-a-thons throughout 2016 at almost every major science conference, in collaboration with the world’s leading scientific societies and associations.

At our edit-a-thons, we leveraged the collective brainpower of scientists, giving them basic training on Wikipedia guidelines and facilitating marathon editing sessions — powered by free pizza, coffee and sometimes beer — during which they made copious contributions within their respective areas of expertise.

These efforts, combined with the Wiki Education Foundation’s powerful classroom model, have had a clear impact. To date, we’ve reached over 150 universities including more than 6,000 students and scientists. As for output, 6,306 articles have been created or edited, garnering more than 304 million views; over 2,000 scientific images have been donated; and countless new scientist-editors have been minted, many of whom will likely continue to update Wikipedia content. The most common response we got from scientists and conference organizers about the edit-a-thons was: “Can we do that again next year?”

That’s where this guide comes in.

Through collaboration, input from Wikipedians and scientists, and more than a little trial and error, we arrived at a model that can help you organize your own edit-a-thons. This informal guide captures our main takeaways and lessons learned….Our hope is that edit-a-thons will become another integral part of science conferences, just like tweetups, communication workshops and other recent outreach initiatives. This would ensure that the content of the public’s most common gateway to science research will continually improve in quality and scope.

Download: “Crowdsourcing Expertise: A working guide for organizing Wikipedia edit-a-thons at science conferences

Fighting Illegal Fishing With Big Data


Emily Matchar in Smithsonian: “In many ways, the ocean is the Wild West. The distances are vast, the law enforcement agents few and far between, and the legal jurisdiction often unclear. In this environment, illegal activity flourishes. Illegal fishing is so common that experts estimate as much as a third of fish sold in the U.S. was fished illegally. This illegal fishing decimates the ocean’s already dwindling fish populations and gives rise to modern slavery, where fishermen are tricked onto vessels and forced to work, sometimes for years.

A new use of data technology aims to help curb these abuses by shining a light on the high seas. The technology uses ships’ satellite signals to detect instances of transshipment, when two vessels meet at sea to exchange cargo. As transshipment is a major way illegally caught fish makes it into the legal supply chain, tracking it could potentially help stop the practice.

“[Transshipment] really allows people to do something out of sight,” says David Kroodsma, the research program director at Global Fishing Watch, an online data platform launched by Google in partnership with the nonprofits Oceana and SkyTruth. “It’s something that obscures supply chains. It’s basically being able to do things without any oversight. And that’s a problem when you’re using a shared resource like the oceans.”

Global Fishing Watch analyzed some 21 billion satellite signals broadcast by ships, which are required to carry transceivers for collision avoidance, from between 2012 and 2016. It then used an artificial intelligence system it created to identify which ships were refrigerated cargo vessels (known in the industry as “reefers”). They then verified this information with fishery registries and other sources, eventually identifying 794 reefers—90 percent of the world’s total number of such vessels. They tracked instances where a reefer and a fishing vessel were moving at similar speeds in close proximity, labeling these instances as “likely transshipments,” and also traced instances where reefers were traveling in a way that indicated a rendezvous with a fishing vessel, even if no fishing vessel was present—fishing vessels often turn off their satellite systems when they don’t want to be seen. All in all there were more than 90,000 likely or potential transshipments recorded.

Even if these encounters were in fact transshipments, they would not all have been for nefarious purposes. They may have taken place to refuel or load up on supplies. But looking at the patterns of where the potential transshipments happen is revealing. Very few are seen close to the coasts of the U.S., Canada and much of Europe, all places with tight fishery regulations. There are hotspots off the coast of Peru and Argentina, all over Africa, and off the coast of Russia. Some 40 percent of encounters happen in international waters, far enough off the coast that no country has jurisdiction.

The tracked reefers were flying flags from some 40 different countries. But that doesn’t necessarily tell us much about where they really come from. Nearly half of the reefers tracked were flying “flags of convenience,” meaning they’re registered in countries other than where the ship’s owners are from to take advantage of those countries’ lax regulations….(More)”

Read more: http://www.smithsonianmag.com/innovation/fighting-illegal-fishing-big-data-180962321/#7eCwGrGS5v5gWjFz.99
Give the gift of Smithsonian magazine for only $12! http://bit.ly/1cGUiGv
Follow us: @SmithsonianMag on Twitter

The Datafied Society. Studying Culture through Data


(Open Access) book edited by Mirko Tobias Schäfer & Karin van Es: “As more and more aspects of everyday life are turned into machine-readable data, researchers are provided with rich resources for researching society. The novel methods and innovative tools to work with this data not only require new knowledge and skills, but also raise issues concerning the practices of investigation and publication. This book critically reflects on the role of data in academia and society and challenges overly optimistic expectations considering data practices as means for understanding social reality. It introduces its readers to the practices and methods for data analysis and visualization and raises questions not only about the politics of data tools, but also about the ethics in collecting, sifting through data, and presenting data research. AUP S17 Catalogue text As machine-readable data comes to play an increasingly important role in everyday life, researchers find themselves with rich resources for studying society. The novel methods and tools needed to work with such data require not only new knowledge and skills, but also a new way of thinking about best research practices. This book critically reflects on the role and usefulness of big data, challenging overly optimistic expectations about what such information can reveal, introducing practices and methods for its analysis and visualization, and raising important political and ethical questions regarding its collection, handling, and presentation….(More)”.

Education startup helps refugees earn university degree


Springwise: “Berlin-based Kiron works with refugee students to put together an online course of study, rigorous enough to provide entry into a partner university’s second year of study. Using Massive Open Online Courses (MOOCs), Kiron helps students master their new country’s language while studying basic prerequisites for a chosen university degree. Already working with more than 1,500 students in Germany, Kiron recently expanded into France.

With less than one percent of all refugees able to access higher education, MOOCs help get new students to the necessary level of knowledge for in-person university study. Kiron also provides off-line support including study buddy programs and career guidance. Once a participant completes the two-year online program, he or she has the opportunity to enroll for free (as a second year student) in one of Kiron’s partner university’s programs.

A number of projects are finding ways to use the talents of refugees to help them integrate into their country through knowledge-sharing and employment opportunities. Locals and refugees work together in this new Dutch ideas hub, and this French catering company hires refugee chefs….(More)”.

The Techno-Politics of Data and Smart Devolution in City-Regions: Comparing Glasgow, Bristol, Barcelona, and Bilbao


Paper by Igor Calzada: “This paper explores the substantial effect that the critical understanding and techno-political consideration of data are having in some smart city strategies. Particularly, the paper presents some results of a comparative study of four cases of smart city transitions: Glasgow, Bristol, Barcelona, and Bilbao. Likewise, considering how relevant the city-regional path-dependency is in each territorial context, the paper will elucidate the notion of smart devolution as a key governance component that is enabling some cities to formulate their own smart city-regional governance policies and implement them by considering the role of the smart citizens as decision makers rather than mere data providers. The paper concludes by identifying an implicit smart city-regional governance strategy for each case based on the techno-politics of data and smart devolution….(More)”

Thesis, antithesis and synthesis: A constructive direction for politics and policy after Brexit and Trump


Geoff Mulgan at Nesta: “In the heady days of 1989, with communism collapsing and the Cold War seemingly over, the political theorist Francis Fukuyama declared that we were witnessing the “end of history” which had culminated in the triumph of liberal democracy and the free market.

Fukuyama was drawing on the ideas of German philosopher Georg Hegel, but of course, history didn’t come to an end, and, as recent events have shown, the Cold War was just sleeping, not dead.

Now, following the political convulsions of 2016, we’re at a very different turning point, which many are trying to make sense of. I want to suggest that we can again usefully turn to Hegel, but this time to his idea that history evolves in dialectical ways, with successive phases of thesis, antithesis and synthesis.

This framework fits well with where we stand today.  The ‘thesis’ that has dominated mainstream politics for the last generation – and continues to be articulated shrilly by many proponents – is the claim that the combination of globalisation, technological progress and liberalisation empowers the great majority.

The antithesis, which, in part, fuelled the votes for Brexit and Trump, as well as the rise of populist parties and populist authoritarian leaders in Europe and beyond, is the argument that this technocratic combination merely empowers a minority and disempowers the majority of citizens.

A more progressive synthesis – which I will outline – then has to address the flaws of the thesis and the grievances of the antithesis, in fields ranging from education and health to democracy and migration, dealing head on with questions of power and its distribution: questions about who has power, and who feels powerful….(More)”

Open innovation in the public sector


Sabrina Diaz Rato in OpenDemocracy: “For some years now, we have been witnessing the emergence of relational, cross-over, participative power. This is the territory that gives technopolitics its meaning and prominence, the basis on which a new vision of democracy – more open, more direct, more interactive – is being developed and embraced. It is a framework that overcomes the closed architecture on which the praxis of governance (closed, hierarchical, one-way) have been cemented in almost all areas. The series The ecosystem of open democracy explores the different aspects of this ongoing transformation….

How can innovation contribute to building an open democracy? The answer is summed up in these ten connectors of innovation.

  1. placing innovation and collective intelligence at the center of public management strategies,
  2. aligning all government areas with clearly-defined goals on associative platforms,
  3. shifting the frontiers of knowledge and action from the institutions to public deliberation on local challenges,
  4. establishing leadership roles, in a language that everyone can easily understand, to organize and plan the wealth of information coming out of citizens’ ideas and to engage those involved in the sustainability of the projects,
  5. mapping the ecosystem and establishing dynamic relations with internal and, particularly, external agents: the citizens,
  6. systematizing the accumulation of information and the creative processes, while communicating progress and giving feedback to the whole community,
  7. preparing society as a whole to experience a new form of governance of the common good,
  8. cooperating with universities, research centers and entrepreneurs in establishing reward mechanisms,
  9. aligning people, technologies, institutions and the narrative with the new urban habits, especially those related to environmental sustainability and public services,
  10. creating education and training programs in tune with the new skills of the 21st century,
  11. building incubation spaces for startups responding to local challenges,
  12. inviting venture capital to generate a satisfactory mix of open innovation, inclusive development policies and local productivity.

Two items in this list are probably the determining factors of any effective innovation process. The first has to do with the correct decision on the mechanisms through which we have pushed the boundaries outwards, so as to bring citizen ideas into the design and co-creation of solutions. This is not an easy task, because it requires a shared organizational mentality on previously non-existent patterns of cooperation, which must now be sustained through dialog and operational dynamics aimed at solving problems defined by external actors – not just any problem.

Another key aspect of the process, related to the breaking down of the institutional barriers that surround and condition action frameworks, is the revaluation of a central figure that we have not yet mentioned here: the policy makers. They are not exactly political leaders or public officials. They are not innovators either. They are the ones within Public Administration who possess highly valuable management skills and knowledge, but who are constantly colliding against the glittering institutional constellations that no longer work….(More)”

From big data to smart data: FDA’s INFORMED initiative


Sean KhozinGeoffrey Kim & Richard Pazdur in Nature: “….Recent advances in our understanding of disease mechanisms have led to the development of new drugs that are enabling precision medicine. For example, the co-development of kinase inhibitors that target ‘driver mutations’ in metastatic non-small-cell lung cancer (NSCLC) with companion diagnostics has led to substantial improvements in the treatment of some patients. However, growing evidence suggests that most patients with metastatic NSCLC and other advanced cancers may not have tumours with single driver mutations. Furthermore, the generation of clinical evidence in genomically diverse and geographically dispersed groups of patients using traditional trial designs and multiple competing therapies is becoming more costly and challenging.

Strategies aimed at creating new efficiencies in clinical evidence generation and extending the benefits of precision medicine to larger groups of patients are driving a transformation from a reductionist approach to drug development (for example, a single drug targeting a driver mutation and traditional clinical trials) to a holistic approach (for example, combination therapies targeting complex multiomic signatures and real-world evidence). This transition is largely fuelled by the rapid expansion in the four dimensions of biomedical big data, which has created a need for greater organizational and technical capabilities (Fig. 1). Appropriate management and analysis of such data requires specialized tools and expertise in health information technology, data science and high-performance computing. For example, efforts to generate clinical evidence using real-world data are being limited by challenges such as capturing clinically relevant variables from vast volumes of unstructured content (such as physician notes) in electronic health records and organizing various structured data elements that are primarily designed to support billing rather than clinical research. So, new standards and quality-control mechanisms are needed to ensure the validity of the design and analysis of studies based on electronic health records.

Figure 1: Conceptual map of technical and organizational capacity for biomedical big data.
Conceptual map of technical and organizational capacity for biomedical big data.

Big data can be defined as having four dimensions: volume (data size), variety (data type), veracity (data noise and uncertainty) and velocity (data flow and processing). Currently, FDA approval decisions are generally based on data of limited variety, mainly from clinical trials and preclinical studies (1) that are mostly structured (2), in data sets usually no more than a few gigabytes in size (3), that are processed intermittently as part of regulatory submissions (4). The expansion of big data in the four dimensions (grey lines) calls for increasing organizational and technical capacity. This could transform big data into smart data by enabling a holistic approach to personalization of therapies that takes patient, disease and environmental characteristics into account. (Full size image (309 KB);Download PowerPoint slide (492 KB)More)”

Crowdsourcing Cybersecurity: Cyber Attack Detection using Social Media


Paper by Rupinder Paul Khandpur, Taoran Ji, Steve Jan, Gang Wang, Chang-Tien Lu, Naren Ramakrishnan: “Social media is often viewed as a sensor into various societal events such as disease outbreaks, protests, and elections. We describe the use of social media as a crowdsourced sensor to gain insight into ongoing cyber-attacks. Our approach detects a broad range of cyber-attacks (e.g., distributed denial of service (DDOS) attacks, data breaches, and account hijacking) in an unsupervised manner using just a limited fixed set of seed event triggers. A new query expansion strategy based on convolutional kernels and dependency parses helps model reporting structure and aids in identifying key event characteristics. Through a large-scale analysis over Twitter, we demonstrate that our approach consistently identifies and encodes events, outperforming existing methods….(More)”

Open Data Privacy Playbook


A data privacy playbook by Ben Green, Gabe Cunningham, Ariel Ekblaw, Paul Kominers, Andrew Linzer, and Susan Crawford: “Cities today collect and store a wide range of data that may contain sensitive or identifiable information about residents. As cities embrace open data initiatives, more of this information is available to the public. While releasing data has many important benefits, sharing data comes with inherent risks to individual privacy: released data can reveal information about individuals that would otherwise not be public knowledge. In recent years, open data such as taxi trips, voter registration files, and police records have revealed information that many believe should not be released.

Effective data governance is a prerequisite for successful open data programs. The goal of this document is to codify responsible privacy-protective approaches and processes that could be adopted by cities and other government organizations that are publicly releasing data. Our report is organized around four recommendations:

  • Conduct risk-benefit analyses to inform the design and implementation of open data programs.
  • Consider privacy at each stage of the data lifecycle: collect, maintain, release, delete.
  • Develop operational structures and processes that codify privacy management widely throughout the City.
  • Emphasize public engagement and public priorities as essential aspects of data management programs.

Each chapter of this report is dedicated to one of these four recommendations, and provides fundamental context along with specific suggestions to carry them out. In particular, we provide case studies of best practices from numerous cities and a set of forms and tactics for cities to implement our recommendations. The Appendix synthesizes key elements of the report into an Open Data Privacy Toolkit that cities can use to manage privacy when releasing data….(More)”