Cinq expériences de démocratie 2.0


Le Monde: “Du 23 au 27 novembre, à Strasbourg, les participants au Forum mondial pour la démocratie examineront des initiatives de démocratie participative à l’oeuvre sur tous les continents. En voici quelques exemples. ( Lire aussi l’entretien : “Internet renforce le pouvoir de la société civile”)

  • EN FRANCE, LES ÉLECTEURS PASSENT À L’ÈRE NUMÉRIQUE

Depuis trois ans, les initiatives françaises de démocratie 2.0 se multiplient, avec pour objectif de stimuler la participation citoyenne aux instances démocratiques, qu’elles soient locales ou nationales. Dans la perspective des élections municipales de mars 2014, Questionnezvoselus.org propose ainsi aux internautes d’interroger les candidats à la mairie des 39 villes de France métropolitaine de plus de 100 000 habitants. Objectif ? Etablir la confiance entre les citoyens et leurs élus grâce à davantage de transparence, d’autonomisation et de responsabilité. La démarche rappelle celle de Voxe.org : lors de l’élection présidentielle de 2012, ce comparateur neutre et indépendant des programmes des candidats a enregistré un million de connexions. En complément, Laboxdesmunicipales.com propose des outils d’aide au vote, tandis que Candidat-et-citoyens.fr offre à ceux qui se présentent la possibilité d’associer des citoyens à la construction de leur programme.
Aux adeptes de la transparence, le collectif Democratieouverte.org propose d’interpeller les élus afin qu’ils affichent ouvertement leurs pratiques, et Regardscitoyens.org offre « un accès simplifié au fonctionnement de nos institutions démocratiques à partir des informations publiques »….

Informer, débattre et donner le pouvoir d’agir », tel est le slogan de Puzzled by Policy (PBP, « perplexe quant à la politique »), une plate-forme Internet lancée en octobre 2010 afin d’aider chacun à mieux comprendre les décisions politiques prises au niveau européen et à améliorer ainsi la qualité du débat public….

  • A PORTO ALEGRE,  UN WIKI RELIE HABITANTS ET ÉDILES

Cartographier le territoire et identifier les problèmes que rencontrent les habitants de la ville en utilisant un système « wiki » (c’est-à-dire un site Internet qui s’enrichit des contributions des internautes), telle est la vocation de Porto Alegre.cc.
Conçu pour donner de la visibilité aux causes défendues par les habitants, ce site s’inscrit dans le cadre de la plate-forme « wikicity » (Wikicidade.cc). Un concept dondé sur la méthode de l’intelligence collective, qui s’articule autour de quatre axes : culture de la citoyenneté, éthique de l’attention, responsabilité partagée et engagement civique….

  • EN FINLANDE,  CHACUN LÉGIFÈRE EN LIGNE

Depuis mars 2012, la Constitution finlandaise laisse à tout citoyen ayant atteint la majorité la possibilité d’inscrire des propositions de loi sur l’agenda parlementaire. Ces dernières sont examinées par les élus à condition de recevoir le soutien de 50 000 autres Finlandais (soit 1 % de la population).
Afin d’optimiser l’usage et l’impact de ce dispositif de participation citoyenne, l’ONG Open Ministry a lancé en octobre 2012 une plate-forme facilitant l’implication de tout un chacun. Participation en ligne, ateliers de travail ouverts ou tables rondes sont autant de techniques utilisées à cette fin….

  • AUX ETATS-UNIS, LA FINANCE PARTICIPATIVE GAGNE LES PROJETS PUBLICS

Citizinvestor.com propose aux citoyens américains de participer au financement d’infrastructures publiques. « L’administration n’a jamais assez d’argent pour financer tous les projets et services dont rêvent les citoyens », observent Tony De Sisto et Jordan Tyler Raynor, les cofondateurs du projet, conscients des choix difficiles opérés lors de l’allocation des budgets municipaux et de l’envie des habitants d’avoir leur voix dans ce choix….”

The Good Judgment Project: Harnessing the Wisdom of the Crowd to Forecast World Events


The Economist: “But then comes the challenge of generating real insight into forecasting accuracy. How can one compare forecasting ability?
The only reliable method is to conduct a forecasting tournament in which independent judges ask all participants to make the same forecasts in the same timeframes. And forecasts must be expressed numerically, so there can be no hiding behind vague verbiage. Words like “may” or “possible” can mean anything from probabilities as low as 0.001% to as high as 60% or 70%. But 80% always and only means 80%.
In the late 1980s one of us (Philip Tetlock) launched such a tournament. It involved 284 economists, political scientists, intelligence analysts and journalists and collected almost 28,000 predictions. The results were startling. The average expert did only slightly better than random guessing. Even more disconcerting, experts with the most inflated views of their own batting averages tended to attract the most media attention. Their more self-effacing colleagues, the ones we should be heeding, often don’t get on to our radar screens.
That project proved to be a pilot for a far more ambitious tournament currently sponsored by the Intelligence Advanced Research Projects Activity (IARPA), part of the American intelligence world. Over 5,000 forecasters have made more than 1m forecasts on more than 250 questions, from euro-zone exits to the Syrian civil war. Results are pouring in and they are revealing. We can discover who has better batting averages, not take it on faith; discover which methods of training promote accuracy, not just track the latest gurus and fads; and discover methods of distilling the wisdom of the crowd.
The big surprise has been the support for the unabashedly elitist “super-forecaster” hypothesis. The top 2% of forecasters in Year 1 showed that there is more than luck at play. If it were just luck, the “supers” would regress to the mean: yesterday’s champs would be today’s chumps. But they actually got better. When we randomly assigned “supers” into elite teams, they blew the lid off IARPA’s performance goals. They beat the unweighted average (wisdom-of-overall-crowd) by 65%; beat the best algorithms of four competitor institutions by 35-60%; and beat two prediction markets by 20-35%.
Over to you
To avoid slipping back to business as usual—believing we know things that we don’t—more tournaments in more fields are needed, and more forecasters. So we invite you, our readers, to join the 2014-15 round of the IARPA tournament. Current questions include: Will America and the EU reach a trade deal? Will Turkey get a new constitution? Will talks on North Korea’s nuclear programme resume? To volunteer, go to the tournament’s website at www.goodjudgmentproject.com. We predict with 80% confidence that at least 70% of you will enjoy it—and we are 90% confident that at least 50% of you will beat our dart-throwing chimps.”
See also https://web.archive.org/web/2013/http://www.iarpa.gov/Programs/ia/ACE/ace.html
 

6 Projects That Make Data More Accessible Win $100,000 Each From Gates


Chronicle of Philanthropy: “Six nonprofit projects that aim to combine multiple sets of data to help solve social problems have each won $100,000 grants from the Bill & Melinda Gates Foundation…The winners:
• Pushpa Aman Singh, who founded GuideStar India as an effort of the Civil Society Information Services India. GuideStar India is the most comprehensive database of India’s registered charities. It has profiles of more than 4,000 organizations, and Ms. Singh plans to expand that number and the types of information included.
• Development Initiatives, an international aid organization, to support its partnership with the Ugandan nonprofit Development Research and Training. Together, they are trying to help residents of two districts in Uganda identify a key problem the communities face and use existing data sets to build both online and offline tools to help tackle that challenge…
• H.V. Jagadish, at the University of Michigan, to develop a prototype that will merge sets of incompatible geographic data to make them comparable. Mr. Jagadish, a professor of electrical engineering and computer science, points to crime precincts and school districts as an example. “We want to understand the impact of education on crime, but the districts don’t quite overlap with the precincts,” he says. “This tool will address the lack of overlap.”
• Vijay Modi, at Columbia University, to work with government agencies and charities in Nigeria on a tool similar to Foursquare, the social network that allows people to share their location with friends. Mr. Modi, a mechanical-engineering professor and faculty member of the university’s Earth Institute, envisions a tool that will help people find important resources more easily…
• Gisli Olafsson and his team at NetHope, a network of aid organizations. The group is building a tool to help humanitarian charities share their data more widely and in real time—potentially saving more lives during disasters…
• Development Gateway, a nonprofit that assists international development charities with technology, and GroundTruth Initiative, a nonprofit that helps residents of communities learn mapping and media skills. The two groups want to give people living in the slums of Nairobi, Kenya, more detailed information about local schools…”

Crisis response needs to be a science, not an art


Jimmy Whitworth in the Financial Times:”…It is an imperative to offer shelter, nutrition, sanitation and medical care to those suddenly bereft of it. Without aid, humanitarian crises would cause still greater suffering. Yet admiration for the agencies that deliver relief should not blind us to the need to ensure that it is well delivered. Humanitarian responses must be founded on good evidence.
The evidence base, unfortunately, is weak. We know that storms, earthquakes and conflicts have devastating consequences for health and wellbeing, and that not responding is not an option, but we know surprisingly little about how best to go about it. Not only is evidence-based practice rare in humanitarian relief operations, it is often impossible.
Questions about how best to deliver clean water or adequate shelter, and even about which health needs should be prioritised as the most pressing, have often been barely researched. Indeed, the evidence gap is so great that the Humanitarian Practice Network has highlighted a “dire lack of credible data to help us understand just how much populations in crisis suffer, and to what extent relief operations are able to relieve that suffering”. No wonder aid responses are often characterised as messy.
Good practice often rests on past practice rather than research. The Bible of humanitarian relief is a document called the Sphere handbook, an important initiative to set minimum standards for provision of health, nutrition, sanitation and shelter. Yet analysis of the 2004 handbook has revealed that just 13 per cent of its 346 standards were supported by good evidence of relevance to health. The handbook, for example, recommended that refugee camps should prioritise measles vaccination – a worthwhile goal, but not one that should clearly be favoured over control of other infectious diseases.

Also under-researched is the question of how best to provide types of relief that everybody agrees meet essential needs. Access to clean water is a clear priority for almost all populations in crisis but little is understood about how this is most efficiently delivered. Is it best to ship bottled water to stricken areas? Are tankers of clean water more effective? Or can water purification tablets do the job? The summer floods in northern India made it clear that there is little good evidence one way or another.

Adequate shelter, too, is a human essential in all but the most benign environments but, once again, the evidence base about how best to provide it is limited. There is a school of thought that building transitional shelter from locally available materials is better in the long run than housing people under tents, tarpaulins and plastic, which if accurate would have far-reaching consequences for standard practice. But too little research has been done…
Researchers also face significant challenges to building a better evidence base. They can struggle to secure access to disaster zones when getting relief in is the priority. The timescales involved in applying for funding and ethical approval, too, make it difficult for them to move quickly enough to set up a study in the critical post-disaster period.
It is to address this that Enhancing Learning and Research for Humanitarian Assistance, with the support of the Wellcome Trust and the UK Department for International Development, recently launched an £8m research programme that investigates these issues.”

Typhoon Yolanda: UN Needs Your Help Tagging Crisis Tweets for Disaster Response


Patrick Meyer: “The United Nations Office for the Coordination of Humanitarian Affairs (OCHA) just activated the Digital Humanitarian Network (DHN) in response to Typhoon Yolanda, which has already been described as possibly one of the strongest Category 5 storms in history. The Standby Volunteer Task Force (SBTF) was thus activated by the DHN to carry out a rapid needs & damage assessment by tagging reports posted to social media. So Ji Lucas and I at QCRI (+ Hemant & Andrew) have launched MicroMappers in partnership with the SBTF to micro-task the tagging of tweets. We need all the help we can get given the volume we’ve collected (and are continuing to collect). This is where you come in!
TweetClicker_PH2
You don’t need any prior experience or training, nor do you need to create an account or even login to use the MicroMappers TweetClicker. If you can read and use a computer mouse, then you’re all set to be a Digital Humanitarian! Just click here to get started. Every tweet will get tagged by 3 different volunteers (to ensure quality control) and those tweets that get identical tags will be shared with our UN colleagues in the Philippines. All this and more is explained in the link above, which will give you a quick intro so you can get started right away. Our UN colleagues need these tags to better understand who needs help and what areas have been affected.”

Crowdsourcing the sounds of cities’ quiet spots


Springwise: “Finding a place in the city to collect your thoughts and enjoy some quietude is a rare thing. While startups such as Breather are set to open up private spaces for work and relaxation in several US cities, a new project called Stereopublic is hoping to map the ones already there, recruiting citizens to collect the sounds of those spaces.
Participants can download the free iOS app created by design studio Freerange Future, which enables them to become an ‘earwitness’. When they discover a tranquil spot in their city, they can use their GPS co-ordinates to record its exact location on the Stereopublic map, as well as record a 30-second sound clip and take a photo to give others a better idea of what it’s like. The team then works with sound experts to create quiet tours of each participating city, which currently includes Adelaide, London, LA, New York City, Singapore and 26 other global cities. The video below offers some more information about the project:

 

Big Data


Special Report on Big Data by Volta – A newsletter on Science, Technology and Society in Europe:  “Locating crime spots, or the next outbreak of a contagious disease, Big Data promises benefits for society as well as business. But more means messier. Do policy-makers know how to use this scale of data-driven decision-making in an effective way for their citizens and ensure their privacy?90% of the world’s data have been created in the last two years. Every minute, more than 100 million new emails are created, 72 hours of new video are uploaded to YouTube and Google processes more than 2 million searches. Nowadays, almost everyone walks around with a small computer in their pocket, uses the internet on a daily basis and shares photos and information with their friends, family and networks. The digital exhaust we leave behind every day contributes to an enormous amount of data produced, and at the same time leaves electronic traces that contain a great deal of personal information….
Until recently, traditional technology and analysis techniques have not been able to handle this quantity and type of data. But recent technological developments have enabled us to collect, store and process data in new ways. There seems to be no limitations, either to the volume of data or technology for storing and analyzing them. Big Data can map a driver’s sitting position to identify a car thief, it can use Google searches to predict outbreaks of the H1N1 flu virus, it can data-mine Twitter to predict the price of rice or use mobile phone top-ups to describe unemployment in Asia.
The word ‘data’ means ‘given’ in Latin. It commonly refers to a description of something that can be recorded and analyzed. While there is no clear definition of the concept of ‘Big Data’, it usually refers to the processing of huge amounts and new types of data that have not been possible with traditional tools.

‘The new development is not necessarily that there are so much more data. It’s rather that data is available to us in a new way.’

The notion of Big Data is kind of misleading, argues Robindra Prabhu, a project manager at the Norwegian Board of Technology. “The new development is not necessarily that there are so much more data. It’s rather that data is available to us in a new way. The digitalization of society gives us access to both ‘traditional’, structured data – like the content of a database or register – and unstructured data, for example the content in a text, pictures and videos. Information designed to be read by humans is now also readable by machines. And this development makes a whole new world of  data gathering and analysis available. Big Data is exciting not just because of the amount and variety of data out there, but that we can process data about so much more than before.”

Smart Citizens


FutureEverything: “This publication aims to shift the debate on the future of cities towards the central place of citizens, and of decentralised, open urban infrastructures. It provides a global perspective on how cities can create the policies, structures and tools to engender a more innovative and participatory society. The publication contains a series of 23 short essays representing some of the key voices developing an emerging discourse around Smart Citizens.  Contributors include:

  • Dan Hill, Smart Citizens pioneer and CEO of communications research centre and transdisciplinary studio Fabrica on why Smart Citizens Make Smart Cities.
  • Anthony Townsend, urban planner, forecaster and author of Smart Cities: Big Data, Civic Hackers, and the Quest for a New Utopia on the tensions between place-making and city-making on the role of mobile technologies in changing the way that people interact with their surroundings.
  • Paul Maltby, Director of the Government Innovation Group and of the Open Data and Transparency in the UK Cabinet Office on how government can support a smarter society.
  • Aditya Dev Sood, Founder and CEO of the Center for Knowledge Societies, presents polarised hypothetical futures for India in 2025 that argues for the use of technology to bridge gaps in social inequality.
  • Adam Greenfield, New York City-based writer and urbanist, on Recuperating the Smart City.

Editors: Drew Hemment, Anthony Townsend
Download Here.

What Government Can and Should Learn From Hacker Culture


in The Atlantic: “Can the open-source model work for federal government? Not in every way—for security purposes, the government’s inner workings will never be completely open to the public. Even in the inner workings of government, fears of triggering the next Wikileaks or Snowden scandal may scare officials away from being more open with one another. While not every area of government can be more open, there are a few areas ripe for change.

Perhaps the most glaring need for an open-source approach is in information sharing. Today, among and within several federal agencies, a culture of reflexive and unnecessary information withholding prevails. This knee-jerk secrecy can backfire with fatal consequences, as seen in the 1998 embassy bombings in Africa, the 9/11 attacks, and the Boston Marathon bombings. What’s most troubling is that decades after the dangers of information-sharing were identified, the problem persists.
What’s preventing reform? The answer starts with the government’s hierarchical structure—though an information-is-power mentality and “need to know” Cold War-era culture contribute too. To improve the practice of information sharing, government needs to change the structure of information sharing. Specifically, it needs to flatten the hierarchy.
Former Obama Administration regulation czar Cass Sunstein’s “nudge” approach shows how this could work. In his book Simpler: The Future of Government, he describes how making even small changes to an environment can affect significant changes in behavior. While Sunstein focuses on regulations, the broader lesson is clear: Change the environment to encourage better behavior and people tend to exhibit better behavior. Without such strict adherence to the many tiers of the hierarchy, those working within it could be nudged towards, rather than fight to, share information.
One example of where this worked is in with the State Department’s annual Religious Engagement Report (RER). In 2011, the office in charge of the RER decided that instead of having every embassy submit their data via email, they would post it on a secure wiki. On the surface, this was a decision to change an information-sharing procedure. But it also changed the information-sharing culture. Instead of sharing information only along the supervisor-subordinate axis, it created a norm of sharing laterally, among colleagues.
Another advantage to flattening information-sharing hierarchies is that it reduces the risk of creating “single points of failure,” to quote technology scholar Beth Noveck. The massive amounts of data now available to us may need massive amounts of eyeballs in order to spot patterns of problems—small pools of supervisors atop the hierarchy cannot be expected to shoulder those burdens alone. And while having the right tech tools to share information is part of the solution—as the wiki made it possible for the RER—it’s not enough. Leadership must also create a culture that nudges their staff to use these tools, even if that means relinquishing a degree of their own power.
Finally, a more open work culture would help connect interested parties across government to let them share the hard work of bringing new ideas to fruition. Government is filled with examples of interesting new projects that stall in their infancy. Creating a large pool of collaborators dedicated to a project increases the likelihood that when one torchbearer burns out, others in the agency will pick up for them.
When Linus Torvalds released Linux, it was considered, in Raymond’s words, “subversive” and “a distinct shock.” Could the federal government withstand such a shock?
Evidence suggests it can—and the transformation is already happening in small ways. One of the winners of the Harvard Kennedy School’s Innovations in Government award is State’s Consular Team India (CTI), which won for joining their embassy and four consular posts—each of which used to have its own distinct set of procedures-into a single, more effective unit who could deliver standardized services. As CTI describes it, “this is no top-down bureaucracy” but shares “a common base of information and shared responsibilities.” They flattened the hierarchy, and not only lived, but thrived.”

Open Data Index provides first major assessment of state of open government data


Press Release from the Open Knowledge Foundation: “In the week of a major international summit on government transparency in London, the Open Knowledge Foundation has published its 2013 Open Data Index, showing that governments are still not providing enough information in an accessible form to their citizens and businesses.
The UK and US top the 2013 Index, which is a result of community-based surveys in 70 countries. They are followed by Denmark, Norway and the Netherlands. Of the countries assessed, Cyprus, St Kitts & Nevis, the British Virgin Islands, Kenya and Burkina Faso ranked lowest. There are many countries where the governments are less open but that were not assessed because of lack of openness or a sufficiently engaged civil society. This includes 30 countries who are members of the Open Government Partnership.
The Index ranks countries based on the availability and accessibility of information in ten key areas, including government spending, election results, transport timetables, and pollution levels, and reveals that whilst some good progress is being made, much remains to be done.
Rufus Pollock, Founder and CEO of the Open Knowledge Foundation said:

Opening up government data drives democracy, accountability and innovation. It enables citizens to know and exercise their rights, and it brings benefits across society: from transport, to education and health. There has been a welcome increase in support for open data from governments in the last few years, but this Index reveals that too much valuable information is still unavailable.

The UK and US are leaders on open government data but even they have room for improvement: the US for example does not provide a single consolidated and open register of corporations, while the UK Electoral Commission lets down the UK’s good overall performance by not allowing open reuse of UK election data.
There is a very disappointing degree of openness of company registers across the board: only 5 out of the 20 leading countries have even basic information available via a truly open licence, and only 10 allow any form of bulk download. This information is critical for range of reasons – including tackling tax evasion and other forms of financial crime and corruption.
Less than half of the key datasets in the top 20 countries are available to re-use as open data, showing that even the leading countries do not fully understand the importance of citizens and businesses being able to legally and technically use, reuse and redistribute data. This enables them to build and share commercial and non-commercial services.
To see the full results: https://index.okfn.org. For graphs of the data: https://index.okfn.org/visualisations.”