Curated findings and actionable knowledge at the intersection of technology, innovation, and governance
Country and Region: Latin America and the Caribbean
Anguilla, Antigua and Barbuda, Argentina, Aruba, Bahamas, Barbados, Belize, Bolivia, Bonaire, Bouvet Island, Brazil, British Virgin Islands, Cayman Islands, Chile, Colombia, Costa Rica, Cuba, Curaçao, Dominica, Dominican Republic, Ecuador, El Salvador, Falkland Islands (Malvinas), French Guiana, Grenada, Guadeloupe, Guatemala, Guyana, Haiti, Honduras, Jamaica, Martinique, Mexico, Montserrat, Nicaragua, Panama, Paraguay, Peru, Puerto Rico, Saint Barthélemy, Saint Kitts and Nevis, Saint Lucia, Saint Martin (French Part), Saint Vincent and the Grenadines, Sint Eustatius and Saba, Sint Maarten (Dutch part), South Georgia and the South Sandwich Islands, Suriname, Trinidad and Tobago, Turks and Caicos Islands, United States Virgin Islands, Uruguay, Venezuela
Data Driven Journalism: “UN-Habitat has launched a new web portal featuring a wealth of city data based on its repository of research on urban trends.
Launched during the 25th Governing Council, the Urban Data Portal allows users to explore data from 741 cities in 220 countries, and compare these for 103 indicators such as slum prevalence and city prosperity.
Image: A comparison of share in national urban population and average annual rate of urban population change for San Salvador, El Salvador, and Asuncion, Paraguay.
The urban indicators data available are analyzed, compiled and published by UN-Habitat’s Global Urban Observatory, which supports governments, local authorities and civil society organizations to develop urban indicators, data and statistics.
Leveraging GIS technology, the Observatory collects data by taking aerial photographs, zooming into particular areas, and then sending in survey teams to answer any remaining questions about the area’s urban development.
The Portal also contains data collected by national statistics authorities, via household surveys and censuses, with analysis conducted by leading urbanists in UN-HABITAT’s State of the World’s Cities and the Global Report on Human Settlements report series.
For the first time, these datasets are available for use under an open licence agreement, and can be downloaded in straightforward database formats like CSV and JSON….(More)
Ana Campoy at Quartz: “Mexico City just launched a massive experiment in digital democracy. It is asking its nearly 9 million residents to help draft a new constitution through social media. The crowdsourcing exercise is unprecedented in Mexico—and pretty much everywhere else.
as locals are known, can petition for issues to be included in the constitution through Change.org (link inSpanish), and make their case in person if they gather more than 10,000 signatures. They can also annotate proposals by the constitution drafters via PubPub, an editing platform (Spanish) similar to GoogleDocs.
The idea, in the words of the mayor, Miguel Angel Mancera, is to“bestow the constitution project (Spanish) with a democratic,progressive, inclusive, civic and plural character.”
There’s a big catch, however. The constitutional assembly—the body that has the final word on the new city’s basic law—is under no obligation to consider any of the citizen input. And then there are the practical difficulties of collecting and summarizing the myriad of views dispersed throughout one of the world’s largest cities.
That makes Mexico City’s public-consultation experiment a big test for the people’s digital power, one being watched around the world.Fittingly, the idea of crowdsourcing a constitution came about in response to an attempt to limit people power.
Fittingly, the idea of crowdsourcing a constitution came about in response to an attempt to limit people power.
For decades, city officials had fought to get out from under the thumb of the federal government, which had the final word on decisions such as who should be the city’s chief of police. This year, finally, they won a legal change that turns the Distrito Federal (federal district), similar to the US’s District of Columbia, into Ciudad de México (Mexico City), a more autonomous entity, more akin to a state. (Confusingly, it’s just part of the larger urban area also colloquially known as Mexico City, which spills into neighboring states.)
However, trying to retain some control, the Mexican congress decided that only 60% of the delegates to the city’s constitutional assembly would be elected by popular vote. The rest will be assigned by the president, congress, and Mancera, the mayor. Mancera is also the only one who can submit a draft constitution to the assembly.
Mancera’s response was to create a committee of some 30 citizens(Spanish), including politicians, human-rights advocates, journalists,and even a Paralympic gold medalist, to write his draft. He also calledfor the development of mechanisms to gather citizens’ “aspirations,values, and longing for freedom and justice” so they can beincorporated into the final document.
The mechanisms, embedded in an online platform (Spanish) that offersvarious ways to weigh in, were launched at the end of March and willcollect inputs until September 1. The drafting group has until themiddle of that month to file its text with the assembly, which has toapprove the new constitution by the end of January.
An experiment with few precedents
Mexico City didn’t have a lot of examples to draw on, since not a lot ofplaces have experience with crowdsourcing laws. In the US, a few locallawmakers have used Wiki pages and GitHub to draft bills, says MarilynBautista, a lecturer at Stanford Law School who has researched thepractice. Iceland—with a population some 27 times smaller than MexicoCity’s—famously had its citizens contribute to its constitution withinput from social media. The effort failed after the new constitution gotstuck in parliament.
In Mexico City, where many citizens already feel left out, the first bighurdle is to convince them it’s worth participating….
Then comes the task of making sense of the cacophony that will likelyemerge. Some of the input can be very easily organized—the results ofthe survey, for example, are being graphed in real time. But there could be thousands of documents and comments on the Change.org petitionsand the editing platform.
Ideas are grouped into 18 topics, such as direct democracy,transparency and economic rights. They are prioritized based on theamount of support they’ve garnered and how relevant they are, saidBernardo Rivera, an adviser for the city. Drafters get a weekly deliveryof summarized citizen petitions….
An essay about human rights on the PubPub platform.(PubPub)
The most elaborate part of the system is PubPub, an open publishing platform similar to Google Docs, which is based on a project originally developed by MIT’s Media Lab. The drafters are supposed to post essays on how to address constitutional issues, and potentially, the constitution draft itself, once there is one. Only they—or whoever they authorize—will be able to reword the original document.
User comments and edits are recorded on a side panel, with links to the portion of text they refer to. Another screen records every change, so everyone can track which suggestions have made it into the text. Members of the public can also vote comments up or down, or post their own essays….(More).
The Open Data Barometer: “Once the preserve of academics and statisticians, data has become a development cause embraced by everyone from grassroots activists to the UN Secretary-General. There’s now a clear understanding that we need robust data to drive democracy and development — and a lot of it.
Last year, the world agreed the Sustainable Development Goals (SDGs) — seventeen global commitments that set an ambitious agenda to end poverty, fight inequality and tackle climate change by 2030. Recognising that good data is essential to the success of the SDGs, the Global Partnership for Sustainable Development Data and the International Open Data Charter were launched as the SDGs were unveiled. These alliances mean the “data revolution” now has over 100 champions willing to fight for it. Meanwhile, Africa adopted the African Data Consensus — a roadmap to improving data standards and availability in a region that has notoriously struggled to capture even basic information such as birth registration.
But while much has been made of the need for bigger and better data to power the SDGs, this year’s Barometer follows the lead set by the International Open Data Charter by focusing on how much of this data will be openly available to the public.
Open data is essential to building accountable and effective institutions, and to ensuring public access to information — both goals of SDG 16. It is also essential for meaningful monitoring of progress on all 169 SDG targets. Yet the promise and possibilities offered by opening up data to journalists, human rights defenders, parliamentarians, and citizens at large go far beyond even these….
At a glance, here are this year’s key findings on the state of open data around the world:
Open data is entering the mainstream.The majority of the countries in the survey (55%) now have an open data initiative in place and a national data catalogue providing access to datasets available for re-use. Moreover, new open data initiatives are getting underway or are promised for the near future in a number of countries, including Ecuador, Jamaica, St. Lucia, Nepal, Thailand, Botswana, Ethiopia, Nigeria, Rwanda and Uganda. Demand is high: civil society and the tech community are using government data in 93% of countries surveyed, even in countries where that data is not yet fully open.
Despite this, there’s been little to no progress on the number of truly open datasets around the world.Even with the rapid spread of open government data plans and policies, too much critical data remains locked in government filing cabinets. For example, only two countries publish acceptable detailed open public spending data. Of all 1,380 government datasets surveyed, almost 90% are still closed — roughly the same as in the last edition of the Open Data Barometer (when only 130 out of 1,290 datasets, or 10%, were open). What is more, much of the approximately 10% of data that meets the open definition is of poor quality, making it difficult for potential data users to access, process and work with it effectively.
“Open-washing” is jeopardising progress. Many governments have advertised their open data policies as a way to burnish their democratic and transparent credentials. But open data, while extremely important, is just one component of a responsive and accountable government. Open data initiatives cannot be effective if not supported by a culture of openness where citizens are encouraged to ask questions and engage, and supported by a legal framework. Disturbingly, in this edition we saw a backslide on freedom of information, transparency, accountability, and privacy indicators in some countries. Until all these factors are in place, open data cannot be a true SDG accelerator.
Implementation and resourcing are the weakest links.Progress on the Barometer’s implementation and impact indicators has stalled or even gone into reverse in some cases. Open data can result in net savings for the public purse, but getting individual ministries to allocate the budget and staff needed to publish their data is often an uphill battle, and investment in building user capacity (both inside and outside of government) is scarce. Open data is not yet entrenched in law or policy, and the legal frameworks supporting most open data initiatives are weak. This is a symptom of the tendency of governments to view open data as a fad or experiment with little to no long-term strategy behind its implementation. This results in haphazard implementation, weak demand and limited impact.
The gap between data haves and have-nots needs urgent attention.Twenty-six of the top 30 countries in the ranking are high-income countries. Half of open datasets in our study are found in just the top 10 OECD countries, while almost none are in African countries. As the UN pointed out last year, such gaps could create “a whole new inequality frontier” if allowed to persist. Open data champions in several developing countries have launched fledgling initiatives, but too often those good open data intentions are not adequately resourced, resulting in weak momentum and limited success.
Governments at the top of the Barometer are being challenged by a new generation of open data adopters. Traditional open data stalwarts such as the USA and UK have seen their rate of progress on open data slow, signalling that new political will and momentum may be needed as more difficult elements of open data are tackled. Fortunately, a new generation of open data adopters, including France, Canada, Mexico, Uruguay, South Korea and the Philippines, are starting to challenge the ranking leaders and are adopting a leadership attitude in their respective regions. The International Open Data Charter could be an important vehicle to sustain and increase momentum in challenger countries, while also stimulating renewed energy in traditional open data leaders….(More)”
Guilo Quaggiotto at Nesta: “If you were looking for the cutting edge of the development sector, where would you go these days? You would probably look at startups like Premise who have predicted food trends 25 days faster than national statistics in Brazil, or GiveDirectly who are pushing the boundaries on evidence – from RCTs to new ways of mapping poverty – to fast track the adoption of cash transfers.
Or perhaps you might draw your attention to PetaJakarta who are experimenting with new responses to crises by harnessing human sensor networks. You might be tempted to consider Airbnb’s Disaster Response programme as an indicator of an emerging alternative infrastructure for disaster response (and perhaps raising questions about the political economy of this all).
And could Bitnation’s Refugee Emergency programme in response to the European refugee crisis be the possible precursor of future solutions for transnational issues – among the development sector’s hardest challenges? Are the business models of One Acre Fund, which provides services for smallholder farmers, or Floodtags, which analyses citizen data during floods for water and disaster managers, an indicator of future pathways to scale – that elusive development unicorn?
If you want to look at the future of procuring solutions for the development sector, should you be looking at initiatives like Citymart, which works with municipalities across the world to rethink traditional procurement and unleash the expertise and innovation capabilities of their citizens? By the same token, projects like Pathogen Box, Poverty Stoplight or Patient Innovation point to a brave new world where lead-user innovation and harnessing ‘sticky’ local knowledge becomes the norm, rather than the exception. You would also be forgiven for thinking that social movements across the world are the place to look for signs of future mechanisms for harnessing collective intelligence – Kawal Pamilu’s “citizen experts” self-organising around the Indonesian elections in 2014 is a textbook case study in this department.
The list could go on and on: welcome to the era of development mutants. While established players in the development sector are engrossed in soul-searching and their fitness for purpose is being scrutinised from all quarters, a whole new set of players is emerging, unfettered by legacy and borrowing from a variety of different disciplines. They point to a potentially different future – indeed, many potentially different futures – for the sector…..
But what if we wanted to invert this paradigm? How could we move from denial to fruitful collaboration with the ‘edgeryders’ of the development sector and accelerate its transformation?
Adopting new programming principles
Based on our experience working with development organisations, we believe that partnering with the mutants involves two types of shifts for traditional players: at the programmatic and the operational level. At the programmatic level, our work on the ground led us to articulate the following emerging principles:
Mapping what people have, not what they need: even though approaches like jugaad and positive deviance have been around for a long time, unfortunately the default starting point for many development projects is still mapping needs, not assets. Inverting this paradigm allows for potentially disruptive project design and partnerships to emerge. (Signs of the future: Patient Innovation, Edgeryders, Community Mirror, Premise)
Getting ready for multiple futures: When distributed across an organisation and not limited to a centralised function, the discipline of scanning the horizon for emergent solutions that contradict the dominant paradigm can help move beyond the denial phase and develop new interfaces to collaborate with the mutants. Here the link between analysis (to understand not only what is probable, but also what is possible) and action is critical – otherwise this remains purely an academic exercise. (Signs of the future: OpenCare, Improstuctures, Seeds of Good Anthropocene, Museum of the Future)
Running multiple parallel experiments: According to Dave Snowden, in order to intervene in a complex system “you need multiple parallel experiments and they should be based on different and competing theories/hypotheses”. Unfortunately, many development projects are still based on linear narratives and assumptions such as “if only we run an awareness raising campaign citizens will change their behaviour”. Turning linear narratives into hypotheses to be tested (without becoming religious on a specific approach) opens up the possibility to explore the solution landscape and collaborate with non-obvious partners that bring new approaches to the table. (Signs of the future: Chukua Hakua, GiveDirectly, Finnish PM’s Office of Experiments, Ideas42, Cognitive Edge)
Embracing obliquity: A deep, granular understanding of local assets and dynamics along with system mapping (see point 5 below) and pairing behavioural experts with development practitioners can help identify entry points for exploring new types of intervention based on obliquity principles. Mutants are often faster in adopting this approach and partnering with them is a way to bypass organisational inertia and explore nonlinear interventions. (Signs of the future: Sardex, social prescriptions, forensic architecture)
From projects to systems: development organisations genuinely interested in developing new partnerships need to make the shift from the project logic to system investments. This involves, among other things, shifting the focus from providing solutions to helping every actor in the system to develop a higher level of consciousness about the issues they are facing and to take better decisions over time. It also entails partnering with mutants to explore entirely new financial mechanisms. (Signs of the future: Lankelly Chase, Indonesia waste banks,Dark Matter Labs)
Adopting new interfaces for working with the mutants
Harvard Business School professor Carliss Baldwin argued that most bureaucracies these days have a ‘non-contractible’ problem: they don’t know where smart people are, or how to evaluate how good they are. Most importantly, most smart people don’t want to work for them because they find them either too callous, unrewarding or slow (or a combination of all of these)….(More)”
Paul Ford at the New Republic: “I have a great fondness for government data, and the government has a great fondness for making more of it. Federal elections financial data, for example, with every contribution identified, connected to a name and address. Or the results of the census. I don’t know if you’ve ever had the experience of downloading census data but it’s pretty exciting. You can hold America on your hard drive! Meditate on the miracles of zip codes, the way the country is held together and addressable by arbitrary sets of digits.
You can download whole books, in PDF format, about the foreign policy of the Reagan Administration as it related to Russia. Negotiations over which door the Soviet ambassador would use to enter a building. Gigabytes and gigabytes of pure joy for the ephemeralist. The government is the greatest creator of ephemera ever.
Consider the Financial Crisis Inquiry Commission, or FCIC, created in 2009 to figure out exactly how the global economic pooch was screwed. The FCIC has made so much data, and has done an admirable job (caveats noted below) of arranging it. So much stuff. There are reams of treasure on a single FCIC web site, hosted at Stanford Law School: Hundreds of MP3 files, for example, with interviews with Jamie Dimonof JPMorgan Chase and Lloyd Blankfein of Goldman Sachs. I am desperate to find time to write some code that automatically extracts random audio snippets from each and puts them on top of a slow ambient drone with plenty of reverb, so that I can relax to the dulcet tones of the financial industry explaining away its failings. (There’s a Paul Krugman interview that I assume is more critical.)
The recordings are just the beginning. They’ve released so many documents, and with the documents, a finding aid that you can download in handy PDF format, which will tell you where to, well, find things, pointing to thousands of documents. That aid alone is 1,439 pages.
Look, it is excellent that this exists, in public, on the web. But it also presents a very contemporary problem: What is transparency in the age of massive database drops? The data is available, but locked in MP3s and PDFs and other documents; it’s not searchable in the way a web page is searchable, not easy to comment on or share.
Consider the WikiLeaks release of State Department cables. They were exhausting, there were so many of them, they were in all caps. Or the trove of data Edward Snowden gathered on aUSB drive, or Chelsea Manning on CD. And the Ashley Madison leak, spread across database files and logs of credit card receipts. The massive and sprawling Sony leak, complete with whole email inboxes. And with the just-released Panama Papers, we see two exciting new developments: First, the consortium of media organizations that managed the leak actually came together and collectively, well, branded the papers, down to a hashtag (#panamapapers), informational website, etc. Second, the size of the leak itself—2.5 terabytes!—become a talking point, even though that exact description of what was contained within those terabytes was harder to understand. This, said the consortia of journalists that notably did not include The New York Times, The Washington Post, etc., is the big one. Stay tuned. And we are. But the fact remains: These artifacts are not accessible to any but the most assiduous amateur conspiracist; they’re the domain of professionals with the time and money to deal with them. Who else could be bothered?
If you watched the movie Spotlight, you saw journalists at work, pawing through reams of documents, going through, essentially, phone books. I am an inveterate downloader of such things. I love what they represent. And I’m also comfortable with many-gigabyte corpora spread across web sites. I know how to fetch data, how to consolidate it, and how to search it. I share this skill set with many data journalists, and these capacities have, in some ways, become the sole province of the media. Organs of journalism are among the only remaining cultural institutions that can fund investigations of this size and tease the data apart, identifying linkages and thus constructing informational webs that can, with great effort, be turned into narratives, yielding something like what we call “a story” or “the truth.”
Spotlight was set around 2001, and it features a lot of people looking at things on paper. The problem has changed greatly since then: The data is everywhere. The media has been forced into a new cultural role, that of the arbiter of the giant and semi-legal database. ProPublica, a nonprofit that does a great deal of data gathering and data journalism and then shares its findings with other media outlets, is one example; it funded a project called DocumentCloud with other media organizations that simplifies the process of searching through giant piles of PDFs (e.g., court records, or the results of Freedom of Information Act requests).
At some level the sheer boredom and drudgery of managing these large data leaks make them immune to casual interest; even the Ashley Madison leak, which I downloaded, was basically an opaque pile of data and really quite boring unless you had some motive to poke around.
If this is the age of the citizen journalist, or at least the citizen opinion columnist, it’s also the age of the data journalist, with the news media acting as product managers of data leaks, making the information usable, browsable, attractive. There is an uneasy partnership between leakers and the media, just as there is an uneasy partnership between the press and the government, which would like some credit for its efforts, thank you very much, and wouldn’t mind if you gave it some points for transparency while you’re at it.
Pause for a second. There’s a glut of data, but most of it comes to us in ugly formats. What would happen if the things released in the interest of transparency were released in actual transparent formats?…(More)”
Steve Lohr at the New York Times: “For years, the federal government, states and some cities have enthusiastically made vast troves of data open to the public. Acres of paper records on demographics, public health, traffic patterns, energy consumption, family incomes and many other topics have been digitized and posted on the web.
This abundance of data can be a gold mine for discovery and insights, but finding the nuggets can be arduous, requiring special skills.
A project coming out of the M.I.T. Media Lab on Monday seeks to ease that challenge and to make the value of government data available to a wider audience. The project, called Data USA, bills itself as “the most comprehensive visualization of U.S. public data.” It is free, and its software code is open source, meaning that developers can build custom applications by adding other data.
Cesar A. Hidalgo, an assistant professor of media arts and sciences at the M.I.T. Media Lab who led the development of Data USA, said the website was devised to “transform data into stories.” Those stories are typically presented as graphics, charts and written summaries….Type “New York” into the Data USA search box, and a drop-down menu presents choices — the city, the metropolitan area, the state and other options. Select the city, and the page displays an aerial shot of Manhattan with three basic statistics: population (8.49 million), median household income ($52,996) and median age (35.8).
Lower on the page are six icons for related subject categories, including economy, demographics and education. If you click on demographics, one of the so-called data stories appears, based largely on data from the American Community Survey of the United States Census Bureau.
Using colorful graphics and short sentences, it shows the median age of foreign-born residents of New York (44.7) and of residents born in the United States (28.6); the most common countries of origin for immigrants (the Dominican Republic, China and Mexico); and the percentage of residents who are American citizens (82.8 percent, compared with a national average of 93 percent).
Adam Vaughan at The Guardian: They’ve been driven from Trafalgar square for being a nuisance, derided as rats with wings and maligned as a risk to public health.
But now pigeons could play a small part in helping Londoners overcome one of the capital’s biggest health problems – its illegal levels of air pollution blamed for thousands of deaths a year.
On Monday, a flock of half a dozen racing pigeons were set loose from a rooftop in Brick Lane by pigeon fancier, Brian Woodhouse, with one strapped with a pollution sensor to its back and one with a GPS tracker.
But while the 25g sensor records the nitrogen dioxide produced by the city’s diesel cars, buses, and trucks and tweets it at anyone who asks for a reading, its real purpose – and the use of the pigeons – is to raise awareness.
“It is a scandal. It is a health and environmental scandal for humans – and pigeons. We’re making the invisible visible,” said Pierre Duquesnoy, who won a London Design Festival award for the idea last year.
“Most of the time when we talk about pollution people think about Beijing or other places, but there are some days in the year when pollution was higher and more toxic in London than Beijing, that’s the reality.”
He said he was inspired by the use of pigeons in the first and second world wars to deliver information and save lives, but they were also a practical way of taking mobile air quality readings and beating London’s congested roads. They fly relatively low, at 100-150ft, and fast, at speeds up to 80mph.
“There’s something about taking what is seen as a flying rat and reversing that into something quite positive,” said Duquesnoy, who is creative director at marketing agency DigitasLBI.
Gary Fuller, an air quality expert at King’s College London, said it was the first time he had heard of urban animals being put to such use.
“It’s great that unemployed pigeons from Trafalgar Square are being put to work. Around 15 years ago tests were done on around 150 stray dogs in Mexico City, showing the ways in which air pollution was affecting lungs and heart health. But this is the first time that I’ve heard of urban wild animals being used to carry sensors to give us a picture of the air pollution over our heads.”
The release of the pigeons for three days this week, dubbed the Pigeon Air Patrol, came as moderate to high pollution affected much of the city, with Battersea recording ‘very high’, the top of the scale.
Elsewhere in the UK, Stockton-on-tees and Middlesbrough recorded high pollution readings and the forecast is for moderate and possibly high pollution in urban areas in northern England and Scotland on Tuesday. Other areas will have low pollution levels….(More).
Anne Frances Johnson in ThrivingEarthExchange: “In the shadow of a rumbling volcano, Quito, Ecuador solicits just-in-time advice from the world’s disaster experts…
Cotopaxi’s last large-scale eruption was in 1877, and the volcano’s level of activity suggests another one is inevitable. In addition to spewing lava, a major eruption would melt Cotopaxi’s glaciers and send a large flow of material barreling down the mountain, posing an immediate risk to people and potentially causing rivers to overflow their banks. Some 120,000 people living in the valley beneath the volcano would have a mere 12 minutes to escape the lava’s path, and more than 325,000 other area residents would have only slightly more time to evacuate. An eruption could also create significant long-term challenges across a broad area, including dangerous air quality and disruptions to infrastructure, food systems and water supplies.
As danger looms, a city gets coaching from the crowd
Aware that the city was underprepared for a significant eruption, The Governance Lab, a program of the New York University Tandon School of Engineering, volunteered its time and expertise to help local officials accelerate preparation efforts. The GovLab, which helps governments and other institutions work collaboratively to solve problems, teamed up with Linq, the city’s innovation agency.
“We were very aware that this was a time-sensitive matter—we needed experts, and we needed them fast,” explained Dinorah Cantú-Pedraza, a human rights lawyer and Research Fellow at The GovLab who collaborated on the project. “So that’s why we decided to create online sessions focused on how innovations can solve specific problems facing the city.”…
GovLab’s “fail-fast, learn-by-doing” approach is crucial to its projects’ success in remaining responsive to the problems at hand. “That was a central element in how we worked with our partners and improved the approach as we went forward,” said Cantú-Pedraza.
To help translate the Cotopaxi crowdsourcing model for other circumstances, GovLab is working to build a network of innovators and experts that can be tapped on short notice to address problems as they emerge around the world. Although we can hope for the best in Quito and elsewhere, the reality is that we must plan for the worst…(More)
But unlike many other global pandemics, the spread of Zika has been harder to identify, map and contain. It’s believed that 4 in 5 people with the virus don’t show any symptoms, and the primary transmitter for the disease, the Aedes mosquito species, is both widespread and challenging to eliminate. That means that fighting Zika requires raising awareness on how people can protect themselves, as well as supporting organizations who can help drive the development of rapid diagnostics and vaccines. We also have to find better ways to visualize the threat so that public health officials and NGO’s can support communities at risk….
A volunteer team of Google engineers, designers, and data scientists is helping UNICEF build a platform to process data from different sources (i.e., weather and travel patterns) in order to visualize potential outbreaks. Ultimately, the goal of this open source platform is to identify the risk of Zika transmission for different regions and help UNICEF, governments and NGO’s decide how and where to focus their time and resources. This set of tools is being prototyped for the Zika response, but will also be applicable to future emergencies….
We already include robust information for 900+ health conditions directly on Search for people in the U.S. We’ve now also added extensive information about Zika globally in 16 languages, with an overview of the virus, symptom information, and Public Health Alerts from that can be updated with new information as it becomes available.
We’re also working with popular YouTube creators across Latin America, including Sesame Street and Brazilian physician Drauzio Varella, to raise awareness about Zika prevention via their channels.
We hope these efforts are helpful in fighting this new public health emergency, and we will continue to do our part to help combat this outbreak.
And if you’re curious about what that 3,000 percent search increase looks like, take a look:….(More)
Tina Rosenberg at The New York Times: “Have you thought of a clever product to mitigate climate change? Did you invent an ingenious gadget to light African villages at night? Have you come up with a new kind of school, or new ideas for lowering the rate of urban shootings?
Thanks, but we have lots of those.
Whatever problem possesses you, we already have plenty of ways to solve it. Many have been rigorously tested and have a lot of evidence behind them — and yet they’re sitting on a shelf.
So don’t invent something new. If you want to make a contribution, choose one of those ideas — and spread it.
Spreading an idea can mean two different things. One is to take something that’s working in one place and introduce it somewhere else. If you want to reduce infant mortality in Cleveland, why not try what’s working in Baltimore?
Well, you might not know about what’s working because there’s no quick system for finding it.
Even when a few people do search out the answer, innovative ideas don’t spread by themselves. To become well known, they require effort from their originators. For example, a Bogotá, Colombia, maternity hospital invented Kangaroo Care — a method of keeping premature babies warm by strapping them 24/7 to Mom’s chest. It saved a lot of lives in Bogotá. But what allowed it to save lives around the world was a campaign to spread it to other countries.
The Colombians established Fundación Canguro and got grants from wealthy countries to bring groups of doctors and nurses from all over to visit Bogotá for two or three weeks. Once the visitors had gone back and set up a program in their hospital, the foundation loaned them a doctor and nurse to help get them started. Save the Children now leads a global partnership to spread Kangaroo Care, with the goal of reaching half the world.
In short, this work requires dedicated organizations, a smart program and lots of money.
The other meaning of spreading an idea is creating ways to get new inventions out to people who need them.
“When I talk to college students or anyone who’s thinking about entrepreneurship or targeting global poverty, the gadget is where 99 percent of people start thinking,” said Nicholas Fusso, the director of D-Prize (its slogan: “Distribution is development”). “That’s important — but the biggest problems in the poverty world aren’t a lack of gadgets or new products. It’s figuring out how people can have access to them.” So D-Prize gives seed money, in chunks of $10,000 to $20,000, to tiny new organizations that have good ideas for how to distribute useful things.
This analysis may be familiar to regular readers of Fixes. Indeed, the first Fixes column, more than five years ago, focused on distribution: getting health care to people in rural Africa by putting health care workers on motorcycles and keeping the bikes running….
Philanthropists and government aid agencies are only starting to get interested in the challenges of distribution — one new philanthropy that does have this focus is Good Ventures. As for academia, it still rewards invention almost exclusively. “There’s a lot of attention and award-giving and prize-giving and credit to people who come up with fancy new ideas instead of reaching people and having impact,” said Brodbar. “The incentives aren’t aligned. The culture of social entrepreneurship needs to change.”
Recognizing the true value of spreading an idea would also allow people who aren’t inventors (which is most of us) to get involved in social change. “The notion that if you want to engage in [social entrepreneurship] you have to have the big idea does a disservice to this space and people who want to play a role in it,” said Brodbar. “It’s a much wider front door.”…(More)