Mission Control: A History of the Urban Dashboard


Futuristic control rooms have proliferated in dozens of global cities. Baltimore has its CitiStat Room, where department heads stand at a podium before a wall of screens and account for their units’ performance.  The Mayor’s office in London’s City Hall features a 4×3 array of iPads mounted in a wooden panel, which seems an almost parodic, Terry Gilliam-esque take on the Brazilian Ops Center. Meanwhile, British Prime Minister David Cameron commissioned an iPad app – the “No. 10 Dashboard” (a reference to his residence at 10 Downing Street) – which gives him access to financial, housing, employment, and public opinion data. As The Guardian reported, “the prime minister said that he could run government remotely from his smartphone.”

This is the age of Dashboard Governance, heralded by gurus like Stephen Few, founder of the “visual business intelligence” and “sensemaking” consultancy Perceptual Edge, who defines the dashboard as a “visual display of the most important information needed to achieve one or more objectives; consolidated and arranged on a single screen so the information can be monitored at a glance.” A well-designed dashboard, he says — one that makes proper use of bullet graphs, sparklines, and other visualization techniques informed by the “brain science” of aesthetics and cognition — can afford its users not only a perceptual edge, but a performance edge, too. The ideal display offers a big-picture view of what is happening in real time, along with information on historical trends, so that users can divine the how and why and redirect future action. As David Nettleton emphasizes, the dashboard’s utility extends beyond monitoring “the current situation”; it also “allows a manager to … make provisions, and take appropriate actions.”….

The dashboard market now extends far beyond the corporate world. In 1994, New York City police commissioner William Bratton adapted former officer Jack Maple’s analog crime maps to create the CompStat model of aggregating and mapping crime statistics. Around the same time, the administrators of Charlotte, North Carolina, borrowed a business idea — Robert Kaplan’s and David Norton’s “total quality management” strategy known as the “Balanced Scorecard” — and began tracking performance in five “focus areas” defined by the City Council: housing and neighborhood development, community safety, transportation, economic development, and the environment. Atlanta followed Charlotte’s example in creating its own city dashboard.

In 1999, Baltimore mayor Martin O’Malley, confronting a crippling crime rate and high taxes, designed CitiStat, “an internal process of using metrics to create accountability within his government.” (This rhetoric of data-tested internal “accountability” is prevalent in early dashboard development efforts.) The project turned to face the public in 2003, when Baltimore launched a website of city operational statistics, which inspired DCStat (2005), Maryland’s StateStat (2007), and NYCStat (2008). Since then, myriad other states and metro areas — driven by a “new managerialist” approach to urban governance, committed to “benchmarking” their performance against other regions, and obligated to demonstrate compliance with sustainability agendas — have developed their own dashboards.

The Open Michigan Mi Dashboard is typical of these efforts. The state website presents data on education, health and wellness, infrastructure, “talent” (employment, innovation), public safety, energy and environment, financial health, and seniors. You (or “Mi”) can monitor the state’s performance through a side-by-side comparison of “prior” and “current” data, punctuated with a thumbs-up or thumbs-down icon indicating the state’s “progress” on each metric. Another click reveals a graph of annual trends and a citation for the data source, but little detail about how the data are actually derived. How the public is supposed to use this information is an open question….(More)”

31 cities agree to use EU-funded open innovation platform for better smart cities’ services


European Commission Press Release: “At CEBIT, 25 cities from 6 EU countries (Belgium, Denmark, Finland, Italy, Portugal and Spain) and 6 cities from Brazil will present Open & Agile Smart Cities Task Force (OASC), an initiative making it easier for city councils  and startups to improve smart city services (such as transport, energy efficiency, environmental or e-health services). This will be achieved thanks to FIWARE, an EU-funded, open source platform and cloud-based building blocks developed in the EU that can be used to develop a huge range of applications, from Smart Cities to eHealth, and from transport to disaster management. Many applications have already been built using FIWARE – from warnings of earthquakes to preventing food waste to Smartaxi apps. Find a full list of cities in the Background.

The OASC deal will allow cities to share their open data (collected from sensors measuring, for example, traffic flows) so that startups can develop apps and tools that benefit all citizens (for example, an app with traffic information for people on the move). Moreover, these systems will be shared between cities (so, an app with transport information developed in city A can be also adopted by city B, without the latter having to develop it from scratch); FIWARE will also give startups and app developers in these cities access to a global market for smart city services.

Cities from across the globe are trying to make the most of open innovation. This will allow them to include a variety of stakeholders in their activities (services are increasingly connected to other systems and innovative startups are a big part of this trend) and encourage a competitive yet attractive market for developers, thus reducing costs, increasing quality and avoiding vendor lock-in….(More)”

Index: Prizes and Challenges


The Living Library Index – inspired by the Harper’s Index – provides important statistics and highlights global trends in governance innovation. This installment focuses on prizes and challenges and was originally published in 2015.

This index highlights recent findings about two key techniques in shifting innovation from institutions to the general public:

  • Prize-Induced Contests – using monetary rewards to incentivize individuals and other entities to develop solutions to public problems; and
  • Grand Challenges – posing large, audacious goals to the public to spur collaborative, non-governmental efforts to solve them.

You can read more about Governing through Prizes and Challenges here. You can also watch Alph Bingham, co-founder of Innocentive, answer the GovLab’s questions about challenge authoring and defining the problem here.

Previous installments of the Index include Measuring Impact with Evidence, The Data Universe, Participation and Civic Engagement and Trust in Institutions. Please share any additional statistics and research findings on the intersection of technology in governance with us by emailing shruti at thegovlab.org.

Prize-Induced Contests

  • Year the British Government introduced the Longitude Prize, one of the first instances of prizes by government to spur innovation: 1714
  • President Obama calls on “all agencies to increase their use of prizes to address some of our Nation’s most pressing challenges” in his Strategy for American Innovation: September 2009
  • The US Office of Management and Budget issues “a policy framework to guide agencies in using prizes to mobilize American ingenuity and advance their respective core missions”:  March 2010
  • Launch of Challenge.gov, “a one-stop shop where entrepreneurs and citizen solvers can find public-sector prize competitions”: September 2010
    • Number of competitions currently live on Challenge.gov in February 2015: 22 of 399 total
    • How many competitions on Challenge.gov are for $1 million or above: 23
  • The America COMPETES Reauthorization Act is introduced, which grants “all Federal agencies authority to conduct prize competitions to spur innovation, solve tough problems, and advance their core missions”: 2010
  • Value of prizes authorized by COMPETES: prizes up to $50 million
  • Fact Sheet and Frequently Asked Questions memorandum issued by the Office of Science and Technology Policy and the Office of Management and Budget to aid agencies to take advantage of authorities in COMPETES: August 2011
  • Number of prize competitions run by the Federal government from 2010 to April 2012: 150
  • How many Federal agencies have run prize competitions by 2012: 40
  • Prior to 1991, the percentage of prize money that recognized prior achievements according to an analysis by McKinsey and Company: 97%
    • Since 1991, percentage of new prize money that “has been dedicated to inducement-style prizes that focus on achieving a specific, future goal”: 78%
  • Value of the prize sector as estimated by McKinsey in 2009: $1-2 billion
  • Growth rate of the total value of new prizes: 18% annually
  • Growth rate in charitable giving in the US: 2.5% annually
  • Value of the first Horizon Prize awarded in 2014 by the European Commission to German biopharmaceutical company CureVac GmbH “for progress towards a novel technology to bring life-saving vaccines to people across the planet in safe and affordable ways”: €2 million
  • Number of solvers registered on InnoCentive, a crowdsourcing company: 355,000+ from nearly 200 countries
    • Total Challenges Posted: 2,000+ External Challenges
    • Total Solution Submissions: 40,000+
    • Value of the awards: $5,000 to $1+ million
    • Success Rate for premium challenges: 85%

Grand Challenges

  • Value of the Progressive Insurance Automotive X Prize, sponsored in part by DOE to develop production-capable super fuel-efficient vehicles: $10 million
    • Number of teams around the world who took part in the challenge “to develop a new generation of technologies” for production-capable super fuel-efficient vehicles: 111 teams
  • Time it took for the Air Force Research Laboratory to receive a workable solution on “a problem that had vexed military security forces and civilian police for years” by opening the challenge to the world: 60 days
  • Value of the HHS Investing in Innovation initiative to spur innovation in Health IT, launched under the new COMPETES act: $5 million program
  • Number of responses received by NASA for its Asteroid Grand Challenge RFI which seeks to identify and address all asteroid threats to the human population: over 400
  • The decreased cost of sequencing a single human genome as a result of the Human Genome Project Grand Challenge: $7000 from $100 million
  • Amount the Human Genome Project Grand Challenge has contributed to the US economy for every $1 invested by the US federal government: $141 for every $1 invested
  • The amount of funding for research available for the “Brain Initiative,” a collaboration between the National Institute of Health, DARPA and the National Science Foundation, which seeks to uncover new prevention and treatment methods for brain disorders like Alzheimer’s, autism and schizophrenia: $100 million
  • Total amount offered in cash awards by the Department of Energy’s “SunShot Grand Challenge,” which seeks to eliminate the cost disparity between solar energy and coal by the end of the decade: $10 million

Sources

The Future of Public Space Analytics


The AGILE Landscape Project: “Public space is an essential component of any great city. It brings people together to socialize, recreate, and work. More pointedly, it attracts people to the city, builds relationships, and spurs innovation and new ideas that fuel a city’s economic growth. How we optimize the investments made in our streetscapes, plazas, parks, and greenways is important to each individual project’s success and the city as a whole. Are the places being built fulfilling their promise? If so, could they be doing more and if not, how do they need to change?…
“Places in the Making”, a publication published in 2013 by MIT’s Department of Urban Studies and Planning, underscored this problem. They cited the lack of active metrics as a common challenge to placemaking efforts. They reported that “It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” They went on to explain the existing placemaking culture focuses on fuzzy, unmeasurable goals and that this has created “an inertia in assessment efforts”. The report stressed that this lack of assessment is detrimental to the field as a whole because valuable insights are left undiscovered and the same mistakes are made repeatedly….
“Places in the Making”, a publication published in 2013 by MIT’s Department of Urban Studies and Planning, underscored this problem. They cited the lack of active metrics as a common challenge to placemaking efforts. They reported that “It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” They went on to explain the existing placemaking culture focuses on fuzzy, unmeasurable goals and that this has created “an inertia in assessment efforts”. The report stressed that this lack of assessment is detrimental to the field as a whole because valuable insights are left undiscovered and the same mistakes are made repeatedly.

“It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” – “Places in the Making”, DUSP (2013)

Both William H. Whyte’s seminal Street Life Project and resulting book and documentary, Social Life of Small Urban Spaces (1980), along with Jan Gehl’s extensive public life studies and influential book Life Between Buildings (1987) serve as the foundation of the urban design profession’s knowledge base. Their observations and insights were amassed through considerable time and manpower. We are lucky to have their contribution from which to draw upon when making future decisions. Though we shouldn’t stop observing and building upon their significant contributions. We must continually add clarity and resolution particularly in the unique and specific context of each place.
social media metrics
Unfortunately, the POE and previous research are only snapshots in time. With the rise of embedded sensor infrastructure, “smart cities”, social networks, and ubiquity of smartphones that are becoming increasingly more powerful and sophisticated every year, the ability to gain real-time continuous 24/7 quantitative monitoring of public life data is now a reality. The ability to capture so much data is unprecedented and will only continue to get easier. The amount of and type of information that can and often is being collected about the world around us and ourselves is vast. When it comes to cities, transportation, parking, energy, waste collection, and water systems has attracted the most attention. In the private sector, retail analytics have been the focus.
Collecting data for data sake should not be a goal. Rather we need be knowledgeable of what is available and possible while being discerning about the data that is useful for improving the management and design of public space. By selectively incorporating sensors into the design of physical space, the data collection system can provide real-time information about the use of the space. Like website analytic systems, the network can provide useful information about the number of people using the space, how long they are using it, as well well as other meaningful measures of public life. The data gathered can be combined with a wide array of environmental data such as temperature, humidity, sun/shade patterns, rain fall, water quality, and air quality. In addition, other available data sets such as sales tax collections, building permits, real estate transactions, and programming/event schedules can be overlaid with data to make relationships between the use of the public space and its context more visible.
Array of information that can help us understand public space
While the real-time nature of the data can be used to quickly understand the immediate impact that programming and design decisions have on the space and vice versa, it can also capture data and trends over extended durations that would not otherwise be possible without considerable resources. All of this enables us to make more informed decisions rather than anecdotal observations and assumptions. With the interest of tactical urbanism and other short-term public space investments, this need for data is greater than ever. The data can be used to justify investment in long-term improvements and/or make slight adjustments in temporary spaces to increase their effectiveness….(More)”

Choosing Not to Choose: Understanding the Value of Choice


New book by Cass Sunstein: “Our ability to make choices is fundamental to our sense of ourselves as human beings, and essential to the political values of freedom-protecting nations. Whom we love; where we work; how we spend our time; what we buy; such choices define us in the eyes of ourselves and others, and much blood and ink has been spilt to establish and protect our rights to make them freely.
Choice can also be a burden. Our cognitive capacity to research and make the best decisions is limited, so every active choice comes at a cost. In modern life the requirement to make active choices can often be overwhelming. So, across broad areas of our lives, from health plans to energy suppliers, many of us choose not to choose. By following our default options, we save ourselves the costs of making active choices. By setting those options, governments and corporations dictate the outcomes for when we decide by default. This is among the most significant ways in which they effect social change, yet we are just beginning to understand the power and impact of default rules. Many central questions remain unanswered: When should governments set such defaults, and when should they insist on active choices? How should such defaults be made? What makes some defaults successful while others fail?….
The onset of big data gives corporations and governments the power to make ever more sophisticated decisions on our behalf, defaulting us to buy the goods we predictably want, or vote for the parties and policies we predictably support. As consumers we are starting to embrace the benefits this can bring. But should we? What will be the long-term effects of limiting our active choices on our agency? And can such personalized defaults be imported from the marketplace to politics and the law? Confronting the challenging future of data-driven decision-making, Sunstein presents a manifesto for how personalized defaults should be used to enhance, rather than restrict, our freedom and well-being. (More)”

The Internet’s hidden science factory


Jenny Marder at PBS Newshour: “….Marshall is a worker for Amazon’s Mechanical Turk, an online job forum where “requesters” post jobs, and an army of crowdsourced workers complete them, earning fantastically small fees for each task. The work has been called microlabor, and the jobs, known as Human Intelligence Tasks, or HITs, range wildly. Some are tedious: transcribing interviews or cropping photos. Some are funny: prank calling someone’s buddy (that’s worth $1) or writing the title to a pornographic movie based on a collection of dirty screen grabs (6 cents). And others are downright bizarre. One task, for example, asked workers to strap live fish to their chests and upload the photos. That paid $5 — a lot by Mechanical Turk standards….
These aren’t obscure studies that Turkers are feeding. They span dozens of fields of research, including social, cognitive and clinical psychology, economics, political science and medicine. They teach us about human behavior. They deal in subjects like energy conservation, adolescent alcohol use, managing money and developing effective teaching methods.


….In 2010, the researcher Joseph Henrich and his team published a paper showing that an American undergraduate was about 4,000 times more likely than an average American to be the subject of a research study.
But that output pales in comparison to Mechanical Turk workers. The typical “Turker” completes more studies in a week than the typical undergraduate completes in a lifetime. That’s according to research by Rand, who surveyed both groups. Among those he surveyed, he found that the median traditional lab subject had completed 15 total academic studies — an average of one per week. The median Turker, on the other hand, had completed 300 total academic studies — an average of 20 per week….(More)”

Can Selfies Save Nutrition Science?


Trevor Butterworth at Stats.org: “You may have never heard of the Energy Balance Working Group, but this collection of 45 experts on nutrition, exercise, biochemistry, and other related disciplines have collectively thrown a “House-like” wrench into the research literature on everything from obesity to cancer and heart disease. Gregory House, the fictional and fantastically brilliant physician played by Hugh Laurie in the eponymous TV show frequently found his patients wanting in the court of self-reported truth: “I’ve found that when you want to know the truth about someone that someone is probably the last person you should ask.”
This is more or less what the Energy Balance Working Group have concluded in an “expert report” recently published in the International Journal of Obesity. If you want to know the truth about how much someone eats and exercises that someone is probably the last person you should ask….The problem is that self-reporting is a cheap and convenient source of data for research, while more accurate alternatives are either expensive and challenging or, as yet, more promise than reality (see sidebar)….
“There are at least two categories of solutions on the horizon. In one category, there are wearable monitoring devices that can collect objective, real-time data. Examples in the works or in use include photographic food diaries, records of chewing and swallowing behavior, and evaluating the time and intensity of movement using accelerometers and GPS, among others. It is important to note that there are still challenges converting these measurements into reliable estimates of energy intake and expenditure, but work is ongoing… David Allison, Distinguished Professor, Quetelet Endowed Professor of Public Health, University of Alabam”…(More)

Untangling Complexity


Essay by Jacqueline Wallace at civicquarterly.com : “The next phase of the digital revolution will be defined by products and services that facilitate shared understanding, allowing concerted participation around complex issues. In working to show the way, civic designers will need to call upon the powers of systems research, design research, social science, and open data….
Creating next-gen civic applications will require designers to embody a systems-based approach to civic participation, marrying systems-based research, user-centered design, social science, and data. This article chronicles my own experience leveraging these tools to facilitate shared understand amongst my community vis. a vis. the Kinder Morgan pipeline….
I believe the contention around the pipeline evinces a bigger problem in our civic sphere: While individual issues such as the Kinder Morgan pipeline continue to absorb a great deal of energy from citizens, user-centered designers must use their unique skillset to address these issues more broadly. Because we’re not only failing our fellow citizens; we’re failing our representatives as well. In the words of digital strategist and civic design advocate Mike Connery: “there has been almost zero investment in giving our representatives the tools they need to understand feedback from citizens.”
The opportunity is two-fold. As previously stated, there is a need to develop tools that support citizens and representatives to understand and engage with complex social issues. There is also a need to develop processes, processes that cultivate our abilities to understand policy development in order to more efficiently spend our tax dollars. Writing about a recent project that used social science methods to analyze the long-term success of a welfare project, NPR correspondent Shankar Vedantam said, “it really makes no sense that marketers selling toys have better data on what works and what doesn’t than policy makers who are spending billions and billions of dollars.”…(More)

Innovate or Stagnate


Mohammed bin Rashid Al Maktoum at Project Syndicate: “Companies, like people, grow old. They start life small and eager to survive, fueled by youthful energy and fresh ideas. They compete, expand, mature, and eventually, with few exceptions, fade into obscurity. The same is true of governments: they, too, can lose the hunger and ambition of youth and allow themselves to become complacent….

The key to corporations’ rejuvenation, civilizations’ evolution, and human development in general is simple: innovation. I am always amazed when governments think they are an exception to this rule. Innovation in government is not an intellectual luxury, a topic confined to seminars and panel discussions, or a matter only of administrative reforms. It is the recipe for human survival and development, the fuel for constant progress, and the blueprint for a country’s rise.

The first key to business-like innovation in government is a focus on skills. Top-tier companies continuously invest in their employees to provide them with the right skills for the marketplace. Governments must do the same, by constantly upgrading skills and nurturing innovation – among their own employees, across key sectors of the economy, and at the foundations of the education system. Governments that fail to equip new generations with the skills needed to become leaders for their time are condemning them to be led by other, more innovative societies….

The second key to transforming governments into engines of innovation is to shift the balance of investment toward intangibles, as in the private sector. Whereas more than 80% of the value of the Standard & Poor’s 500 consisted of tangible assets 40 years ago, today that ratio is reversed: more than 80% of the largest companies’ value is intangible – the knowledge and skills of their employees and the intellectual property embedded in their products.

Governments, too, should think strategically about shifting their spending away from tangible infrastructure like roads and buildings, and toward intangibles like education and research and development…. (More)”.

Mapping the Nation: Building a More Resilient Future


New book from Esri: “The fifth book in Esri’s Mapping the Nation series, Mapping the Nation: Building a More Resilient Future is a collection of geographic information system (GIS) maps that illustrate how federal government agencies rely on GIS analysis to build stronger, more resilient communities and help make the world a better place.
The print version of the book includes 118 full-color maps produced by more than 50 federal government agencies, including the US Forest Service, US Department of Defense, US Department of Education, and the Bureau of Ocean Energy Management. The digital version of Mapping the Nation offers enhanced and interactive maps and videos showcasing four start-up companies that are using ArcGIS technology in partnership with Esri and the government.
The maps depict how federal employees and officials use GIS to evaluate, plan, and respond to social, economic, and environmental concerns at local, regional, national, and global levels. Topics such as green government, economic recovery and sustainability, and climate protection show how government agencies use GIS to facilitate initiatives, improve transparency, and deliver strong business models…
Mapping and Apping the Nation 2015, an interactive digital adaptation of the printed map book, is available free of charge from the Esri Books app on Apple iTunes and the Google Play store.”