Five Headlines from a Big Month for the Data Revolution


Sarah T. Lucas at Post2015.org: “If the history of the data revolution were written today, it would include three major dates. May 2013, when theHigh Level Panel on the Post-2015 Development Agenda first coined the phrase “data revolution.” November 2014, when the UN Secretary-General’s Independent Expert Advisory Group (IEAG) set a vision for it. And April 2015, when five headliner stories pushed the data revolution from great idea to a concrete roadmap for action.

The April 2015 Data Revolution Headlines

1. The African Data Consensus puts Africa in the lead on bringing the data revolution to the regional level. TheAfrica Data Consensus (ADC) envisions “a profound shift in the way that data is harnessed to impact on development decision-making, with a particular emphasis on building a culture of usage.” The ADC finds consensus across 15 “data communities”—ranging from open data to official statistics to geospatial data, and is endorsed by Africa’s ministers of finance. The ADC gets top billing in my book, as the first contribution that truly reflects a large diversity of voices and creates a political hook for action. (Stay tuned for a blog from my colleague Rachel Quint on the ADC).

2. The Sustainable Development Solutions Network (SDSN) gets our minds (and wallets) around the data needed to measure the SDGs. The SDSN Needs Assessment for SDG Monitoring and Statistical Capacity Development maps the investments needed to improve official statistics. My favorite parts are the clear typology of data (see pg. 12), and that the authors are very open about the methods, assumptions, and leaps of faith they had to take in the costing exercise. They also start an important discussion about how advances in information and communications technology, satellite imagery, and other new technologies have the potential to expand coverage, increase analytic capacity, and reduce the cost of data systems.

3. The Overseas Development Institute (ODI) calls on us to find the “missing millions.” ODI’s The Data Revolution: Finding the Missing Millions presents the stark reality of data gaps and what they mean for understanding and addressing development challenges. The authors highlight that even that most fundamental of measures—of poverty levels—could be understated by as much as a quarter. And that’s just the beginning. The report also pushes us to think beyond the costs of data, and focus on how much good data can save. With examples of data lowering the cost of doing government business, the authors remind us to think about data as an investment with real economic and social returns.

4. Paris21 offers a roadmap for putting national statistic offices (NSOs) at the heart of the data revolution.Paris21’s Roadmap for a Country-Led Data Revolution does not mince words. It calls on the data revolution to “turn a vicious cycle of [NSO] underperformance and inadequate resources into a virtuous one where increased demand leads to improved performance and an increase in resources and capacity.” It makes the case for why NSOs are central and need more support, while also pushing them to modernize, innovate, and open up. The roadmap gets my vote for best design. This ain’t your grandfather’s statistics report!

5. The Cartagena Data Festival features real-live data heroes and fosters new partnerships. The Festival featured data innovators (such as terra-i using satellite data to track deforestation), NSOs on the leading edge of modernization and reform (such as Colombia and the Philippines), traditional actors using old data in new ways (such as the Inter-American Development Bank’s fantastic energy database), groups focused on citizen-generated data (such as The Data Shift and UN My World), private firms working with big data for social good (such asTelefónica), and many others—all reminding us that the data revolution is well underway and will not be stopped. Most importantly, it brought these actors together in one place. You could see the sparks flying as folks learned from each other and hatched plans together. The Festival gets my vote for best conference of a lifetime, with the perfect blend of substantive sessions, intense debate, learning, inspiration, new connections, and a lot of fun. (Stay tuned for a post from my colleague Kristen Stelljes and me for more on Cartagena).

This month full of headlines leaves no room for doubt—momentum is building fast on the data revolution. And just in time.

With the Financing for Development (FFD) conference in Addis Ababa in July, the agreement of Sustainable Development Goals in New York in September, and the Climate Summit in Paris in December, this is a big political year for global development. Data revolutionaries must seize this moment to push past vision, past roadmaps, to actual action and results…..(More)”

Monithon


“Moni-thon” comes from “monitor” and “marathon”, and this is precisely what this platform seeks to help with: anintensive activity of observing and reporting of public policies in Italy.

What’s there to monitor?  Monithon was born as an independently developed initiative to promote the citizen monitoring of development projects funded both by the Italian government and the EU through the Cohesion (aka. Regional) Policy. Projects include a wide range of interventions such as large transport, digital, research or environmental infrastructures (railroads, highways, broadband networks, waste management systems…), aids to enterprises to support innovation and competitiveness, and other funding for energy efficiency, social inclusion, education and training, occupation and workers mobility, tourism, etc.

Citizen monitoring of these projects is possible thanks to a combination of open government data and citizens’ collaboration, joined by the goal of controlling how the projects are progressing, and whether they deliver actual results.

The Italian government releases the information on all the 800k+ projects funded (worth almost 100 billion Euros), the beneficiaries of the subsidies and all the actors involved as open data, including the location and the timing of the intervention. All the data is integrated with interactive visualizations on the national portal OpenCoesione, where people can play with the data and find the most interesting projects to follow.

The Monithon initiative takes this transparency further: it asks citizens to actively engage with open government data and to produce valuable information through it.

How does it work? Monithon means active involvement of communities and a shared methodology. Citizens, journalist, experts, researchers, students – or all combined – collect information on a specific project chosen from the OpenCoesione database. Then this information can be uploaded on the Monithon platform (based on Ushahidi) by selecting the projects from a list and it can be geo-referenced and enriched with interviews, quantitative data, pictures, videos. The result is a form of civic, bottom-down, collective data storytelling. All the “wannabe monithoners” can download this simple toolkit, a 10-page document that describes the initiative and explains how to pick a project to monitor and get things started.  ….

How to achieve actual impact? The Monithon platform is method and a model whereby citizen monitoring may be initiated and a tool for civic partners to press forward, to report on malpractice, but also to collaborate in making all these projects work, in accelerating their completion and understanding whether they actually respond to local demand. ….

Monithon has rapidly evolved from being an innovative new platform into a transferable civic engagement format.  Since its launch in September 2013, Monithon has drawn dozens of national and local communities (some formed on purpose, other based on existing associations) and around 500 people into civic monitoring activities, mostly in Southern Italy, where cohesion funds are more concentrated. Specific activities are carried out by established citizen groups, like Libera, a national anti-Mafia association, which became Monithon partner, focusing their monitoring on the rehabilitation of Mafia-seized properties. Action Aid is now partnering with Monithon to promote citizen empowerment. Existing, local groups of activists are using the Monithon methodology to test local transportation systems that benefited from EU funding, while new groups have formed to begin monitoring social innovation and cultural heritage projects.

Now more than 50 “citizen monitoring reports”, which take the form of collective investigations on project development and results, are publicly available on the Monithon website, many of which spurred further dialogue with public administrations….(More)

Growing Data Collection Inspires Openness at NGA


at Secrecy News: “A flood of information from the ongoing proliferation of space-based sensors and ground-based data collection devices is promoting a new era of transparency in at least one corner of the U.S. intelligence community.

The “explosion” of geospatial information “makes geospatial intelligence increasingly transparent because of the huge number and diversity of commercial and open sources of information,” said Robert Cardillo, director of the National Geospatial-Intelligence Agency (NGA), in a speech last month.

Hundreds of small satellites are expected to be launched within the next three years — what Mr. Cardillo called a “darkening of the skies” — and they will provide continuous, commercially available coverage of the entire Earth’s surface.

“The challenges of taking advantage of all of that data are daunting for all of us,” Mr. Cardillo said.

Meanwhile, the emerging “Internet of Things” is “spreading rapidly as more people carry more handheld devices to more places” generating an abundance of geolocation data.

This is, of course, a matter of intelligence interest since “Every local, regional, and global challenge — violent extremism in the Middle East and Africa, Russian aggression, the rise of China, Iranian and North Korean nuclear weapons, cyber security, energy resources, and many more — has geolocation at its heart.”

Consequently, “We must open up GEOINT far more toward the unclassified world,” Director Cardillo said in another speech last week.

“In the past, we have excelled in our closed system. We enjoyed a monopoly on sources and methods. That monopoly has long since ended. Today and in the future, we must thrive and excel in the open.”

So far, NGA has already distinguished itself in the area of disaster relief, Mr. Cardillo said.

“Consider Team NGA’s response to the Ebola crisis. We are the first intelligence agency to create a World Wide Web site with access to our relevant unclassified content. It is open to everyone — no passwords, no closed groups.”

NGA provided “more than a terabyte of up-to-date commercial imagery.”

“You can imagine how important it is for the Liberian government to have accurate maps of the areas hardest hit by the Ebola epidemic as well as the medical and transportation infrastructure to combat the disease,” Mr. Cardillo said.

But there are caveats. Just because information is unclassified does not mean that it is freely available.

“Although 99 percent of all of our Ebola data is unclassified, most of that is restricted by our agreements [with commercial providers],” Mr. Cardillo said. “We are negotiating with many sources to release more data.”

Last week, Director Cardillo announced a new project called GEOINT Pathfinder that will attempt “to answer key intelligence questions using only unclassified data.”….(More)

Mission Control: A History of the Urban Dashboard


Futuristic control rooms have proliferated in dozens of global cities. Baltimore has its CitiStat Room, where department heads stand at a podium before a wall of screens and account for their units’ performance.  The Mayor’s office in London’s City Hall features a 4×3 array of iPads mounted in a wooden panel, which seems an almost parodic, Terry Gilliam-esque take on the Brazilian Ops Center. Meanwhile, British Prime Minister David Cameron commissioned an iPad app – the “No. 10 Dashboard” (a reference to his residence at 10 Downing Street) – which gives him access to financial, housing, employment, and public opinion data. As The Guardian reported, “the prime minister said that he could run government remotely from his smartphone.”

This is the age of Dashboard Governance, heralded by gurus like Stephen Few, founder of the “visual business intelligence” and “sensemaking” consultancy Perceptual Edge, who defines the dashboard as a “visual display of the most important information needed to achieve one or more objectives; consolidated and arranged on a single screen so the information can be monitored at a glance.” A well-designed dashboard, he says — one that makes proper use of bullet graphs, sparklines, and other visualization techniques informed by the “brain science” of aesthetics and cognition — can afford its users not only a perceptual edge, but a performance edge, too. The ideal display offers a big-picture view of what is happening in real time, along with information on historical trends, so that users can divine the how and why and redirect future action. As David Nettleton emphasizes, the dashboard’s utility extends beyond monitoring “the current situation”; it also “allows a manager to … make provisions, and take appropriate actions.”….

The dashboard market now extends far beyond the corporate world. In 1994, New York City police commissioner William Bratton adapted former officer Jack Maple’s analog crime maps to create the CompStat model of aggregating and mapping crime statistics. Around the same time, the administrators of Charlotte, North Carolina, borrowed a business idea — Robert Kaplan’s and David Norton’s “total quality management” strategy known as the “Balanced Scorecard” — and began tracking performance in five “focus areas” defined by the City Council: housing and neighborhood development, community safety, transportation, economic development, and the environment. Atlanta followed Charlotte’s example in creating its own city dashboard.

In 1999, Baltimore mayor Martin O’Malley, confronting a crippling crime rate and high taxes, designed CitiStat, “an internal process of using metrics to create accountability within his government.” (This rhetoric of data-tested internal “accountability” is prevalent in early dashboard development efforts.) The project turned to face the public in 2003, when Baltimore launched a website of city operational statistics, which inspired DCStat (2005), Maryland’s StateStat (2007), and NYCStat (2008). Since then, myriad other states and metro areas — driven by a “new managerialist” approach to urban governance, committed to “benchmarking” their performance against other regions, and obligated to demonstrate compliance with sustainability agendas — have developed their own dashboards.

The Open Michigan Mi Dashboard is typical of these efforts. The state website presents data on education, health and wellness, infrastructure, “talent” (employment, innovation), public safety, energy and environment, financial health, and seniors. You (or “Mi”) can monitor the state’s performance through a side-by-side comparison of “prior” and “current” data, punctuated with a thumbs-up or thumbs-down icon indicating the state’s “progress” on each metric. Another click reveals a graph of annual trends and a citation for the data source, but little detail about how the data are actually derived. How the public is supposed to use this information is an open question….(More)”

31 cities agree to use EU-funded open innovation platform for better smart cities’ services


European Commission Press Release: “At CEBIT, 25 cities from 6 EU countries (Belgium, Denmark, Finland, Italy, Portugal and Spain) and 6 cities from Brazil will present Open & Agile Smart Cities Task Force (OASC), an initiative making it easier for city councils  and startups to improve smart city services (such as transport, energy efficiency, environmental or e-health services). This will be achieved thanks to FIWARE, an EU-funded, open source platform and cloud-based building blocks developed in the EU that can be used to develop a huge range of applications, from Smart Cities to eHealth, and from transport to disaster management. Many applications have already been built using FIWARE – from warnings of earthquakes to preventing food waste to Smartaxi apps. Find a full list of cities in the Background.

The OASC deal will allow cities to share their open data (collected from sensors measuring, for example, traffic flows) so that startups can develop apps and tools that benefit all citizens (for example, an app with traffic information for people on the move). Moreover, these systems will be shared between cities (so, an app with transport information developed in city A can be also adopted by city B, without the latter having to develop it from scratch); FIWARE will also give startups and app developers in these cities access to a global market for smart city services.

Cities from across the globe are trying to make the most of open innovation. This will allow them to include a variety of stakeholders in their activities (services are increasingly connected to other systems and innovative startups are a big part of this trend) and encourage a competitive yet attractive market for developers, thus reducing costs, increasing quality and avoiding vendor lock-in….(More)”

Index: Prizes and Challenges


The Living Library Index – inspired by the Harper’s Index – provides important statistics and highlights global trends in governance innovation. This installment focuses on prizes and challenges and was originally published in 2015.

This index highlights recent findings about two key techniques in shifting innovation from institutions to the general public:

  • Prize-Induced Contests – using monetary rewards to incentivize individuals and other entities to develop solutions to public problems; and
  • Grand Challenges – posing large, audacious goals to the public to spur collaborative, non-governmental efforts to solve them.

You can read more about Governing through Prizes and Challenges here. You can also watch Alph Bingham, co-founder of Innocentive, answer the GovLab’s questions about challenge authoring and defining the problem here.

Previous installments of the Index include Measuring Impact with Evidence, The Data Universe, Participation and Civic Engagement and Trust in Institutions. Please share any additional statistics and research findings on the intersection of technology in governance with us by emailing shruti at thegovlab.org.

Prize-Induced Contests

  • Year the British Government introduced the Longitude Prize, one of the first instances of prizes by government to spur innovation: 1714
  • President Obama calls on “all agencies to increase their use of prizes to address some of our Nation’s most pressing challenges” in his Strategy for American Innovation: September 2009
  • The US Office of Management and Budget issues “a policy framework to guide agencies in using prizes to mobilize American ingenuity and advance their respective core missions”:  March 2010
  • Launch of Challenge.gov, “a one-stop shop where entrepreneurs and citizen solvers can find public-sector prize competitions”: September 2010
    • Number of competitions currently live on Challenge.gov in February 2015: 22 of 399 total
    • How many competitions on Challenge.gov are for $1 million or above: 23
  • The America COMPETES Reauthorization Act is introduced, which grants “all Federal agencies authority to conduct prize competitions to spur innovation, solve tough problems, and advance their core missions”: 2010
  • Value of prizes authorized by COMPETES: prizes up to $50 million
  • Fact Sheet and Frequently Asked Questions memorandum issued by the Office of Science and Technology Policy and the Office of Management and Budget to aid agencies to take advantage of authorities in COMPETES: August 2011
  • Number of prize competitions run by the Federal government from 2010 to April 2012: 150
  • How many Federal agencies have run prize competitions by 2012: 40
  • Prior to 1991, the percentage of prize money that recognized prior achievements according to an analysis by McKinsey and Company: 97%
    • Since 1991, percentage of new prize money that “has been dedicated to inducement-style prizes that focus on achieving a specific, future goal”: 78%
  • Value of the prize sector as estimated by McKinsey in 2009: $1-2 billion
  • Growth rate of the total value of new prizes: 18% annually
  • Growth rate in charitable giving in the US: 2.5% annually
  • Value of the first Horizon Prize awarded in 2014 by the European Commission to German biopharmaceutical company CureVac GmbH “for progress towards a novel technology to bring life-saving vaccines to people across the planet in safe and affordable ways”: €2 million
  • Number of solvers registered on InnoCentive, a crowdsourcing company: 355,000+ from nearly 200 countries
    • Total Challenges Posted: 2,000+ External Challenges
    • Total Solution Submissions: 40,000+
    • Value of the awards: $5,000 to $1+ million
    • Success Rate for premium challenges: 85%

Grand Challenges

  • Value of the Progressive Insurance Automotive X Prize, sponsored in part by DOE to develop production-capable super fuel-efficient vehicles: $10 million
    • Number of teams around the world who took part in the challenge “to develop a new generation of technologies” for production-capable super fuel-efficient vehicles: 111 teams
  • Time it took for the Air Force Research Laboratory to receive a workable solution on “a problem that had vexed military security forces and civilian police for years” by opening the challenge to the world: 60 days
  • Value of the HHS Investing in Innovation initiative to spur innovation in Health IT, launched under the new COMPETES act: $5 million program
  • Number of responses received by NASA for its Asteroid Grand Challenge RFI which seeks to identify and address all asteroid threats to the human population: over 400
  • The decreased cost of sequencing a single human genome as a result of the Human Genome Project Grand Challenge: $7000 from $100 million
  • Amount the Human Genome Project Grand Challenge has contributed to the US economy for every $1 invested by the US federal government: $141 for every $1 invested
  • The amount of funding for research available for the “Brain Initiative,” a collaboration between the National Institute of Health, DARPA and the National Science Foundation, which seeks to uncover new prevention and treatment methods for brain disorders like Alzheimer’s, autism and schizophrenia: $100 million
  • Total amount offered in cash awards by the Department of Energy’s “SunShot Grand Challenge,” which seeks to eliminate the cost disparity between solar energy and coal by the end of the decade: $10 million

Sources

The Future of Public Space Analytics


The AGILE Landscape Project: “Public space is an essential component of any great city. It brings people together to socialize, recreate, and work. More pointedly, it attracts people to the city, builds relationships, and spurs innovation and new ideas that fuel a city’s economic growth. How we optimize the investments made in our streetscapes, plazas, parks, and greenways is important to each individual project’s success and the city as a whole. Are the places being built fulfilling their promise? If so, could they be doing more and if not, how do they need to change?…
“Places in the Making”, a publication published in 2013 by MIT’s Department of Urban Studies and Planning, underscored this problem. They cited the lack of active metrics as a common challenge to placemaking efforts. They reported that “It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” They went on to explain the existing placemaking culture focuses on fuzzy, unmeasurable goals and that this has created “an inertia in assessment efforts”. The report stressed that this lack of assessment is detrimental to the field as a whole because valuable insights are left undiscovered and the same mistakes are made repeatedly….
“Places in the Making”, a publication published in 2013 by MIT’s Department of Urban Studies and Planning, underscored this problem. They cited the lack of active metrics as a common challenge to placemaking efforts. They reported that “It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” They went on to explain the existing placemaking culture focuses on fuzzy, unmeasurable goals and that this has created “an inertia in assessment efforts”. The report stressed that this lack of assessment is detrimental to the field as a whole because valuable insights are left undiscovered and the same mistakes are made repeatedly.

“It is astonishing how few placemaking projects actively and honestly assess their own successes and failures.” – “Places in the Making”, DUSP (2013)

Both William H. Whyte’s seminal Street Life Project and resulting book and documentary, Social Life of Small Urban Spaces (1980), along with Jan Gehl’s extensive public life studies and influential book Life Between Buildings (1987) serve as the foundation of the urban design profession’s knowledge base. Their observations and insights were amassed through considerable time and manpower. We are lucky to have their contribution from which to draw upon when making future decisions. Though we shouldn’t stop observing and building upon their significant contributions. We must continually add clarity and resolution particularly in the unique and specific context of each place.
social media metrics
Unfortunately, the POE and previous research are only snapshots in time. With the rise of embedded sensor infrastructure, “smart cities”, social networks, and ubiquity of smartphones that are becoming increasingly more powerful and sophisticated every year, the ability to gain real-time continuous 24/7 quantitative monitoring of public life data is now a reality. The ability to capture so much data is unprecedented and will only continue to get easier. The amount of and type of information that can and often is being collected about the world around us and ourselves is vast. When it comes to cities, transportation, parking, energy, waste collection, and water systems has attracted the most attention. In the private sector, retail analytics have been the focus.
Collecting data for data sake should not be a goal. Rather we need be knowledgeable of what is available and possible while being discerning about the data that is useful for improving the management and design of public space. By selectively incorporating sensors into the design of physical space, the data collection system can provide real-time information about the use of the space. Like website analytic systems, the network can provide useful information about the number of people using the space, how long they are using it, as well well as other meaningful measures of public life. The data gathered can be combined with a wide array of environmental data such as temperature, humidity, sun/shade patterns, rain fall, water quality, and air quality. In addition, other available data sets such as sales tax collections, building permits, real estate transactions, and programming/event schedules can be overlaid with data to make relationships between the use of the public space and its context more visible.
Array of information that can help us understand public space
While the real-time nature of the data can be used to quickly understand the immediate impact that programming and design decisions have on the space and vice versa, it can also capture data and trends over extended durations that would not otherwise be possible without considerable resources. All of this enables us to make more informed decisions rather than anecdotal observations and assumptions. With the interest of tactical urbanism and other short-term public space investments, this need for data is greater than ever. The data can be used to justify investment in long-term improvements and/or make slight adjustments in temporary spaces to increase their effectiveness….(More)”

Choosing Not to Choose: Understanding the Value of Choice


New book by Cass Sunstein: “Our ability to make choices is fundamental to our sense of ourselves as human beings, and essential to the political values of freedom-protecting nations. Whom we love; where we work; how we spend our time; what we buy; such choices define us in the eyes of ourselves and others, and much blood and ink has been spilt to establish and protect our rights to make them freely.
Choice can also be a burden. Our cognitive capacity to research and make the best decisions is limited, so every active choice comes at a cost. In modern life the requirement to make active choices can often be overwhelming. So, across broad areas of our lives, from health plans to energy suppliers, many of us choose not to choose. By following our default options, we save ourselves the costs of making active choices. By setting those options, governments and corporations dictate the outcomes for when we decide by default. This is among the most significant ways in which they effect social change, yet we are just beginning to understand the power and impact of default rules. Many central questions remain unanswered: When should governments set such defaults, and when should they insist on active choices? How should such defaults be made? What makes some defaults successful while others fail?….
The onset of big data gives corporations and governments the power to make ever more sophisticated decisions on our behalf, defaulting us to buy the goods we predictably want, or vote for the parties and policies we predictably support. As consumers we are starting to embrace the benefits this can bring. But should we? What will be the long-term effects of limiting our active choices on our agency? And can such personalized defaults be imported from the marketplace to politics and the law? Confronting the challenging future of data-driven decision-making, Sunstein presents a manifesto for how personalized defaults should be used to enhance, rather than restrict, our freedom and well-being. (More)”

The Internet’s hidden science factory


Jenny Marder at PBS Newshour: “….Marshall is a worker for Amazon’s Mechanical Turk, an online job forum where “requesters” post jobs, and an army of crowdsourced workers complete them, earning fantastically small fees for each task. The work has been called microlabor, and the jobs, known as Human Intelligence Tasks, or HITs, range wildly. Some are tedious: transcribing interviews or cropping photos. Some are funny: prank calling someone’s buddy (that’s worth $1) or writing the title to a pornographic movie based on a collection of dirty screen grabs (6 cents). And others are downright bizarre. One task, for example, asked workers to strap live fish to their chests and upload the photos. That paid $5 — a lot by Mechanical Turk standards….
These aren’t obscure studies that Turkers are feeding. They span dozens of fields of research, including social, cognitive and clinical psychology, economics, political science and medicine. They teach us about human behavior. They deal in subjects like energy conservation, adolescent alcohol use, managing money and developing effective teaching methods.


….In 2010, the researcher Joseph Henrich and his team published a paper showing that an American undergraduate was about 4,000 times more likely than an average American to be the subject of a research study.
But that output pales in comparison to Mechanical Turk workers. The typical “Turker” completes more studies in a week than the typical undergraduate completes in a lifetime. That’s according to research by Rand, who surveyed both groups. Among those he surveyed, he found that the median traditional lab subject had completed 15 total academic studies — an average of one per week. The median Turker, on the other hand, had completed 300 total academic studies — an average of 20 per week….(More)”

Can Selfies Save Nutrition Science?


Trevor Butterworth at Stats.org: “You may have never heard of the Energy Balance Working Group, but this collection of 45 experts on nutrition, exercise, biochemistry, and other related disciplines have collectively thrown a “House-like” wrench into the research literature on everything from obesity to cancer and heart disease. Gregory House, the fictional and fantastically brilliant physician played by Hugh Laurie in the eponymous TV show frequently found his patients wanting in the court of self-reported truth: “I’ve found that when you want to know the truth about someone that someone is probably the last person you should ask.”
This is more or less what the Energy Balance Working Group have concluded in an “expert report” recently published in the International Journal of Obesity. If you want to know the truth about how much someone eats and exercises that someone is probably the last person you should ask….The problem is that self-reporting is a cheap and convenient source of data for research, while more accurate alternatives are either expensive and challenging or, as yet, more promise than reality (see sidebar)….
“There are at least two categories of solutions on the horizon. In one category, there are wearable monitoring devices that can collect objective, real-time data. Examples in the works or in use include photographic food diaries, records of chewing and swallowing behavior, and evaluating the time and intensity of movement using accelerometers and GPS, among others. It is important to note that there are still challenges converting these measurements into reliable estimates of energy intake and expenditure, but work is ongoing… David Allison, Distinguished Professor, Quetelet Endowed Professor of Public Health, University of Alabam”…(More)