Can national statistical offices shape the data revolution?


Article by Juan Daniel Oviedo, Katharina Fenz, François Fonteneau, and Simon Riedl: “In recent years, breakthrough technologies in artificial intelligence (AI) and the use of satellite imagery made it possible to disrupt the way we collect, process, and analyze data. Facilitated by the intersection of new statistical techniques and the availability of (big) data, it is now possible to create hypergranular estimates.

National statistical offices (NSOs) could be at the forefront of this change. Conventional tasks of statistical offices, such as the coordination of household surveys and censuses, will remain at the core of their work. However, just like AI can enhance the capabilities of doctors, it also has the potential to make statistical offices better, faster, and eventually cheaper.

Still, many countries struggle to make this happen. In a COVID-19 world marked by constrained financial and statistical capacities, making innovation work for statistical offices is of prime importance to create better lives for all…

In the case of Colombia, this novel method facilitated a scale-up from existing poverty estimates that contained 1,123 data points to 78,000 data points, which represents a 70-fold increase. This results in much more granular estimates highlighting Colombia’s heterogeneity between and within municipalities (see Figure 1).

Figure 1. Poverty shares (%) Colombia, in 2018

Figure 1. Poverty shares (%) Colombia, in 2018

Traditional methods don´t allow for cost-efficient hypergranular estimations but serve as a reference point, due to their ground-truthing capacity. Hence, we have combined existing data with novel AI techniques, to go down to granular estimates of up to 4×4 kilometers. In particular, we have trained an algorithm to connect daytime and nighttime satellite images….(More)”.

The promise and the premise: How digital media present big data


Paper by Gastón Becerra: “This paper analyzes the thematic and discursive construction of big data by the Argentine digital press. Using text mining techniques — topic modelling and enriched associative networks — together with qualitative and quantitative content analysis — in both discourse and images — over 2,026 articles, we sought to identify the topics wherein big data is treated, the promises and risks it addresses, its definition within the semantic field in which is explicitly expressed, and the pictures that illustrate it. Results herein presented compare how big data is portrayed in news about politics, business, and technological innovations, as well as in focal pieces targeted to a generic and massive audience, and critical reflections about its risks. Although in each of those thematic contexts big data is anchored differently, there is a common idea that associates big data with a socio-technological premise and an epistemic promise: because of the availability of large volumes of data, something new that will allow better decisions can be known. Our exploration contributes to a more detailed knowledge on how the news media social systems make sense of novel phenomena such as big data….(More)”.

The geography of AI


Report by Mark Muro and Sifan Liu: “Much of the U.S. artificial intelligence (AI) discussion revolves around futuristic dreams of both utopia and dystopia. From extreme to extreme, the promises range from solutions to global climate change to a “robot apocalypse.”

However, it bears remembering that AI is also becoming a real-world economic fact with major implications for national and regional economic development as the U.S. crawls out of the COVID-19 pandemic.

Based on advanced uses of statistics, algorithms, and fast computer processing, AI has become a focal point of U.S. innovation debates. Even more, AI is increasingly viewed as the next great “general purpose technology”—one that has the power to boost the productivity of sector after sector of the economy.

All of which is why state and city leaders are increasingly assessing AI for its potential to spur economic growth. Such leaders are analyzing where their regions stand and what they need to do to ensure their locations are not left behind.

In response to such questions, this analysis examines the extent, location, and concentration of AI technology creation and business activity in U.S. metropolitan areas.

Employing seven basic measures of AI capacity, the report benchmarks regions on the basis of their core AI assets and capabilities as they relate to two basic dimensions: AI research and AI commercialization. In doing so, the assessment categorizes metro areas into five tiers of regional AI involvement and extracts four main findings reflecting that involvement…(More)”.

The Mobility Data Sharing Assessment


New Tool from the Mobility Data Collaborative (MDC): “…released a set of resources to support transparent and accountable decision making about how and when to share mobility data between organizations. …The Mobility Data Sharing Assessment (MDSA) is a practical and customizable assessment that provides operational guidance to support an organization’s existing processes when sharing or receiving mobility data. It consists of a collection of resources:

  • 1. A Tool that provides a practical, customizable and open-source assessment for organizations to conduct a self-assessment.
  • 2. An Operator’s Manual that provides detailed instructions, guidance and additional resources to assist organizations as they complete the tool.
  • 3. An Infographic that provides a visual overview of the MDSA process.

“We were excited to work with the MDC to create a practical set of resources to support mobility data sharing between organizations,” said Chelsey Colbert, policy counsel at FPF. “Through collaboration, we designed version one of a technology-neutral tool, which is consistent and interoperable with leading industry frameworks. The MDSA was designed to be a flexible and scalable approach that enables mobility data sharing initiatives by encouraging organizations of all sizes to assess the legal, privacy, and ethical considerations.”

New mobility options, such as shared cars and e-scooters, have rapidly emerged in cities over the past decade. Data generated by these mobility services offers an exciting opportunity to provide valuable and timely insight to effectively develop transportation policy and infrastructure. As the world becomes more data-driven, tools like the MDSA help remove barriers to safe data sharing without compromising consumer trust….(More)”.

Kansas City expands civic engagement with data stories, virtual ‘lunch-and-learns’


Ryan Johnston at Statescoop: “…The city is currently running a series of virtual “lunch-and-learns,” as well as publishing data-driven “stories” using Socrata software to improve civic engagement, said Kate Bender, a senior management analyst in the city’s data division.

The work is especially important in reaching residents that aren’t equipped with digital literacy or data analysis skills, Bender said. The free lunch-and-learns — managed under the new Office of Citizen Engagement — teaches residents how to use digital tools like the city’s open data portal and 311 mobile app.

New data stories, meanwhile, published on the city’s open data portal, allow residents to see the context behind raw data around COVID-19, 311 requests or city hiring practices that they might not otherwise be able to parse themselves. They’re both part of an effort to reach residents that aren’t already plugged in to the city’s digital channels, Bender said.

“Knowing that we have more digital options and we have good engagement, how can we open up residents’ exposure to other things, and specifically tools, that we make available, that we put on our website or that we tweet about?” Bender said. “Unless you’re already pretty engaged, you might not know or think to download the city’s 311 app, or you might have heard of open data, but not really know how it pertains to you. So that was our concept.”

Bender’s office, DataKC, has “always been pretty closely aligned in working with 311 and advising on citizen engagement,” Bender said. But when COVID-19 hit and people could no longer gather in-person for citizen engagement events, like the city’s “Community Engagement University,” a free, 8-week, in-person program that taught residents about how various city agencies work, Bender and her team decided to take the education component virtual….(More)”.

Data in Crisis — Rethinking Disaster Preparedness in the United States


Paper by Satchit Balsari, Mathew V. Kiang, and Caroline O. Buckee: “…In recent years, large-scale streams of digital data on medical needs, population vulnerabilities, physical and medical infrastructure, human mobility, and environmental conditions have become available in near-real time. Sophisticated analytic methods for combining them meaningfully are being developed and are rapidly evolving. However, the translation of these data and methods into improved disaster response faces substantial challenges. The data exist but are not readily accessible to hospitals and response agencies. The analytic pipelines to rapidly translate them into policy-relevant insights are lacking, and there is no clear designation of responsibility or mandate to integrate them into disaster-mitigation or disaster-response strategies. Building these integrated translational pipelines that use data rapidly and effectively to address the health effects of natural disasters will require substantial investments, and these investments will, in turn, rely on clear evidence of which approaches actually improve outcomes. Public health institutions face some ongoing barriers to achieving this goal, but promising solutions are available….(More)”

The U.S. Is Getting a Crash Course in Scientific Uncertainty


Apoorva Mandavilli at the New York Times: “When the coronavirus surfaced last year, no one was prepared for it to invade every aspect of daily life for so long, so insidiously. The pandemic has forced Americans to wrestle with life-or-death choices every day of the past 18 months — and there’s no end in sight.

Scientific understanding of the virus changes by the hour, it seems. The virus spreads only by close contact or on contaminated surfaces, then turns out to be airborne. The virus mutates slowly, but then emerges in a series of dangerous new forms. Americans don’t need to wear masks. Wait, they do.

At no point in this ordeal has the ground beneath our feet seemed so uncertain. In just the past week, federal health officials said they would begin offering booster shots to all Americans in the coming months. Days earlier, those officials had assured the public that the vaccines were holding strong against the Delta variant of the virus, and that boosters would not be necessary.

As early as Monday, the Food and Drug Administration is expected to formally approve the Pfizer-BioNTech vaccine, which has already been given to scores of millions of Americans. Some holdouts found it suspicious that the vaccine was not formally approved yet somehow widely dispensed. For them, “emergency authorization” has never seemed quite enough.

Americans are living with science as it unfolds in real time. The process has always been fluid, unpredictable. But rarely has it moved at this speed, leaving citizens to confront research findings as soon as they land at the front door, a stream of deliveries that no one ordered and no one wants.

Is a visit to my ailing parent too dangerous? Do the benefits of in-person schooling outweigh the possibility of physical harm to my child? Will our family gathering turn into a superspreader event?

Living with a capricious enemy has been unsettling even for researchers, public health officials and journalists who are used to the mutable nature of science. They, too, have frequently agonized over the best way to keep themselves and their loved ones safe.

But to frustrated Americans unfamiliar with the circuitous and often contentious path to scientific discovery, public health officials have seemed at times to be moving the goal posts and flip-flopping, or misleading, even lying to, the country.

Most of the time, scientists are “edging forward in a very incremental way,” said Richard Sever, assistant director of Cold Spring Harbor Laboratory Press and a co-founder of two popular websites, bioRxiv and medRxiv, where scientists post new research.

“There are blind alleys that people go down, and a lot of the time you kind of don’t know what you don’t know.”

Biology and medicine are particularly demanding fields. Ideas are evaluated for years, sometimes decades, before they are accepted….(More)”.

The Secret Bias Hidden in Mortgage-Approval Algorithms


An investigation by The Markup: “…has found that lenders in 2019 were more likely to deny home loans to people of color than to White people with similar financial characteristics—even when we controlled for newly available financial factors that the mortgage industry for years has said would explain racial disparities in lending.

Holding 17 different factors steady in a complex statistical analysis of more than two million conventional mortgage applications for home purchases, we found that lenders were 40 percent more likely to turn down Latino applicants for loans, 50 percent more likely to deny Asian/Pacific Islander applicants, and 70 percent more likely to deny Native American applicants than similar White applicants. Lenders were 80 percent more likely to reject Black applicants than similar White applicants. These are national rates.

In every case, the prospective borrowers of color looked almost exactly the same on paper as the White applicants, except for their race.

The industry had criticized previous similar analyses for not including financial factors they said would explain disparities in lending rates but were not public at the time: debts as a percentage of income, how much of the property’s assessed worth the person is asking to borrow, and the applicant’s credit score.

The first two are now public in the Home Mortgage Disclosure Act data. Including these financial data points in our analysis not only failed to eliminate racial disparities in loan denials, it highlighted new, devastating ones.

We found that lenders gave fewer loans to Black applicants than White applicants even when their incomes were high—$100,000 a year or more—and had the same debt ratios. In fact, high-earning Black applicants with less debt were rejected more often than high-earning White applicants who have more debt….(More)”

How local governments are scaring tech companies


Ben Brody at Protocol: “Congress has failed to regulate tech, so states and cities are stepping in with their own approaches to food delivery apps, AI regulation and, yes, privacy. Tech doesn’t like what it sees….

Andrew Rigie said it isn’t worth waiting around for tech regulation in Washington.

“New York City is a restaurant capital of the world,” Rigie told Protocol. “We need to lead on these issues.”

Rigie, executive director of the New York City Hospitality Alliance, has pushed for New York City’s new laws on food delivery apps such as Uber Eats. His group supported measures to make permanent a cap on the service fees the apps charge to restaurants, ban the apps from listing eateries without permission and share customer information with restaurants that ask for it.

While Rigie’s official purview is dining in the Big Apple, his belief that the local government should lead on regulating tech companies in a way Washington hasn’t has become increasingly common.

“It wouldn’t be a surprise if lawmakers elsewhere seek to implement similar policies,” Rigie said. “Some of it could potentially come from the federal government, but New York City can’t wait for the federal government to maybe act.”

New York is not the only city to take action. While the Federal Trade Commission has faced calls to regulate third-party food delivery apps at a national level, San Francisco was first to pass a permanent fee cap for them in June.

Food apps are just a microcosm highlighting the patchworks of local-level regulation that are developing, or are already a fact of life, for tech. These regulatory patchworks occur when state and local governments move ahead of Congress to pass their own, often divergent, laws and rules. So far, states and municipalities are racing ahead of the feds on issues such as cybersecurity, municipal broadbandcontent moderationgig work, the use of facial recognition, digital taxes, mobile app store fees and consumer rights to repair their own devices, among others.

Many in tech became familiar with the idea when the California Consumer Privacy Act passed in 2018, making it clear more states would follow suit, although the possibility has popped up throughout modern tech policy history on issues such as privacy requirements on ISPs, net neutrality and even cybersecurity breach notification.

Many patchworks reflect the stance of advocates, consumers and legislators that Washington has simply failed to do its job on tech. The resulting uncompromising or inconsistent approaches by local governments also has tech companies worried enough to push Congress to overrule states and establish one uniform U.S. standard.

“With a bit of a vacuum at the federal level, states are looking to step in, whether that’s on content moderation, whether that’s on speech on platforms, antitrust and anticompetitive conduct regulation, data privacy,” said April Doss, executive director of Georgetown University’s Institute for Technology Law and Policy. “It is the whole bundle of issues.”…(More)

Abundance: On the Experience of Living in a World of Information Plenty


Book by Pablo J. Boczkowski: “The book examines the experience of living in a society that has more information available to the public than ever before. It focuses on the interpretations, emotions, and practices of dealing with this abundance in everyday life. Drawing upon extensive fieldwork and survey research conducted in Argentina, the book inquiries into the role of cultural and structural factors that mediate between the availability of information and the actual consequences for individuals, media, politics, and society. Providing the first book-length account of the topic in the Global South, it concludes that the experience of information abundance is tied to an overall unsettling of society, a reconstitution of how we understand and perform our relationships with others, and a twin depreciation of facts and appreciation of fictions….(More)”.