Omnivorous Analysis


Essay by Anne Lee Steele: “Satellite imagery has woven itself into the fabric of the internet. We recognize these crisp, high-definition, bird’s-eye-view images most commonly from Google Earth—but we employ them in much more besides: from reporting on stuck shipping containers to getting directions to a friend’s house, to tracking forest fires in real time and scrolling through real estate listings. Given their ever-widening range of commercial, consumer, and civic uses, it won’t surprise most people to hear that the industry that produces them (also known as Earth Observation, or EO) is growing at an exponential rate, and is only expected to expand further in the coming years.

Yet despite the prominence of satellite imagery in the geographical imagination of the internet, the imperatives of the industry are much less clear. The corporations that produce them are much less well known, and the military interests that back them remain as murky as ever. The highly visible commercial side of the industry is still deeply intertwined with its classified counterpart, and two companies, Maxar and Planet, have emerged to dominate the industry—supporting civilian functions with one hand, while supplying US defense needs with the other. 

Indeed, the ubiquity of commercial satellite imagery gives nearly anyone godlike powers of reconnaissance and surveillance not that far removed from those enjoyed by militaries and intelligence agencies—a fact that causes no small amount of anxiety within the Pentagon. The pervasiveness and power of their imagery compels us to ask: Where do they come from? And how are they being put to use?…(More)”.

Imagining Governance for Emerging Technologies


Essay by Debra J.H. Mathews, Rachel Fabi and Anaeze C. Offodile: “…How should such technologies be regulated and governed? It is increasingly clear that past governance structures and strategies are not up to the task. What these technologies require is a new governance approach that accounts for their interdisciplinary impacts and potential for both good and ill at both the individual and societal level. 

To help lay the groundwork for a novel governance framework that will enable policymakers to better understand these technologies’ cross-sectoral footprint and anticipate and address the social, legal, ethical, and governance issues they raise, our team worked under the auspices of the National Academy of Medicine’s Committee on Emerging Science, Technology, and Innovation in health and medicine (CESTI) to develop an analytical approach to technology impacts and governance. The approach is grounded in detailed case studies—including the vignettes about Robyn and Liam—which have informed the development of a set of guiding principles (see sidebar).

Based on careful analysis of past governance, these case studies also contain a plausible vision of what might happen in the future. They illuminate ethical issues and help reveal governance tools and choices that could be crucial to delivering social benefits and reducing or avoiding harms. We believe that the approach taken by the committee will be widely applicable to considering the governance of emerging health technologies. Our methodology and process, as we describe here, may also be useful to a range of stakeholders involved in governance issues like these…(More)”.

How We Can Encode Human Rights In The Blockchain


Essay by Nathan Schneider: “Imagine there is a new decentralized finance app quietly spreading around the world that’s like a payday lender from hell. Call it DevilsBridge. Rather than getting it from the App Store, you access its blockchain contracts directly, using a Web browser with a crypto-wallet plugin. DevilsBridge provides small loans in cryptocurrency that “bridge” people to the next paycheck. The interest rates are far below those of conventional payday lenders, which is life-changing for many users.

But if the payments go unpaid, they grow. They balloon. They reach multiples upon multiples of the principal. As time goes on, pressure ratchets up on borrowers, who become notorious for undertaking desperate, violent crimes to pay back their exorbitant debts. The deal, after all, is that if a debt reaches the magic threshold of $1 million, the debtor becomes a target. A private market of poison-dart-shooting drones receives a bounty to assassinate the mega-debtors.

Anywhere there are laws, of course, this is all wildly illegal. But nobody knows who created DevilDAO, the decentralized autonomous organization that operates DevilsBridge, or who its members are. The identities of the drone owners also hide behind cryptographic gibberish. Sometimes local police can trace the drones back to their bases, or investigators can trace a DevilDAO member’s address to a real person. But in most places where the assassinations happen, authorities are ill-equipped for airborne chases or scrutinizing blockchain analytics.

This may sound like a cartoonish scenario, but it’s freshly plausible thanks to the advent of decentralized, autonomous systems on blockchains. Ethereum co-founder Vitalik Buterin jokingly nodded to such dystopian possibilities in early 2014, when he listed possible uses for his proposed blockchain, from crop insurance to decentralized social networks — or perhaps, he said as he walked away from the mic, it could allow for the creation of Skynet, the robot intelligence in the “Terminator” movies that tries to exterminate the human race.

The potential for blockchain-enabled human-rights abuses is real. At the same time, these technologies introduce new ways of encoding and enforcing rights. Imagine the blockchain that DevilsBridge runs on introduces a software update. It bans any smart contract that kills humans. An anonymous investigator presents evidence of what the app is doing, and an anonymous jury confirms its validity; instantly, the contracts for DevilsBridge and DevilDAO no longer function…(More)”.

The linguistics search engine that overturned the federal mask mandate


Article by Nicole Wetsman: “The COVID-19 pandemic was still raging when a federal judge in Florida made the fateful decision to type “sanitation” into the search bar of the Corpus of Historical American English.

Many parts of the country had already dropped mask requirements, but a federal mask mandate on planes and other public transportation was still in place. A lawsuit challenging the mandate had come before Judge Kathryn Mizelle, a former clerk for Justice Clarence Thomas. The Biden administration said the mandate was valid, based on a law that authorizes the Centers for Disease Control and Prevention (CDC) to introduce rules around “sanitation” to prevent the spread of disease.

Mizelle took a textualist approach to the question — looking specifically at the meaning of the words in the law. But along with consulting dictionaries, she consulted a database of language, called a corpus, built by a Brigham Young University linguistics professor for other linguists. Pulling every example of the word “sanitation” from 1930 to 1944, she concluded that “sanitation” was used to describe actively making something clean — not as a way to keep something clean. So, she decided, masks aren’t actually “sanitation.”

The mask mandate was overturned, one of the final steps in the defanging of public health authorities, even as infectious disease ran rampant…

Using corpora to answer legal questions, a strategy often referred to as legal corpus linguistics, has grown increasingly popular in some legal circles within the past decade. It’s been used by judges on the Michigan Supreme Court and the Utah Supreme Court, and, this past March, was referenced by the US Supreme Court during oral arguments for the first time.

“It’s been growing rapidly since 2018,” says Kevin Tobia, a professor at Georgetown Law. “And it’s only going to continue to grow.”…(More)”.

In this small Va. town, citizens review police like Uber drivers


Article by Emily Davies: “Chris Ford stepped on the gas in his police cruiser and rolled down Gold Cup Drive to catch the SUV pushing 30 mph in a 15 mph zone. Eleven hours and 37 minutes into his shift, the corporal was ready for his first traffic stop of the day.

“Look at him being sneaky,” Fordsaid, his blue lights flashing on a quiet road in this small town where a busy day could mean animals escaped from a local slaughterhouse.

Ford parked, walked toward the SUV and greeted the man who had ignored the speed limit at exactly the wrong time.

“I was doing 15,” said the driver, a Black man in a mostly White neighborhood of a mostly White town.

The officertook his license and registration back to the cruiser.

“Every time I pull over someone of color, they’re standoffish with me. Like, ‘Here’s a White police officer, here we go again.’ ” Ford, 56, said. “So I just try to be nice.”

Ford knew the stop would be scrutinized — and not just by the reporter who was allowed to ride along on his shift.

After every significant encounter with residents, officers in Warrenton are required to hand out a QR code, which is on the back of their business card, asking for feedback on the interaction. Through a series of questions, citizens can use a star-based system to rate officers on their communication, listening skills and fairness. The responses are anonymous and can be completed any time after the interaction to encourage people to give honest assessments. The program, called Guardian Score, is supposed to give power to those stopped by police in a relationship that has historically felt one-sided — and to give police departments a tool to evaluate their force on more than arrests and tickets.

“If we started to measure how officers are treating community members, we realized we could actually infuse this into the overall evaluation process of individual officers,” said Burke Brownfeld, a founder of Guardian Score and a former police officer in Alexandria. “The definition of doing a good job could change. It would also include: How are your listening skills? How fairly are you treating people based on their perception?”…(More)”.

How science could aid the US quest for environmental justice


Jeff Tollefson at Nature: “…The network of US monitoring stations that detect air pollution catches only broad trends across cities and regions, and isn’t equipped for assessing air quality at the level of streets and neighbourhoods. So environmental scientists are exploring ways to fill the gaps.

In one project funded by NASA, researchers are developing methods to assess street-level pollution using measurements of aerosols and other contaminants from space. When the team trained its tools on Washington DC, the scientists found1 that sections in the city’s southeast, which have a larger share of Black residents, are exposed to much higher levels of fine-soot pollution than wealthier — and whiter — areas in the northwest of the city, primarily because of the presence of major roads and bus depots in the southeast.

Cumulative burden: Air-pollution levels tend to be higher in poorer and predominantly Black neighbourhoods of Washington DC.
Source: Ref. 1

The detailed pollution data painted a more accurate picture of the burden on a community that also lacks access to high-quality medical facilities and has high rates of cardiovascular disorders and other diseases. The results help to explain a more than 15-year difference in life expectancy between predominantly white neighbourhoods and some predominantly Black ones.

The analysis underscores the need to consider pollution and socio-economic data in parallel, says Susan Anenberg, director of the Climate and Health Institute at the George Washington University in Washington DC and co-leader of the project. “We can actually get neighbourhood-scale observations from space, which is quite incredible,” she says, “but if you don’t have the demographic, economic and health data as well, you’re missing a very important piece of the puzzle.”

Other projects, including one from technology company Aclima, in San Francisco, California, are focusing on ubiquitous, low-cost sensors that measure air pollution at the street level. Over the past few years, Aclima has deployed a fleet of vehicles to collect street-level data on air pollutants such as soot and greenhouse gases across 101 municipalities in the San Francisco Bay area. Their data have shown that air-pollution levels can vary as much as 800% from one neighbourhood block to the next.

Working directly with disadvantaged communities and environmental regulators in California, as well as with other states and localities, the company provides pollution monitoring on a subscription basis. It also offers the use of its screening tool, which integrates a suite of socio-economic data and can be used to assess cumulative impacts…(More)”.

Stories to Work By


Essay by William E. Spriggs: “In Charlie Chaplin’s 1936 film Modern Times, humans in a factory are reduced to adjuncts to a massive series of cogs and belts. Overlords bark commands from afar to a servant class, and Chaplin’s hapless hero is literally consumed by the machine … and then spit out by it. In the film, the bosses have all the power, and machines keep workers in check.

Modern Times’s dystopian narrative remains with us today. In particular, it is still held by many policymakers who assume that increasing technological progress, whether mechanical or informational, inevitably means that ordinary workers will lose. This view perpetuates itself when policies that could give workers more power in times of technological change are overlooked, while those that disempower workers are adopted. If we are to truly consider science policy for the future, we need to understand how this narrative about workers and technology functions, where it is misleading, and how deliberate policies can build a better world for all….

Today’s tales of pending technological dystopia—echoed in economics papers as well as in movies and news reports—blind us to the lessons we could glean from the massive disruptions of earlier periods of even greater change. Today the threat of AI is portrayed as revolutionary, and previous technological change as slow and inconsequential—but this was never the case. These narratives of technological inevitability limit the tools we have at our disposal to promote equality and opportunity.

The challenges we face today are far from insurmountable: technology is not destiny. Workers are not doomed to be Chaplin’s victim of technology with one toe caught in the gears of progress. We have choices, and the central challenge of science and technology policy for the next century will be confronting those choices head on. Policymakers should focus on the fundamental tasks of shaping how technology is deployed and enacting the economic rules we need to ensure that technology works for us all, rather than only the few….(More)”.

I tried to read all my app privacy policies. It was 1 million words.


Article by Geoffrey A. Fowler: “…So here’s an idea: Let’s abolish the notion that we’re supposed to read privacy policies.

I’m not suggesting companies shouldn’t have to explain what they’re up to. Maybe we call them “data disclosures” for the regulators, lawyers, investigative journalists and curious consumers to pore over.

But to protect our privacy, the best place to start is for companies to simply collect less data. “Maybe don’t do things that need a million words of explanation? Do it differently,” said Slaughter. “You can’t abuse, misuse, leverage data that you haven’t collected in the first place.”

Apps and services should only collect the information they really need to provide that service — unless we opt in to let them collect more, and it’s truly an option.

I’m not holding my breath that companies will do that voluntarily, but a federal privacy law would help. While we wait for one, Slaughter said the FTC (where Democratic commissioners recently gained a majority) is thinking about how to use its existing authority “to pursue practices — including data collection, use and misuse — that are unfair to users.”

Second, we need to replace the theater of pressing “agree” with real choices about our privacy.

Today, when we do have choices to make, companies often present them in ways that pressure us into making the worst decisions for ourselves.

Apps and websites should give us the relevant information and our choices in the moment when it matters. Twitter actually does this just-in-time notice better than many other apps and websites: By default, it doesn’t collect your exact location, and only prompts you to do so when you ask to tag your location in a tweet.

Even better, technology could help us manage our choices. Cranor suggests that data disclosures could be coded to be read by machines. Companies already do this for financial information, and the TLDR Act would require consistent tags on privacy information, too. Then your computer could act kind of like a butler, interacting with apps and websites on your behalf.

Picture Siri as a butler who quizzes you briefly about your preferences and then does your bidding. The privacy settings on an iPhone already let you tell all the different apps on your phone not to collect your location. For the past year, they’ve also allowed you to ask apps not to track you.

Web browsers could serve as privacy butlers, too. Mozilla’s Firefox already lets you block certain kinds of privacy invasions. Now a new technology called the Global Privacy Control is emerging that would interact with websites and instruct them not to “sell” our data. It’s grounded in California’s privacy law, which is among the toughest in the nation, though it remains to be seen how the state will enforce GPC…(More)”.

Facial Expressions Do Not Reveal Emotions


Lisa Feldman Barrett at Scientific American: “Do your facial movements broadcast your emotions to other people? If you think the answer is yes, think again. This question is under contentious debate. Some experts maintain that people around the world make specific, recognizable faces that express certain emotions, such as smiling in happiness, scowling in anger and gasping with widened eyes in fear. They point to hundreds of studies that appear to demonstrate that smiles, frowns, and so on are universal facial expressions of emotion. They also often cite Charles Darwin’s 1872 book The Expression of the Emotions in Man and Animals to support the claim that universal expressions evolved by natural selection.

Other scientists point to a mountain of counterevidence showing that facial movements during emotions vary too widely to be universal beacons of emotional meaning. People may smile in hatred when plotting their enemy’s downfall and scowl in delight when they hear a bad pun. In Melanesian culture, a wide-eyed gasping face is a symbol of aggression, not fear. These experts say the alleged universal expressions just represent cultural stereotypes. To be clear, both sides in the debate acknowledge that facial movements vary for a given emotion; the disagreement is about whether there is enough uniformity to detect what someone is feeling.

This debate is not just academic; the outcome has serious consequences. Today you can be turned down for a job because a so-called emotion-reading system watching you on camera applied artificial intelligence to evaluate your facial movements unfavorably during an interview. In a U.S. court of law, a judge or jury may sometimes hand down a harsher sentence, even death, if they think a defendant’s face showed a lack of remorse. Children in preschools across the country are taught to recognize smiles as happiness, scowls as anger and other expressive stereotypes from books, games and posters of disembodied faces. And for children on the autism spectrum, some of whom have difficulty perceiving emotion in others, these teachings do not translate to better communication….Emotion AI systems, therefore, do not detect emotions. They detect physical signals, such as facial muscle movements, not the psychological meaning of those signals. The conflation of movement and meaning is deeply embedded in Western culture and in science. An example is a recent high-profile study that applied machine learning to more than six million internet videos of faces. The human raters, who trained the AI system, were asked to label facial movements in the videos, but the only labels they were given to use were emotion words, such as “angry,” rather than physical descriptions, such as “scowling.” Moreover there was no objective way to confirm what, if anything, the anonymous people in the videos were feeling in those moments…(More)”.

Barcelona bets on ‘digital twin’ as future of city planning


Article by Aitor Hernández-Morales: “In five years’ time, the structure of Europe’s cities won’t be decided in local town halls but inside a quiet 19th-century chapel in a leafy neighborhood of Barcelona.

Housed in the deconsecrated Torre Girona chapel, the MareNostrum supercomputer — one of the world’s most powerful data processors — is already busily analyzing how to improve city planning in Barcelona.

Barcelona is using data to track access to primary health care centers throughout the city | BSC

“We’re using the supercomputer to make sure the urban planning process isn’t just based on clever ideas and good intentions, but on data that allows us to anticipate its impacts and avoid the negative ones,” said Barcelona Deputy Mayor Laia Bonet, who is in charge of the city’s digital transition, climate goals and international partnerships.

As part of a pilot project launched with the Italian city of Bologna earlier this year, Barcelona has created a data-based replica of itself — a digital twin — where it can trial run potential city planning projects.

“Instead of implementing flawed policies and then have to go back and correct them, we’re saving time by making sure those decisions are right before we execute them,” said Bonet.

Although the scheme is still in its test phase, Bonet said she expects the city’s high-tech approach to urban development will soon be the norm in cities across the EU.

“Within a five-year horizon I expect to see this as a basic urban planning tool,” she said.

Looking for blindspots

Barcelona’s popular superilles, or “superblocks,” are a prime example of an urban scheme that could have benefited from data modelling in the planning stages, according to Bonet.

Since 2014 the city has been creating mini-neighborhoods where through-traffic and on-street parking is all but banned, with the goal of establishing a “network of green hubs and squares where pedestrians have priority.” The superblocks were also touted as a way to help tackle air pollution, which is directly responsible for over 1,000 deaths in Barcelona each year…(More)”.