Reimagining the Request for Proposal


Article by Devon Davey, Heather Hiscox & Nicole Markwick : “In recent years, the social sector and the communities it serves have called for deep structural change to address our most serious social injustices. Yet one of the basic tools we use to fund change, the request for proposal (RFP), has remained largely unchanged. We believe that RFPs must become part of the larger call for systemic reform….

At first glance, the RFP process may seem neutral or fair. Yet RFPs are often designed by individuals in high-level positions without meaningful input from community members and frontline staff—those who are most familiar with social injustices and who often hold the least institutional power. What’s more, those who both issue and respond to RFPs often rely on their social capital to find and collaborate on RFP opportunities. Since social networks are highly homogeneous, RFP participation is limited to the professionals who have social connections to the issuer, resulting in a more limited pool of applicants.

This selection process is further compounded by the human propensity to hire people who look the same and who reflect similar ways of thinking. Social sector decision makers and power holders tend to be—among other identities—white. This lack of diversity, furthered by historical oppression, has ensured that white privilege and ways of working have come to dominate within the philanthropic and nonprofit sectors. This concentration of power and lack of diverse perspectives and experiences shaping RFPs results in projects failing to respond to the needs of communities and, in many cases, projects that directly perpetuate racism, colonialism, misogyny, ableism, sexism, and other forms of systemic and individual oppression.

The rigid structure of RFPs plays an important role in many of the negative outcomes of projects. Effective social change work is emergent, is iterative, and centers trust by nature. By contrast, RFPs frequently apply inflexible work scopes, limited timelines and budgets, and unproven solutions that are developed within the blinders of institutional power. Too often, funders force programs into implementation because they want to see results according to a specified plan. This rigidity can produce initiatives that are ineffective and removed from community needs. As consultant Joyce Lee-Ibarra says, “[RFPs] feel fundamentally transactional, when the work I want to do is relational.”…(More)”.

Six Prescriptions for Applied Behavioral Science as It Comes of Age


Article by Dilip Soman and Nina Mažar: “…But it has now been over 14 years since the publication of Nudge and more than 10 years since the first behavioral unit in government started functioning. While we have made a lot of progress as a field, we believe that the applied science is at a critical juncture. Our efforts at this stage will determine whether the field matures in a systematic and stable manner, or grows wildly and erratically. Unless we take stock of the science, the practice, and the mechanisms that we can put into place to align the two, we will run the danger of the promise of behavioral science being an illusion for many—not because the science itself was faulty, but because we did not successfully develop a science for using the science.  

We offer six prescriptions for how the field of applied behavioral science can better align itself so that it grows in a systematic and not in a wild manner. 

1. Offer a balanced and nuanced view of the promise of behavioral science 

We believe that it is incumbent on leaders in both the academic and applied space to offer a balanced view of the promise of behavioral science. While we understand that the nature of the book publication process or of public lectures tends to skew on additives to highlight success, we also believe that it is perhaps more of a contribution for the field to highlight limitations and nuances. Rather than narratives along the lines of “A causes B,” it would be helpful for our leaders to highlight narratives such as “A causes B in some conditions and C in others.” Dissemination of this new narrative could take the form of traditional knowledge mobilization tools, such as books, popular press articles, interviews, podcasts, and essays. Our recent coedited book, Behavioral Science in the Wildis one attempt at this.

2.Publish null and nonsurprising results 

Academic incentives usually create a body of work that (a) is replete with positive results, (b) overrepresents surprising results, (c) is not usually replicated, and (d) is focused on theory and phenomena and not on practical problems. As has been discussed elsewhere, this occurs because of the academic incentive structure, which favors surprising and positive results. We call on our field to change this culture by creating platforms that allow and encourage authors to publish null results, as well as unsurprising results…(More)”.

Omnivorous Analysis


Essay by Anne Lee Steele: “Satellite imagery has woven itself into the fabric of the internet. We recognize these crisp, high-definition, bird’s-eye-view images most commonly from Google Earth—but we employ them in much more besides: from reporting on stuck shipping containers to getting directions to a friend’s house, to tracking forest fires in real time and scrolling through real estate listings. Given their ever-widening range of commercial, consumer, and civic uses, it won’t surprise most people to hear that the industry that produces them (also known as Earth Observation, or EO) is growing at an exponential rate, and is only expected to expand further in the coming years.

Yet despite the prominence of satellite imagery in the geographical imagination of the internet, the imperatives of the industry are much less clear. The corporations that produce them are much less well known, and the military interests that back them remain as murky as ever. The highly visible commercial side of the industry is still deeply intertwined with its classified counterpart, and two companies, Maxar and Planet, have emerged to dominate the industry—supporting civilian functions with one hand, while supplying US defense needs with the other. 

Indeed, the ubiquity of commercial satellite imagery gives nearly anyone godlike powers of reconnaissance and surveillance not that far removed from those enjoyed by militaries and intelligence agencies—a fact that causes no small amount of anxiety within the Pentagon. The pervasiveness and power of their imagery compels us to ask: Where do they come from? And how are they being put to use?…(More)”.

Imagining Governance for Emerging Technologies


Essay by Debra J.H. Mathews, Rachel Fabi and Anaeze C. Offodile: “…How should such technologies be regulated and governed? It is increasingly clear that past governance structures and strategies are not up to the task. What these technologies require is a new governance approach that accounts for their interdisciplinary impacts and potential for both good and ill at both the individual and societal level. 

To help lay the groundwork for a novel governance framework that will enable policymakers to better understand these technologies’ cross-sectoral footprint and anticipate and address the social, legal, ethical, and governance issues they raise, our team worked under the auspices of the National Academy of Medicine’s Committee on Emerging Science, Technology, and Innovation in health and medicine (CESTI) to develop an analytical approach to technology impacts and governance. The approach is grounded in detailed case studies—including the vignettes about Robyn and Liam—which have informed the development of a set of guiding principles (see sidebar).

Based on careful analysis of past governance, these case studies also contain a plausible vision of what might happen in the future. They illuminate ethical issues and help reveal governance tools and choices that could be crucial to delivering social benefits and reducing or avoiding harms. We believe that the approach taken by the committee will be widely applicable to considering the governance of emerging health technologies. Our methodology and process, as we describe here, may also be useful to a range of stakeholders involved in governance issues like these…(More)”.

How We Can Encode Human Rights In The Blockchain


Essay by Nathan Schneider: “Imagine there is a new decentralized finance app quietly spreading around the world that’s like a payday lender from hell. Call it DevilsBridge. Rather than getting it from the App Store, you access its blockchain contracts directly, using a Web browser with a crypto-wallet plugin. DevilsBridge provides small loans in cryptocurrency that “bridge” people to the next paycheck. The interest rates are far below those of conventional payday lenders, which is life-changing for many users.

But if the payments go unpaid, they grow. They balloon. They reach multiples upon multiples of the principal. As time goes on, pressure ratchets up on borrowers, who become notorious for undertaking desperate, violent crimes to pay back their exorbitant debts. The deal, after all, is that if a debt reaches the magic threshold of $1 million, the debtor becomes a target. A private market of poison-dart-shooting drones receives a bounty to assassinate the mega-debtors.

Anywhere there are laws, of course, this is all wildly illegal. But nobody knows who created DevilDAO, the decentralized autonomous organization that operates DevilsBridge, or who its members are. The identities of the drone owners also hide behind cryptographic gibberish. Sometimes local police can trace the drones back to their bases, or investigators can trace a DevilDAO member’s address to a real person. But in most places where the assassinations happen, authorities are ill-equipped for airborne chases or scrutinizing blockchain analytics.

This may sound like a cartoonish scenario, but it’s freshly plausible thanks to the advent of decentralized, autonomous systems on blockchains. Ethereum co-founder Vitalik Buterin jokingly nodded to such dystopian possibilities in early 2014, when he listed possible uses for his proposed blockchain, from crop insurance to decentralized social networks — or perhaps, he said as he walked away from the mic, it could allow for the creation of Skynet, the robot intelligence in the “Terminator” movies that tries to exterminate the human race.

The potential for blockchain-enabled human-rights abuses is real. At the same time, these technologies introduce new ways of encoding and enforcing rights. Imagine the blockchain that DevilsBridge runs on introduces a software update. It bans any smart contract that kills humans. An anonymous investigator presents evidence of what the app is doing, and an anonymous jury confirms its validity; instantly, the contracts for DevilsBridge and DevilDAO no longer function…(More)”.

The linguistics search engine that overturned the federal mask mandate


Article by Nicole Wetsman: “The COVID-19 pandemic was still raging when a federal judge in Florida made the fateful decision to type “sanitation” into the search bar of the Corpus of Historical American English.

Many parts of the country had already dropped mask requirements, but a federal mask mandate on planes and other public transportation was still in place. A lawsuit challenging the mandate had come before Judge Kathryn Mizelle, a former clerk for Justice Clarence Thomas. The Biden administration said the mandate was valid, based on a law that authorizes the Centers for Disease Control and Prevention (CDC) to introduce rules around “sanitation” to prevent the spread of disease.

Mizelle took a textualist approach to the question — looking specifically at the meaning of the words in the law. But along with consulting dictionaries, she consulted a database of language, called a corpus, built by a Brigham Young University linguistics professor for other linguists. Pulling every example of the word “sanitation” from 1930 to 1944, she concluded that “sanitation” was used to describe actively making something clean — not as a way to keep something clean. So, she decided, masks aren’t actually “sanitation.”

The mask mandate was overturned, one of the final steps in the defanging of public health authorities, even as infectious disease ran rampant…

Using corpora to answer legal questions, a strategy often referred to as legal corpus linguistics, has grown increasingly popular in some legal circles within the past decade. It’s been used by judges on the Michigan Supreme Court and the Utah Supreme Court, and, this past March, was referenced by the US Supreme Court during oral arguments for the first time.

“It’s been growing rapidly since 2018,” says Kevin Tobia, a professor at Georgetown Law. “And it’s only going to continue to grow.”…(More)”.

In this small Va. town, citizens review police like Uber drivers


Article by Emily Davies: “Chris Ford stepped on the gas in his police cruiser and rolled down Gold Cup Drive to catch the SUV pushing 30 mph in a 15 mph zone. Eleven hours and 37 minutes into his shift, the corporal was ready for his first traffic stop of the day.

“Look at him being sneaky,” Fordsaid, his blue lights flashing on a quiet road in this small town where a busy day could mean animals escaped from a local slaughterhouse.

Ford parked, walked toward the SUV and greeted the man who had ignored the speed limit at exactly the wrong time.

“I was doing 15,” said the driver, a Black man in a mostly White neighborhood of a mostly White town.

The officertook his license and registration back to the cruiser.

“Every time I pull over someone of color, they’re standoffish with me. Like, ‘Here’s a White police officer, here we go again.’ ” Ford, 56, said. “So I just try to be nice.”

Ford knew the stop would be scrutinized — and not just by the reporter who was allowed to ride along on his shift.

After every significant encounter with residents, officers in Warrenton are required to hand out a QR code, which is on the back of their business card, asking for feedback on the interaction. Through a series of questions, citizens can use a star-based system to rate officers on their communication, listening skills and fairness. The responses are anonymous and can be completed any time after the interaction to encourage people to give honest assessments. The program, called Guardian Score, is supposed to give power to those stopped by police in a relationship that has historically felt one-sided — and to give police departments a tool to evaluate their force on more than arrests and tickets.

“If we started to measure how officers are treating community members, we realized we could actually infuse this into the overall evaluation process of individual officers,” said Burke Brownfeld, a founder of Guardian Score and a former police officer in Alexandria. “The definition of doing a good job could change. It would also include: How are your listening skills? How fairly are you treating people based on their perception?”…(More)”.

How science could aid the US quest for environmental justice


Jeff Tollefson at Nature: “…The network of US monitoring stations that detect air pollution catches only broad trends across cities and regions, and isn’t equipped for assessing air quality at the level of streets and neighbourhoods. So environmental scientists are exploring ways to fill the gaps.

In one project funded by NASA, researchers are developing methods to assess street-level pollution using measurements of aerosols and other contaminants from space. When the team trained its tools on Washington DC, the scientists found1 that sections in the city’s southeast, which have a larger share of Black residents, are exposed to much higher levels of fine-soot pollution than wealthier — and whiter — areas in the northwest of the city, primarily because of the presence of major roads and bus depots in the southeast.

Cumulative burden: Air-pollution levels tend to be higher in poorer and predominantly Black neighbourhoods of Washington DC.
Source: Ref. 1

The detailed pollution data painted a more accurate picture of the burden on a community that also lacks access to high-quality medical facilities and has high rates of cardiovascular disorders and other diseases. The results help to explain a more than 15-year difference in life expectancy between predominantly white neighbourhoods and some predominantly Black ones.

The analysis underscores the need to consider pollution and socio-economic data in parallel, says Susan Anenberg, director of the Climate and Health Institute at the George Washington University in Washington DC and co-leader of the project. “We can actually get neighbourhood-scale observations from space, which is quite incredible,” she says, “but if you don’t have the demographic, economic and health data as well, you’re missing a very important piece of the puzzle.”

Other projects, including one from technology company Aclima, in San Francisco, California, are focusing on ubiquitous, low-cost sensors that measure air pollution at the street level. Over the past few years, Aclima has deployed a fleet of vehicles to collect street-level data on air pollutants such as soot and greenhouse gases across 101 municipalities in the San Francisco Bay area. Their data have shown that air-pollution levels can vary as much as 800% from one neighbourhood block to the next.

Working directly with disadvantaged communities and environmental regulators in California, as well as with other states and localities, the company provides pollution monitoring on a subscription basis. It also offers the use of its screening tool, which integrates a suite of socio-economic data and can be used to assess cumulative impacts…(More)”.

Stories to Work By


Essay by William E. Spriggs: “In Charlie Chaplin’s 1936 film Modern Times, humans in a factory are reduced to adjuncts to a massive series of cogs and belts. Overlords bark commands from afar to a servant class, and Chaplin’s hapless hero is literally consumed by the machine … and then spit out by it. In the film, the bosses have all the power, and machines keep workers in check.

Modern Times’s dystopian narrative remains with us today. In particular, it is still held by many policymakers who assume that increasing technological progress, whether mechanical or informational, inevitably means that ordinary workers will lose. This view perpetuates itself when policies that could give workers more power in times of technological change are overlooked, while those that disempower workers are adopted. If we are to truly consider science policy for the future, we need to understand how this narrative about workers and technology functions, where it is misleading, and how deliberate policies can build a better world for all….

Today’s tales of pending technological dystopia—echoed in economics papers as well as in movies and news reports—blind us to the lessons we could glean from the massive disruptions of earlier periods of even greater change. Today the threat of AI is portrayed as revolutionary, and previous technological change as slow and inconsequential—but this was never the case. These narratives of technological inevitability limit the tools we have at our disposal to promote equality and opportunity.

The challenges we face today are far from insurmountable: technology is not destiny. Workers are not doomed to be Chaplin’s victim of technology with one toe caught in the gears of progress. We have choices, and the central challenge of science and technology policy for the next century will be confronting those choices head on. Policymakers should focus on the fundamental tasks of shaping how technology is deployed and enacting the economic rules we need to ensure that technology works for us all, rather than only the few….(More)”.

I tried to read all my app privacy policies. It was 1 million words.


Article by Geoffrey A. Fowler: “…So here’s an idea: Let’s abolish the notion that we’re supposed to read privacy policies.

I’m not suggesting companies shouldn’t have to explain what they’re up to. Maybe we call them “data disclosures” for the regulators, lawyers, investigative journalists and curious consumers to pore over.

But to protect our privacy, the best place to start is for companies to simply collect less data. “Maybe don’t do things that need a million words of explanation? Do it differently,” said Slaughter. “You can’t abuse, misuse, leverage data that you haven’t collected in the first place.”

Apps and services should only collect the information they really need to provide that service — unless we opt in to let them collect more, and it’s truly an option.

I’m not holding my breath that companies will do that voluntarily, but a federal privacy law would help. While we wait for one, Slaughter said the FTC (where Democratic commissioners recently gained a majority) is thinking about how to use its existing authority “to pursue practices — including data collection, use and misuse — that are unfair to users.”

Second, we need to replace the theater of pressing “agree” with real choices about our privacy.

Today, when we do have choices to make, companies often present them in ways that pressure us into making the worst decisions for ourselves.

Apps and websites should give us the relevant information and our choices in the moment when it matters. Twitter actually does this just-in-time notice better than many other apps and websites: By default, it doesn’t collect your exact location, and only prompts you to do so when you ask to tag your location in a tweet.

Even better, technology could help us manage our choices. Cranor suggests that data disclosures could be coded to be read by machines. Companies already do this for financial information, and the TLDR Act would require consistent tags on privacy information, too. Then your computer could act kind of like a butler, interacting with apps and websites on your behalf.

Picture Siri as a butler who quizzes you briefly about your preferences and then does your bidding. The privacy settings on an iPhone already let you tell all the different apps on your phone not to collect your location. For the past year, they’ve also allowed you to ask apps not to track you.

Web browsers could serve as privacy butlers, too. Mozilla’s Firefox already lets you block certain kinds of privacy invasions. Now a new technology called the Global Privacy Control is emerging that would interact with websites and instruct them not to “sell” our data. It’s grounded in California’s privacy law, which is among the toughest in the nation, though it remains to be seen how the state will enforce GPC…(More)”.