Making forest data fair and open


Paper by Renato A. F. de Lima : “It is a truth universally acknowledged that those in possession of time and good fortune must be in want of information. Nowhere is this more so than for tropical forests, which include the richest and most productive ecosystems on Earth. Information on tropical forest carbon and biodiversity, and how these are changing, is immensely valuable, and many different stakeholders wish to use data on tropical and subtropical forests. These include scientists, governments, nongovernmental organizations and commercial interests, such as those extracting timber or selling carbon credits. Another crucial, often-ignored group are the local communities for whom forest information may help to assert their rights and conserve or restore their forests.

A widespread view is that to lead to better public outcomes it is necessary and sufficient for forest data to be open and ‘Findable, Accessible, Interoperable, Reusable’ (FAIR). There is indeed a powerful case. Open data — those that anyone can use and share without restrictions — can encourage transparency and reproducibility, foster innovation and be used more widely, thus translating into a greater public good (for example, https://creativecommons.org). Open biological collections and genetic sequences such as GBIF or GenBank have enabled species discovery, and open Earth observation data helps people to understand and monitor deforestation (for example, Global Forest Watch). But the perspectives of those who actually make the forest measurements are much less recognized, meaning that open and FAIR data can be extremely unfair indeed. We argue here that forest data policies and practices must be fair in the correct, linguistic use of the term — just and equitable.

In a world in which forest data origination — measuring, monitoring and sustaining forest science — is secured by large, long-term capital investment (such as through space missions and some officially supported national forest inventories), making all data open makes perfect sense. But where data origination depends on insecure funding and precarious employment conditions, top-down calls to make these data open can be deeply problematic. Even when well-intentioned, such calls ignore the socioeconomic context of the places where the forest plots are located and how knowledge is created, entrenching the structural inequalities that characterize scientific research and collaboration among and within nations. A recent review found scant evidence for open data ever lessening such inequalities. Clearly, only a privileged part of the global community is currently able to exploit the potential of open forest data. Meanwhile, some local communities are de facto owners of their forests and associated knowledge, so making information open — for example, the location of valuable species — may carry risks to themselves and their forests….(More)”.

Cities Take the Lead in Setting Rules Around How AI Is Used


Jackie Snow at the Wall Street Journal: “As cities and states roll out algorithms to help them provide services like policing and traffic management, they are also racing to come up with policies for using this new technology.

AI, at its worst, can disadvantage already marginalized groups, adding to human-driven bias in hiring, policing and other areas. And its decisions can often be opaque—making it difficult to tell how to fix that bias, as well as other problems. (The Wall Street Journal discussed calls for regulation of AI, or at least greater transparency about how the systems work, with three experts.)

Cities are looking at a number of solutions to these problems. Some require disclosure when an AI model is used in decisions, while others mandate audits of algorithms, track where AI causes harm or seek public input before putting new AI systems in place.

Here are some ways cities are redefining how AI will work within their borders and beyond.

Explaining the algorithms: Amsterdam and Helsinki

One of the biggest complaints against AI is that it makes decisions that can’t be explained, which can lead to complaints about arbitrary or even biased results.

To let their citizens know more about the technology already in use in their cities, Amsterdam and Helsinki collaborated on websites that document how each city government uses algorithms to deliver services. The registry includes information on the data sets used to train an algorithm, a description of how an algorithm is used, how public servants use the results, the human oversight involved and how the city checks the technology for problems like bias.

Amsterdam has six algorithms fully explained—with a goal of 50 to 100—on the registry website, including how the city’s automated parking-control and trash-complaint reports work. Helsinki, which is only focusing on the city’s most advanced algorithms, also has six listed on its site, with another 10 to 20 left to put up.

“We needed to assess the risk ourselves,” says Linda van de Fliert, an adviser at Amsterdam’s Chief Technology Office. “And we wanted to show the world that it is possible to be transparent.”…(More)” See also AI Localism: The Responsible Use and Design of Artificial Intelligence at the Local Level

The Power of Narrative


Essay by Klaus Schwab and Thierry Mallerett: “…The expression “failure of imagination” captures this by describing the expectation that future opportunities and risks will resemble those of the past. Novelist Graham Greene used it in The Power and the Glory, but the 9/11 Commission made it popular by invoking it as the main reason why intelligence agencies had failed to anticipate the “unimaginable” events of that day.

Ever since, the expression has been associated with situations in which strategic thinking and risk management are stuck in unimaginative and reactive thinking. Considering today’s wide and interdependent array of risks, we can’t afford to be unimaginative, even though, as the astrobiologist Caleb Scharf points out, we risk getting imprisoned in a dangerous cognitive lockdown because of the magnitude of the task. “Indeed, we humans do seem to struggle in general when too many new things are thrown at us at once. Especially when those things are outside of our normal purview. Like, well, weird viruses or new climate patterns,” Scharf writes. “In the face of such things, we can simply go into a state of cognitive lockdown, flipping from one small piece of the problem to another and not quite building a cohesive whole.”

Imagination is precisely what is required to escape a state of “cognitive lockdown” and to build a “cohesive whole.” It gives us the capacity to dream up innovative solutions to successfully address the multitude of risks that confront us. For decades now, we’ve been destabilizing the world, having failed to imagine the consequences of our actions on our societies and our biosphere, and the way in which they are connected. Now, following this failure and the stark realization of what it has entailed, we need to do just the opposite: rely on the power of imagination to get us out of the holes we’ve dug ourselves into. It is incumbent upon us to imagine the contours of a more equitable and sustainable world. Imagination being boundless, the variety of social, economic, and political solutions is infinite.

With respect to the assertion that there are things we don’t imagine to be socially or politically possible, a recent book shows that nothing is preordained. We are in fact only bound by the power of our own imaginations. In The Dawn of Everything, David Graeber and David Wengrow (an anthropologist and an archaeologist) prove this by showing that every imaginable form of social and economic organization has existed from the very beginning of humankind. Over the past 300,000 years, we’ve pursued knowledge, experimentation, happiness, development, freedom, and other human endeavors in myriad different ways. During these times that preceded our modern world, none of the arrangements that we devised to live together exhibited a single point of origin or an invariant pattern. Early societies were peaceful and violent, authoritarian and democratic, patriarchal and matriarchal, slaveholding and abolitionist, some moving between different types of organizations all the time, others not. Antique industrial cities were flourishing at the heart of empires while others existed in the absence of a sovereign entity…(More)”

Opening up Science—to Skeptics


Essay by Rohan R. Arcot  and Hunter Gehlbach: “Recently, the soaring trajectory of science skepticism seems to be rivaled only by global temperatures. Empirically established facts—around vaccines, elections, climate science, and the like—face potent headwinds. Despite the scientific consensus on these issues, much of the public remains unconvinced. In turn, science skepticism threatens our health, the health of our democracy, and the health of our planet.  

The research community is no stranger to skepticism. Its own members have been questioning the integrity of many scientific findings with particular intensity of late. In response, we have seen a swell of open science norms and practices, which provide greater transparency about key procedural details of the research process, mitigating many research skeptics’ misgivings. These open practices greatly facilitate how science is communicated—but only between scientists. 

Given the present historical moment’s critical need for science, we wondered: What if scientists allowed skeptics in the general public to look under the hood at how their studies were conducted? Could opening up the basic ideas of open science beyond scholars help combat the epidemic of science skepticism?  

Intrigued by this possibility, we sought a qualified skeptic and returned to Rohan’s father. If we could chaperone someone through a scientific journey—a person who could vicariously experience the key steps along the way—could our openness assuage their skepticism?…(More)”.

How to use the Civil Society Foresight report


Report by Dominique Barron, Rachel Coldicutt, Stephanie Pau, Anna Williams: “This report is for anyone making plans for the future.  In particular, we hope it will be a useful strategic tool for funders, civil society organisations, and policymakers who are developing strategies for long-term change. 

What is it? 

A way of looking ahead that makes it easier to see past the overwhelming present and focus on creating longer-term change. 

It highlights what is missing now; what is too dominant; and it shows that innovation is something driven by people, not technologies. 

How was it created?

The scenarios here were produced through a relational process. (More on that process in our report “A Constellation of Possible Futures”.) Our team brought together thirteen civil society leaders with lived, learned and practice experience; introduced them to some of the “official” futures created by management consultancies, trade bodies and banks (reports that focus on things like retail and transport and financial capital); and we then all participated in a workshop process that took us to 2036 and beyond. 

What does it cover?

The concepts explored in this report include the notion of care in a climate-altered world; a sketch of what happens when a nation welcomes migrants at scale; the psychological toll of social division; and the possible outcomes of technological breakdown. The outputs focussed on ways to reduce fear, overcome entrenched barriers, and increase spirituality and belonging. 

Importantly, the process never asked for agreement or utopia; instead, it held a space for tension, disagreement, and pragmatism. And it surfaced the strategic knowledge of expertise of people in civil society with a wide range of experience…(More)”.

Regulatory Technology for the 21st Century


White Paper by the World Economic Forum: “Regulation is central to government’s management of complex systems. However, if designed or applied ineffectively, regulation may trigger significant losses, impose unnecessary financial burdens and stifle innovation. Regulatory Technology (RegTech), is the application of new technological solutions to in set, effectuate and meet regulatory requirements. This white paper explores the value of RegTech through a series of case studies and identifies the 7 common success factors that help define best practice deployment of RegTech. It provides government and business with a roadmap to start implementing RegTech without having to upend or rewrite entire regulatory and compliance frameworks to begin the journey…(More)”.

Internet ‘algospeak’ is changing our language in real time, from ‘nip nops’ to ‘le dollar bean’


Article by Taylor Lorenz: “Algospeak” is becoming increasingly common across the Internet as people seek to bypass content moderation filters on social media platforms such as TikTok, YouTube, Instagram and Twitch.

Algospeak refers to code words or turns of phrase users have adopted in an effort to create a brand-safe lexicon that will avoid getting their posts removed or down-ranked by content moderation systems. For instance, in many online videos, it’s common to say “unalive” rather than “dead,” “SA” instead of “sexual assault,” or “spicy eggplant” instead of “vibrator.”

As the pandemic pushed more people to communicate and express themselves online, algorithmic content moderation systems have had an unprecedented impact on the words we choose, particularly on TikTok, and given rise to a new form of internet-driven Aesopian language.

Unlike other mainstream social platforms, the primary way content is distributed on TikTok is through an algorithmically curated “For You” page; having followers doesn’t guarantee people will see your content. This shift has led average users to tailor their videos primarily toward the algorithm, rather than a following, which means abiding by content moderation rules is more crucial than ever.

When the pandemic broke out, people on TikTok and other apps began referring to it as the “Backstreet Boys reunion tour” or calling it the “panini” or “panda express” as platforms down-ranked videos mentioning the pandemic by name in an effort to combat misinformation. When young people began to discuss struggling with mental health, they talked about “becoming unalive” in order to have frank conversations about suicide without algorithmic punishment. Sex workers, who have long been censored by moderation systems, refer to themselves on TikTok as “accountants” and use the corn emoji as a substitute for the word “porn.”

As discussions of major events are filtered through algorithmic content delivery systems, more users are bending their language. Recently, in discussing the invasion of Ukraine, people on YouTube and TikTok have used the sunflower emoji to signify the country. When encouraging fans to follow them elsewhere, users will say “blink in lio” for “link in bio.”

Euphemisms are especially common in radicalized or harmful communities. Pro-anorexia eating disorder communities have long adopted variations on moderated words to evade restrictions. One paper from the School of Interactive Computing, Georgia Institute of Technology found that the complexity of such variants even increased over time. Last year, anti-vaccine groups on Facebook began changing their names to “dance party” or “dinner party” and anti-vaccine influencers on Instagram used similar code words, referring to vaccinated people as “swimmers.”

Tailoring language to avoid scrutiny predates the Internet. Many religions have avoided uttering the devil’s name lest they summon him, while people living in repressive regimes developed code words to discuss taboo topics…(More)”.

Orientation Failure? Why Directionality Matters in Innovation Policy and Implementation


Blog by Mariam Tabatadze and Benjamin Kumpf: “…In the essay “The Moon and the Ghetto” from 1977, Richard Nelson brought renewed attention to the question of directionality of innovation. He asked why societies that are wealthy and technologically advanced are not able to deal effectively with social problems such as poverty or inequities in education. Nelson believed that politics are only a small part of the problem. The main challenge, according to him, was further advancing scientific and technological breakthroughs.

Since the late seventies, humanity has laid claim to many more significant technological and scientific achievements. However, challenges such as poverty, social inequalities and of course environmental degradation persist. This begs the question: is the main problem a lack of directionality?

The COVID-19 pandemic sparked renewed interest in mission-driven innovation in industrial and socio-economic policy (see below for a framing of missions and mission-oriented innovation). The focus is a continuation of a “normative turn” in national and supranational science, technology and innovation (STI) policies over the last 15 years.

The directionality of STI policies shifted from pursuing predominantly growth and competitiveness-related objectives to addressing societal challenges. It brings together elements of innovation policy – focused on economic growth – and transition policy, which seeks beneficial change for society at large. This is important as we are seeing increasingly more evidence on the negative effects of innovation in countries across the globe, from exacerbated inequalities between places to greater inequalities between income groups…(More)”.

A 630-Billion-Word Internet Analysis Shows ‘People’ Is Interpreted as ‘Men’


Dana G. Smith at Scientific American: “A massive linguistic analysis of more than half a trillion words concludes that we assign gender to words that, by their very definition, should be gender-neutral.

Psychologists at New York University analyzed text from nearly three billion Web pages and compared how often words for person (“individual,” “people,” and so on) were associated with terms for a man (“male,” “he”) or a woman (“female,” “she”). They found that male-related words overlapped with “person” more frequently than female words did. The cultural concept of a person, from this perspective, is more often a man than a woman, according to the study, which was published on April 1 in Science Advances.

To conduct the study, the researchers turned to an enormous open-source data set of Web pages called the Common Crawl, which pulls text from everything from corporate white papers to Internet discussion forums. For their analysis of the text—a total of more than 630 billion words—the researchers used word embeddings, a computational linguistic technique that assesses how similar two words are by looking for how often they appear together.

“You can take a word like the word ‘person’ and understand what we mean by ‘person,’ how we represent the word ‘person,’ by looking at the other words that we often use around the word ‘person,’” explains April Bailey, a postdoctoral researcher at N.Y.U., who conducted the study. “We found that there was more overlap between the words for people and words for men than words for people and the words for women…, suggesting that there is this male bias in the concept of a person.”

Scientists have previously studied gender bias in language, such as the idea that women are more closely associated with family and home life and that men are more closely linked with work. “But this is the first to study this really general gender stereotype—the idea that men are sort of the default humans—in this quantitative computational social science way,” says Molly Lewis, a research scientist at the psychology department at Carnegie Mellon University, who was not involved in the study….(More)”.

The rise of the data steward


Article by Sarah Wray: “As data use and collaboration become more advanced, there is a need for a new profession within the public and private sectors, says Stefaan Verhulst, Co-Founder and Chief Research and Development Officer at New York University’s The GovLab. He calls this role the ‘data steward’ and is also seeking to expand existing definitions of the term.

While many cities, government organisations, and private sector companies have chief data officers and chief privacy officers, Verhulst says this new function is broader and necessary as more organisations begin to explore data collaborations which bring together data from various sources to solve problems for the public good.

Many cities, for instance, want to get more value and innovation from the open data they share, and are also increasingly partnering to benefit from private sector data on mobility, spending, and more.

Several examples highlight the challenges, though. There have been disputes about data-sharing and privacy, such as between Uber and the Los Angeles Department of Transportation, while other initiatives have failed to gain traction. Copenhagen’s City Data Exchange facilitated the exchange of public and private data but was disbanded after it struggled to get enough data providers and users on the platform and to become financially sustainable.

Verhulst says that beyond ensuring the security and integrity of data, new skills required by data stewards include the ability to secure partnerships, adequately vet data partners and set up data-sharing agreements, as well as the capacity to steward data-sharing initiatives internally and obtain legal and executive buy-in. Data stewards should also develop financial models for data-sharing to ensure partnerships are sustainable over time.

“That’s quite often ignored,” says Verhulst. “It’s assumed that these things will pay for themselves. Well surprise, surprise, there are costs.”

In addition, there’s an important role for retaining an active focus on insights from data and problems to be solved. Many early open data efforts have taken a ‘build it and they will come’ approach, and usage at scale hasn’t always materialised.

A dynamic regulatory environment is also driving demand for new skills, says Verhulst, noting that the proposed EU Data Act indicates a mandate “to knock on the doors of the private sector [for data] in emergency contexts”.

“The question is: how do you go about that?” Verhulst comments. “Many organisations are going to have to figure this out.”

The GovLab is now running the third cohort of its training for data stewards, and the first focused in the Eastern Hemisphere.

The Developing a Data Reuse Strategy for Public Problems course is part of The GovLab’s Open Data Policy Lab, which is supported by Microsoft..(More)”.