Wikipedia’s Moment of Truth


Article by Jon Gertner at the New York Times: “In early 2021, a Wikipedia editor peered into the future and saw what looked like a funnel cloud on the horizon: the rise of GPT-3, a precursor to the new chatbots from OpenAI. When this editor — a prolific Wikipedian who goes by the handle Barkeep49 on the site — gave the new technology a try, he could see that it was untrustworthy. The bot would readily mix fictional elements (a false name, a false academic citation) into otherwise factual and coherent answers. But he had no doubts about its potential. “I think A.I.’s day of writing a high-quality encyclopedia is coming sooner rather than later,” he wrote in “Death of Wikipedia,” an essay that he posted under his handle on Wikipedia itself. He speculated that a computerized model could, in time, displace his beloved website and its human editors, just as Wikipedia had supplanted the Encyclopaedia Britannica, which in 2012 announced it was discontinuing its print publication.

Recently, when I asked this editor — he asked me to withhold his name because Wikipedia editors can be the targets of abuse — if he still worried about his encyclopedia’s fate, he told me that the newer versions made him more convinced that ChatGPT was a threat. “It wouldn’t surprise me if things are fine for the next three years,” he said of Wikipedia, “and then, all of a sudden, in Year 4 or 5, things drop off a cliff.”..(More)”.

The Gutenberg Parenthesis: The Age of Print and Its Lessons for the Age of the Internet



Book by Jeff Jarvis: “The age of print is a grand exception in history. For five centuries it fostered what some call print culture – a worldview shaped by the completeness, permanence, and authority of the printed word. As a technology, print at its birth was as disruptive as the digital migration of today. Now, as the internet ushers us past print culture, journalist Jeff Jarvis offers important lessons from the era we leave behind.

To understand our transition out of the Gutenberg Age, Jarvis first examines the transition into it. Tracking Western industrialized print to its origins, he explores its invention, spread, and evolution, as well as the bureaucracy and censorship that followed. He also reveals how print gave rise to the idea of the mass – mass media, mass market, mass culture, mass politics, and so on – that came to dominate the public sphere.

What can we glean from the captivating, profound, and challenging history of our devotion to print? Could it be that we are returning to a time before mass media, to a society built on conversation, and that we are relearning how to hold that conversation with ourselves? Brimming with broader implications for today’s debates over communication, authorship, and ownership, Jarvis’ exploration of print on a grand scale is also a complex, compelling history of technology and power…(More)”

Shallowfakes


Essay by James R. Ostrowski: “…This dystopian fantasy, we are told, is what the average social media feed looks like today: a war zone of high-tech disinformation operations, vying for your attention, your support, your compliance. Journalist Joseph Bernstein, in his 2021 Harper’s piece “Bad News,” attributes this perception of social media to “Big Disinfo” — a cartel of think tanks, academic institutions, and prestige media outlets that spend their days spilling barrels of ink into op-eds about foreign powers’ newest disinformation tactics. The technology’s specific impact is always vague, yet somehow devastating. Democracy is dying, shot in the chest by artificial intelligence.

The problem with Big Disinfo isn’t that disinformation campaigns aren’t happening but that claims of mind-warping, AI-enabled propaganda go largely unscrutinized and often amount to mere speculation. There is little systematic public information about the scale at which foreign governments use deepfakes, bot armies, or generative text in influence ops. What little we know is gleaned through irregular investigations or leaked documents. In lieu of data, Big Disinfo squints into the fog, crying “Bigfoot!” at every oak tree.

Any machine learning researcher will admit that there is a critical disconnect between what’s possible in the lab and what’s happening in the field. Take deepfakes. When the technology was first developed, public discourse was saturated with proclamations that it would slacken society’s grip on reality. A 2019 New York Times op-ed, indicative of the general sentiment of this time, was titled “Deepfakes Are Coming. We Can No Longer Believe What We See.” That same week, Politico sounded the alarm in its article “‘Nightmarish’: Lawmakers brace for swarm of 2020 deepfakes.” A Forbes article asked us to imagine a deepfake video of President Trump announcing a nuclear weapons launch against North Korea. These stories, like others in the genre, gloss over questions of practicality…(More)”.

AI Is Tearing Wikipedia Apart


Article by Claire Woodcock: “As generative artificial intelligence continues to permeate all aspects of culture, the people who steward Wikipedia are divided on how best to proceed. 

During a recent community call, it became apparent that there is a community split over whether or not to use large language models to generate content. While some people expressed that tools like Open AI’s ChatGPT could help with generating and summarizing articles, others remained wary. 

The concern is that machine-generated content has to be balanced with a lot of human review and would overwhelm lesser-known wikis with bad content. While AI generators are useful for writing believable, human-like text, they are also prone to including erroneous information, and even citing sources and academic papers which don’t exist. This often results in text summaries which seem accurate, but on closer inspection are revealed to be completely fabricated

“The risk for Wikipedia is people could be lowering the quality by throwing in stuff that they haven’t checked,” Bruckman added. “I don’t think there’s anything wrong with using it as a first draft, but every point has to be verified.” 

The Wikimedia Foundation, the nonprofit organization behind the website, is looking into building tools to make it easier for volunteers to identify bot-generated content. Meanwhile, Wikipedia is working to draft a policy that lays out the limits to how volunteers can use large language models to create content.

The current draft policy notes that anyone unfamiliar with the risks of large language models should avoid using them to create Wikipedia content, because it can open the Wikimedia Foundation up to libel suits and copyright violations—both of which the nonprofit gets protections from but the Wikipedia volunteers do not. These large language models also contain implicit biases, which often result in content skewed against marginalized and underrepresented groups of people

The community is also divided on whether large language models should be allowed to train on Wikipedia content. While open access is a cornerstone of Wikipedia’s design principles, some worry the unrestricted scraping of internet data allows AI companies like OpenAI to exploit the open web to create closed commercial datasets for their models. This is especially a problem if the Wikipedia content itself is AI-generated, creating a feedback loop of potentially biased information, if left unchecked…(More)”.

DMA: rules for digital gatekeepers to ensure open markets start to apply


Press Release: “The EU Digital Markets Act (DMA) applies from today. Now that the DMA applies, potential gatekeepers that meet the quantitative thresholds established have until 3 July to notify their core platform services to the Commission. ..

The DMA aims to ensure contestable and fair markets in the digital sector. It defines gatekeepers as those large online platforms that provide an important gateway between business users and consumers, whose position can grant them the power to act as a private rule maker, and thus create a bottleneck in the digital economy. To address these issues, the DMA defines a series of specific obligations that gatekeepers will need to respect, including prohibiting them from engaging in certain behaviours in a list of do’s and don’ts. More information is available in the dedicated Q&A…(More)”.

Gaming Public Opinion


Article by Albert Zhang , Tilla Hoja & Jasmine Latimore: “The Chinese Communist Party’s (CCP’s) embrace of large-scale online influence operations and spreading of disinformation on Western social-media platforms has escalated since the first major attribution from Silicon Valley companies in 2019. While Chinese public diplomacy may have shifted to a softer tone in 2023 after many years of wolf-warrior online rhetoric, the Chinese Government continues to conduct global covert cyber-enabled influence operations. Those operations are now more frequent, increasingly sophisticated and increasingly effective in supporting the CCP’s strategic goals. They focus on disrupting the domestic, foreign, security and defence policies of foreign countries, and most of all they target democracies.

Currently—in targeted democracies—most political leaders, policymakers, businesses, civil society groups and publics have little understanding of how the CCP currently engages in clandestine activities online in their countries, even though this activity is escalating and evolving quickly. The stakes are high for democracies, given the indispensability of the internet and their reliance on open online spaces, free from interference. Despite years of monitoring covert CCP cyber-enabled influence operations by social-media platforms, governments, and research institutes such as ASPI, definitive public attribution of the actors driving these activities is rare. Covert online operations, by design, are difficult to detect and attribute to state actors. 

Social-media platforms and governments struggle to devote adequate resources to identifying, preventing and deterring increasing levels of malicious activity, and sometimes they don’t want to name and shame the Chinese Government for political, economic and/or commercial reasons…(More)”.

The Synchronized Society: Time and Control From Broadcasting to the Internet


Book by Randall Patnode: “…traces the history of the synchronous broadcast experience of the twentieth century and the transition to the asynchronous media that dominate today. Broadcasting grew out of the latent desire by nineteenth-century industrialists, political thinkers, and social reformers to tame an unruly society by controlling how people used their time. The idea manifested itself in the form of the broadcast schedule, a managed flow of information and entertainment that required audiences to be in a particular place – usually the home – at a particular time and helped to create “water cooler” moments, as audiences reflected on their shared media texts. Audiences began disconnecting from the broadcast schedule at the end of the twentieth century, but promoters of social media and television services still kept audiences under control, replacing the schedule with surveillance of media use. Author Randall Patnode offers compelling new insights into the intermingled roles of broadcasting and industrial/post-industrial work and how Americans spend their time…(More)”.

Exploring data journalism practices in Africa: data politics, media ecosystems and newsroom infrastructures


Paper by Sarah Chiumbu and Allen Munoriyarwa: “Extant research on data journalism in Africa has focused on newsroom factors and the predilections of individual journalists as determinants of the uptake of data journalism on the continent. This article diverts from this literature by examining the slow uptake of data journalism in sub- Saharan Africa through the prisms of non-newsroom factors. Drawing on in-depth interviews with prominent investigative journalists sampled from several African countries, we argue that to understand the slow uptake of data journalism on the continent; there is a need to critique the role of data politics, which encompasses state, market and existing media ecosystems across the continent. Therefore, it is necessary to move beyond newsroom-centric factors that have dominated the contemporary understanding of data journalism practices. A broader, non-newsroom conceptualisation beyond individual journalistic predilections and newsroom resources provides productive clarity on data journalism’s slow uptake on the continent. These arguments are made through the conceptual prisms of materiality, performativity and reflexivity…(More)”.

The Meta Oversight Board’s First Term


Paper by Evelyn Douek: “The Meta Oversight Board was established to oversee one of the most expansive systems of speech regulation in history and to exercise independent review over “some of the most difficult and significant
content decisions” Meta makes. As a voluntary exercise in selfregulation, the Board exercises power over Meta only insofar and for as long as Meta permits it to. And yet, in its inaugural members’ first threeyear term, the Board has in many ways defied its skeptics. The Board has established itself as a regular part of conversations about content moderation governance, receiving significant academic and media attention. It has also instantiated meaningful reforms of Meta’s content moderation systems, and shed light on otherwise completely opaque decisionmaking processes within one of the world’s most powerful
speech regulators. But the Board has also consistently shied away from answering the hardest and most controversial questions that come before it—that is, the very questions it was set up to solve. Although the Board purported to evaluate Meta’s rules under international human rights law, it has almost entirely failed to engage with the necessary the normative question of how international law principles created to constrain governmental power over expression should apply to private content moderation systems. This Essay argues that the Board’s institutional incentives and desire for influence have made it prioritize consensus and simplicity over engagement with the fundamental normative questions that the quest for principled content moderation decisionmaking raises. The result is a tremendous missed opportunity that holds important lessons for the design of future content moderation oversight bodies…(More)”

Foolproof: Why We Fall for Misinformation and How to Build Immunity


Book by Sander van der Linden: “From fake news to conspiracy theories, from pandemics to politics, misinformation may be the defining problem of our era. Like a virus, misinformation infects our minds – altering our beliefs and replicating at astonishing rates. Once the virus takes hold, our primary strategies of fact-checking and debunking are an insufficient cure.

In Foolproof Sander van der Linden describes how to inoculate yourself and others against the spread of misinformation, discern fact from fiction and push back against methods of mass persuasion.

Everyone is susceptible to fake news. There are polarising narratives in society, conspiracy theories are rife, fake experts dole out misleading advice and accuracy is often lost in favour of sensationalist headlines. So how and why does misinformation spread if we’re all aware of its existence? And, more importantly, what can we do about it?…(More)”.