Why is it so hard to establish the death toll?


Article by Smriti Mallapaty: “Given the uncertainty of counting fatalities during conflict, researchers use other ways to estimate mortality.

One common method uses household surveys, says Debarati Guha-Sapir, an epidemiologist who specializes in civil conflicts at the University of Louvain in Louvain-la-Neuve, Belgium, and is based in Brussels. A sample of the population is asked how many people in their family have died over a specific period of time. This approach has been used to count deaths in conflicts elsewhere, including in Iraq3 and the Central African Republic4.

The situation in Gaza right now is not conducive to a survey, given the level of movement and displacement, say researchers. And it would be irresponsible to send data collectors into an active conflict and put their lives at risk, says Ball.

There are also ethical concerns around intruding on people who lack basic access to food and medication to ask about deaths in their families, says Jamaluddine. Surveys will have to wait for the conflict to end and movement to ease, say researchers.

Another approach is to compare multiple independent lists of fatalities and calculate mortality from the overlap between them. The Human Rights Data Analysis Group used this approach to estimate the number of people killed in Syria between 2011 and 2014. Jamaluddine hopes to use the ministry fatality data in conjunction with those posted on social media by several informal groups to estimate mortality in this way. But Guha-Sapir says this method relies on the population being stable and not moving around, which is often not the case in conflict-affected communities.

In addition to deaths immediately caused by the violence, some civilians die of the spread of infectious diseases, starvation or lack of access to health care. In February, Jamaluddine and her colleagues used modelling to make projections of excess deaths due to the war and found that, in a continued scenario of six months of escalated conflict, 68,650 people could die from traumatic injuries, 2,680 from non-communicable diseases such as cancer and 2,720 from infectious diseases — along with thousands more if an epidemic were to break out. On 30 July, the ministry declared a polio epidemic in Gaza after detecting the virus in sewage samples, and in mid-August it confirmed the first case of polio in 25 years, in a 10-month-old baby…

The longer the conflict continues, the harder it will be to get reliable estimates, because “reports by survivors get worse as time goes by”, says Jon Pedersen, a demographer at !Mikro in Oslo, who advises international agencies on mortality estimates…(More)”.

Germany’s botched data revamp leaves economists ‘flying blind’


Article by Olaf Storbeck: “Germany’s statistical office has suspended some of its most important indicators after botching a data update, leaving citizens and economists in the dark at a time when the country is trying to boost flagging growth.

In a nation once famed for its punctuality and reliability, even its notoriously diligent beancounters have become part of a growing perception that “nothing works any more” as Germans moan about delayed trains, derelict roads and bridges, and widespread staff shortages.

“There used to be certain aspects in life that you could just rely on, and the fact that official statistics are published on time was one of them — not any more,” said Jörg Krämer, chief economist of Commerzbank, adding that the suspended data was also closely watched by monetary policymakers and investors.

Since May the Federal Statistical Office (Destatis) has not updated time-series data for retail and wholesale sales, as well as revenue from the services sector, hospitality, car dealers and garages.

These indicators, which are published monthly and adjusted for seasonal changes, are a key component of GDP and crucial for assessing consumer demand in the EU’s largest economy.

Private consumption accounted for 52.7 per cent of German output in 2023. Retail sales made up 28 per cent of private consumption but shrank 3.4 per cent from a year earlier. Overall GDP declined 0.3 per cent last year, Destatis said.

The Wiesbaden-based authority, which was established in 1948, said the outages had been caused by IT issues and a complex methodological change in EU business statistics in a bid to boost accuracy.

Destatis has been working on the project since the EU directive in 2019, and the deadline for implementing the changes is December.

But a series of glitches, data issues and IT delays meant Destatis has been unable to publish retail sales and other services data for four months.

A key complication is that the revenues of companies that operate in both services and manufacturing will now be reported differently for each sector. In the past, all revenue was treated as either services or manufacturing, depending on which unit was bigger…(More)”

Problem-solving matter


Essay by David C Krakauer and Chris Kempes: “What makes computation possible? Seeking answers to that question, a hardware engineer from another planet travels to Earth in the 21st century. After descending through our atmosphere, this extraterrestrial explorer heads to one of our planet’s largest data centres, the China Telecom-Inner Mongolia Information Park, 470 kilometres west of Beijing. But computation is not easily discovered in this sprawling mini-city of server farms. Scanning the almost-uncountable transistors inside the Information Park, the visiting engineer might­ be excused for thinking that the answer to their question lies in the primary materials driving computational processes: silicon and metal oxides. After all, since the 1960s, most computational devices have relied on transistors and semiconductors made from these metalloid materials.

If the off-world engineer had visited Earth several decades earlier, before the arrival of metal-oxide transistors and silicon semiconductors, they might have found entirely different answers to their question. In the 1940s, before silicon semiconductors, computation might appear as a property of thermionic valves made from tungsten, molybdenum, quartz and silica – the most important materials used in vacuum tube computers.

And visiting a century earlier, long before the age of modern computing, an alien observer might come to even stranger conclusions. If they had arrived in 1804, the year the Jacquard loom was patented, they might have concluded that early forms of computation emerged from the plant matter and insect excreta used to make the wooden frames, punch cards and silk threads involved in fabric-weaving looms, the analogue precursors to modern programmable machines.

But if the visiting engineer did come to these conclusions, they would be wrong. Computation does not emerge from silicon, tungsten, insect excreta or other materials. It emerges from procedures of reason or logic.

This speculative tale is not only about the struggles of an off-world engineer. It is also an analogy for humanity’s attempts to answer one of our most difficult problems: life. For, just as an alien engineer would struggle to understand computation through materials, so it is with humans studying our distant origins…(More)”.

Use GenAI to Improve Scenario Planning


Article by Daniel J. Finkenstadt et al: “Businesses are increasingly leveraging strategic foresight and scenario planning to navigate uncertainties stemming from climate change, global conflicts, and technological advancements. Traditional methods, however, struggle with identifying key trends, exploring multiple scenarios, and providing actionable guidance. Generative AI offers a robust alternative, enabling rapid, cost-effective, and comprehensive contingency planning. This AI-driven approach enhances scenario creation, narrative exploration, and strategy generation, providing detailed, adaptable strategies rather than conclusive solutions. This approach demands accurate, relevant data and encourages iterative refinement, transforming how organizations forecast and strategize for the future…(More)”.

We finally have a definition for open-source AI


Article by Rhiannon Williams and James O’Donnell: “Open-source AI is everywhere right now. The problem is, no one agrees on what it actually is. Now we may finally have an answer. The Open Source Initiative (OSI), the self-appointed arbiters of what it means to be open source, has released a new definition, which it hopes will help lawmakers develop regulations to protect consumers from AI risks. 

Though OSI has published much about what constitutes open-source technology in other fields, this marks its first attempt to define the term for AI models. It asked a 70-person group of researchers, lawyers, policymakers, and activists, as well as representatives from big tech companies like Meta, Google, and Amazon, to come up with the working definition. 

According to the group, an open-source AI system can be used for any purpose without the need to secure permission, and researchers should be able to inspect its components and study how the system works.

It should also be possible to modify the system for any purpose—including to change its output—and to share it with others to usewith or without modificationsfor any purpose. In addition, the standard attempts to define a level of transparency for a given model’s training data, source code, and weights. 

The previous lack of an open-source standard presented a problem…(More)”.

Revisiting the ‘Research Parasite’ Debate in the Age of AI


Article by C. Brandon Ogbunu: “A 2016 editorial published in the New England Journal of Medicine lamented the existence of “research parasites,” those who pick over the data of others rather than generating new data themselves. The article touched on the ethics and appropriateness of this practice. The most charitable interpretation of the argument centered around the hard work and effort that goes into the generation of new data, which costs millions of research dollars and takes countless person-hours. Whatever the merits of that argument, the editorial and its associated arguments were widely criticized.

Given recent advances in AI, revisiting the research parasite debate offers a new perspective on the ethics of sharing and data democracy. It is ironic that the critics of research parasites might have made a sound argument — but for the wrong setting, aimed at the wrong target, at the wrong time. Specifically, the large language models, or LLMs, that underlie generative AI tools such as OpenAI’s ChatGPT, have an ethical challenge in how they parasitize freely available data. These discussions bring up new conversations about data security that may undermine, or at least complicate, efforts at openness and data democratization.

The backlash to that 2016 editorial was swift and violent. Many arguments centered around the anti-science spirit of the message. For example, metanalysis – which re-analyzes data from a selection of studies – is a critical practice that should be encouraged. Many groundbreaking discoveries about the natural world and human health have come from this practice, including new pictures of the molecular causes of depression and schizophrenia. Further, the central criticisms of research parasitism undermine the ethical goals of data sharing and ambitions for open science, where scientists and citizen-scientists can benefit from access to data. This differs from the status quo in 2016, when data published in many of the top journals of the world were locked behind a paywall, illegible, poorly labeled, or difficult to use. This remains largely true in 2024…(More)”.

Definitions, digital, and distance: on AI and policymaking


Article by Gavin Freeguard: “Our first question is less, ‘to what extent can AI improve public policymaking?’, but ‘what is currently wrong with policymaking?’, and then, ‘is AI able to help?’.

Ask those in and around policymaking about the problems and you’ll get a list likely to include:

  • the practice not having changed in decades (or centuries)
  • it being an opaque ‘dark art’ with little transparency
  • defaulting to easily accessible stakeholders and evidence
  • a separation between policy and delivery (and digital and other disciplines), and failure to recognise the need for agility and feedback as opposed to distinct stages
  • the challenges in measuring or evaluating the impact of policy interventions and understanding what works, with a lack of awareness, let alone sharing, of case studies elsewhere
  • difficulties in sharing data
  • the siloed nature of government complicating cross-departmental working
  • policy asks often being dictated by politics, with electoral cycles leading to short-termism, ministerial churn changing priorities and personal style, events prompting rushed reactions, or political priorities dictating ‘policy-based evidence making’
  • a rush to answers before understanding the problem
  • definitional issues about what policy actually is making it hard to get a hold of or develop professional expertise.  

If we’re defining ‘policy’ and the problem, we also need to define ‘AI’, or at least acknowledge that we are not only talking about new, shiny generative AI, but a world of other techniques for automating processes and analysing data that have been used in government for years.

So is ‘AI’ able to help? It could support us to make better use of a wider range of data more quickly; but it could privilege that which is easier to measure, strip data of vital context, and embed biases and historical assumptions. It could ‘make decisions more transparent (perhaps through capturing digital records of the process behind them, or by visualising the data that underpins a decision)’; or make them more opaque with ‘black-box’ algorithms, and distract from overcoming the very human cultural problems around greater openness. It could help synthesise submissions or generate ideas to brainstorm; or fail to compensate for deficiencies in underlying government knowledge infrastructure, and generate gibberish. It could be a tempting silver bullet for better policy; or it could paper over the cracks, while underlying technical, organisational and cultural plumbing goes unfixed. It could have real value in some areas, or cause harms in others…(More)”.

New AI standards group wants to make data scraping opt-in


Article by Kate Knibbs: “The first wave of major generative AI tools largely were trained on “publicly available” data—basically, anything and everything that could be scraped from the Internet. Now, sources of training data are increasingly restricting access and pushing for licensing agreements. With the hunt for additional data sources intensifying, new licensing startups have emerged to keep the source material flowing.

The Dataset Providers Alliance, a trade group formed this summer, wants to make the AI industry more standardized and fair. To that end, it has just released a position paper outlining its stances on major AI-related issues. The alliance is made up of seven AI licensing companies, including music copyright-management firm Rightsify, Japanese stock-photo marketplace Pixta, and generative-AI copyright-licensing startup Calliope Networks. (At least five new members will be announced in the fall.)

The DPA advocates for an opt-in system, meaning that data can be used only after consent is explicitly given by creators and rights holders. This represents a significant departure from the way most major AI companies operate. Some have developed their own opt-out systems, which put the burden on data owners to pull their work on a case-by-case basis. Others offer no opt-outs whatsoever…(More)”.

Policies must be justified by their wellbeing-to-cost ratio


Article by Richard Layard: “…What is its value for money — that is, how much wellbeing does it deliver per (net) pound it costs the government? This benefit/cost ratio (or BCR) should be central to every discussion.

The science exists to produce these numbers and, if the British government were to require them of the spending departments, it would be setting an example of rational government to the whole world.

Such a move would, of course, lead to major changes in priorities. At the London School of Economics we have been calculating the benefits and costs of policies across a whole range of government departments.

In our latest report on value for money, the best policies are those that save the government more money than they cost — for example by getting people back to work. Classic examples of this are treatments for mental health. The NHS Talking Therapies programme now treats 750,000 people a year for anxiety disorders and depression. Half of them recover and the service demonstrably pays for itself. It needs to expand.

But we also need a parallel service for those addicted to alcohol, drugs and gambling. These individuals are more difficult to treat — but the savings if they recover are greater. Again, it will pay for itself. And so will the improved therapy service for children and young people that Labour has promised.

However, most spending policies do cost more than they save. For these it is crucial to measure the benefit/cost ratio, converting the wellbeing benefit into its monetary equivalent. For example, we can evaluate the wellbeing gain to a community of having more police and subsequently less crime. Once this is converted into money, we calculate that the benefit/cost ratio is 12:1 — very high…(More)”.

AI has a democracy problem. Citizens’ assemblies can help.


Article by Jack Stilgoe: “…With AI, beneath all the hype, some companies know that they have a democracy problem. OpenAI admitted as much when they funded a program of pilot projects for what they called “Democratic Inputs to AI.” There have been some interesting efforts to involve the public in rethinking cutting-edge AI. A collaboration between Anthropic, one of OpenAI’s competitors, and the Collective Intelligence Project asked 1000 Americans to help shape what they called “Collective Constitutional AI.” They were asked to vote on statements such as “the AI should not be toxic” and “AI should be interesting,” and they were given the option of adding their own statements (one of the stranger statements reads “AI should not spread Marxist communistic ideology”). Anthropic used these inputs to tweak its “Claude” Large Language Model, which, when tested against standard AI benchmarks, seemed to help mitigate the model’s biases.

In using the word “constitutional,” Anthropic admits that, in making AI systems, they are doing politics by other means. We should welcome the attempt to open up. But, ultimately, these companies are interested in questions of design, not regulation. They would like there to be a societal consensus, a set of human values to which they can “align” their systems. Politics is rarely that neat…(More)”.