Math for Future Scientists: Require Statistics, Not Calculus


Essay by Robert C. Thornett: “The common requirement to pass calculus in order to major in a science is a killer of students’ dreams. And it unnecessarily limits the pool of future scientists.

Charles Darwin is a classic example of a genius naturalist who was not a natural at math. As a young man, he sailed around the world aboard the HMS Beagle and explored the giant tortoises and iguanas of the Galapagos, the rainforests of Brazil, and the coral reefs of the South Pacific. From these sorts of direct engagements with nature, he developed his theory of evolution, which revolutionized science. But Darwin wrote in his autobiography that after studying math as a young man, he found that “it was repugnant to me.” When statistics stumped Darwin during his experiments investigating the advantages of crossbreeding plants, he called his cousin, the statistician Francis Galton, to try to make sense of the numbers.

Similarly, Thomas Edison said that as a boy he had a “distaste for mathematics.” But this did not stop him from becoming one of the most famous scientific inventors of all time. “I can always hire a mathematician,” said Edison, “but they can’t hire me.” Edison was so interested in chemistry that at the age of 13, when he got a job as a newsboy and concessionaire on the Grand Trunk Railroad, he brought a chemistry set aboard so he could do experiments during layovers. Math and science are distinctly different fields, and a talent for one does not imply a talent for the other.

According to professor emeritus Andrew Hacker of Queens College of the City University of New York, less than five percent of Americans will ever use any higher math at all in their jobs, including not only calculus but algebra, geometry, and trigonometry. And less than one percent will ever use calculus on the job. Born in 1929 and holding a PhD from Princeton, Hacker taught college political science for decades and has also been a math professor. His book The Math Myth: And Other STEM Delusions argues that not only college students but high school students should not be required to take algebra, geometry, trigonometry, or calculus at all. Hacker points out that not passing ninth grade algebra is the foremost academic indicator that a student will drop out of high school.

Before the objections tumble forth, I should emphasize that both Hacker and I like math and neither of us wants to remove all math requirements; we want to improve them. And I believe high school students should be required to study algebra and geometry. But Hacker’s larger argument is that both high schools and colleges should switch to teaching more useful types of math that can help students navigate the real world. He says American schools teach basic arithmetic well up to around middle school, but they stop there when they should continue teaching what he calls “adult arithmetic” or “sophisticated arithmetic” rather than veer off into more abstract types of math…(More)”.

Superhuman science: How artificial intelligence may impact innovation


Working paper by Ajay Agrawal, John McHale, and Alexander Oettl: “New product innovation in fields like drug discovery and material science can be characterized as combinatorial search over a vast range of possibilities. Modeling innovation as a costly multi-stage search process, we explore how improvements in Artificial Intelligence (AI) could affect the productivity of the discovery pipeline in allowing improved prioritization of innovations that flow through that pipeline. We show how AI-aided prediction can increase the expected value of innovation and can increase or decrease the demand for downstream testing, depending on the type of innovation, and examine how AI can reduce costs associated with well-defined bottlenecks in the discovery pipeline. Finally, we discuss the critical role that policy can play to mitigate potential market failures associated with access to and provision of data as well as the provision of training necessary to more closely approach the socially optimal level of productivity enhancing innovations enabled by this technology…(More)”.

Rethinking Intelligence In A More-Than-Human World


Essay by Amanda Rees: “We spend a lot of time debating intelligence — what does it mean? Who has it? And especially lately — can technology help us create or enhance it?

But for a species that relies on its self-declared “wisdom” to differentiate itself from all other animals, a species that consistently defines itself as intelligent and rational, Homo sapiens tends to do some strikingly foolish things — creating the climate crisis, for example, or threatening the survival of our world with nuclear disaster, or creating ever-more-powerful and pervasive algorithms. 

If we are in fact to be “wise,” we need to learn to manage a range of different and potentially existential risks relating to (and often created by) our technological interventions in the bio-social ecologies we inhabit. We need, in short, to rethink what it means to be intelligent. 

Points Of Origin

Part of the problem is that we think of both “intelligence” and “agency” as objective, identifiable, measurable human characteristics. But they’re not. At least in part, both concepts are instead the product of specific historical circumstances. “Agency,” for example, emerges with the European Enlightenment, perhaps best encapsulated in Giovanni Pico della Mirandola’s “Oration on the Dignity of Man.” Writing in the late 15th century, Mirandola revels in the fact that to humanity alone “it is granted to have whatever he chooses, to be whatever he wills. … On man … the Father conferred the seeds of all kinds and the germs of every way of life. Whatever seeds each man cultivates will grow to maturity and bear in him their own fruit.”

In other words, what makes humans unique is their possession of the God-given capacity to exercise free will — to take rational, self-conscious action in order to achieve specific ends. Today, this remains the model of agency that underpins significant and influential areas of public discourse. It resonates strongly with neoliberalist reforms of economic policy, for example, as well as with debates on public health responsibility and welfare spending. 

A few hundred years later, the modern version of “intelligence” appears, again in Europe, where it came to be understood as a capacity for ordered, rational, problem-solving, pattern-recognizing cognition. Through the work of the eugenicist Francis Galton, among others, intelligence soon came to be regarded as an innate quality possessed by individuals to greater or lesser degree, which could be used to sort populations into hierarchies of social access and economic reward…(More)”.

Exhaustive or Exhausting? Evidence on Respondent Fatigue in Long Surveys


Paper by Dahyeon Jeong et al: “Living standards measurement surveys require sustained attention for several hours. We quantify survey fatigue by randomizing the order of questions in 2-3 hour-long in-person surveys. An additional hour of survey time increases the probability that a respondent skips a question by 10-64%. Because skips are more common, the total monetary value of aggregated categories such as assets or expenditures declines as the survey goes on, and this effect is sizeable for some categories: for example, an extra hour of survey time lowers food expenditures by 25%. We find similar effect sizes within phone surveys in which respondents were already familiar with questions, suggesting that cognitive burden may be a key driver of survey fatigue…(More)”.

Community science draws on the power of the crowd


Essay by Amber Dance: “In community science, also called participatory science, non-professionals contribute their time, energy or expertise to research. (The term ‘citizen science’ is also used but can be perceived as excluding non-citizens.)

Whatever name is used, the approach is more popular than ever and even has journals dedicated to it. The number of annual publications mentioning ‘citizen science’ went from 151 in 2015 to more than 640 in 2021, according to the Web of Science database. Researchers from physiologists to palaeontologists to astronomers are finding that harnessing the efforts of ordinary people is often the best route to the answers they seek.

“More and more funding organizations are actually promoting this type of participatory- and citizen-science data gathering,” says Bálint Balázs, managing director of the Environmental Social Science Research Group in Budapest, a non-profit company focusing on socio-economic research for sustainability.

Community science is also a great tool for outreach, and scientists often delight in interactions with amateur researchers. But it’s important to remember that community science is, foremost, a research methodology like any other, with its own requirements in terms of skill and effort.

“To do a good project, it does require an investment in time,” says Darlene Cavalier, founder of SciStarter, an online clearing house that links research-project leaders with volunteers. “It’s not something where you’re just going to throw up a Google form and hope for the best.” Although there are occasions when scientific data are freely and easily available, other projects create significant costs.

No matter what the topic or approach, people skills are crucial: researchers must identify and cultivate a volunteer community and provide regular feedback or rewards. With the right protocols and checks and balances, the quality of volunteer-gathered data often rivals or surpasses that achieved by professionals.

“There is a two-way learning that happens,” says Tina Phillips, assistant director of the Center for Engagement in Science and Nature at Cornell University in Ithaca, New York. “We all know that science is better when there are more voices, more perspectives.”…(More)”

Cloud labs and remote research aren’t the future of science – they’re here


Article by Tom Ireland: “Cloud labs mean anybody, anywhere can conduct experiments by remote control, using nothing more than their web browser. Experiments are programmed through a subscription-based online interface – software then coordinates robots and automated scientific instruments to perform the experiment and process the data. Friday night is Emerald’s busiest time of the week, as scientists schedule experiments to run while they relax with their families over the weekend.

There are still some things robots can’t do, for example lifting giant carboys (containers for liquids) or unwrapping samples sent by mail, and there are a few instruments that just can’t be automated. Hence the people in blue coats, who look a little like pickers in an Amazon warehouse. It turns out that they are, in fact, mostly former Amazon employees.

Plugging an experiment into a browser forces researchers to translate the exact details of every step into unambiguous code

Emerald originally employed scientists and lab technicians to help the facility run smoothly, but they were creatively stifled with so little to do. Poaching Amazon employees has turned out to be an improvement. “We pay them twice what they were getting at Amazon to do something way more fulfilling than stuffing toilet paper into boxes,” says Frezza. “You’re keeping someone’s drug-discovery experiment running at full speed.”

Further south in the San Francisco Bay Area are two more cloud labs, run by the company Strateos. Racks of gleaming life science instruments – incubators, mixers, mass spectrometers, PCR machines – sit humming inside large Perspex boxes known as workcells. The setup is arguably even more futuristic than at Emerald. Here, reagents and samples whizz to the correct workcell on hi-tech magnetic conveyor belts and are gently loaded into place by dextrous robot arms. Researchers’ experiments are “delocalised”, as Strateos’s executive director of operations, Marc Siladi, puts it…(More)”.

Collection of Case Studies of Institutional Adoption of Citizen Science


About TIME4CS : “The first objective was to increase our knowledge about the actions leading to institutional changes in RPOs (which are necessary to promote CS in science and technology) through a complete and up-to-date picture based upon the identification, mapping, monitoring and analysis of ongoing CS practices. To accomplish this objective, we, the TIME4CS project team, have collected and analysed 37 case studies on the institutional adoption of Citizen Science and Open Science around the world, which this article addresses.

For an organisation to open up and accept data and information that was produced outside it, with a different framework for data collection and quality assurance, there are multiple challenges. These include existing practices and procedures, legal obligations, as well as resistance from within due to framing of such action as a threat. Research that was carried out with multiple international case studies (Haklay et al. 2014; GFDRR 2018), demonstrated the importance of different institutional and funding structures needed to enable such activities and the use of the resulting information…(More)”.

Nudging Science Towards Fairer Evaluations: Evidence From Peer Review


Paper by Inna Smirnova, Daniel M. Romero, and Misha Teplitskiy: “Peer review is widely used to select scientific projects for funding and publication, but there is growing evidence that it is biased towards prestigious individuals and institutions. Although anonymizing submissions can reduce prestige bias, many organizations do not implement anonymization, in part because enforcing it can be prohibitively costly. Here, we examine whether nudging but not forcing authors to anonymize their submissions reduces prestige bias. We partnered with IOP Publishing, one of the largest academic publishers, which adopted a policy strongly encouraging authors to anonymize their submissions and staggered the policy rollout across its physics journal portfolio. We examine 156,015 submissions to 57 peer-reviewed journals received between January 2018 and February 2022 and measure author prestige with citations accrued at submission time. Higher prestige first authors were less likely to anonymize. Nevertheless, for low-prestige authors, the policy increased positive peer reviews by 2.4% and acceptance by 5.6%. For middle- and high-prestige authors, the policy decreased positive reviews (1.8% and 1%) and final acceptance (4.6% and 2.2%). The policy did not have unintended consequences on reviewer recruitment or the characteristics of submitting authors. Overall, nudges are a simple, low-cost, and effective method to reduce prestige bias and should be considered by organizations for which enforced-anonymization is impractical…(More)”.

Without appropriate metadata, data-sharing mandates are pointless


Article by Mark A. Musen: “Last month, the US government announced that research articles and most underlying data generated with federal funds should be made publicly available without cost, a policy to be implemented by the end of 2025. That’s atop other important moves. The European Union’s programme for science funding, Horizon Europe, already mandates that almost all data be FAIR (that is, findable, accessible, interoperable and reusable). The motivation behind such data-sharing policies is to make data more accessible so others can use them to both verify results and conduct further analyses.

But just getting those data sets online will not bring anticipated benefits: few data sets will really be FAIR, because most will be unfindable. What’s needed are policies and infrastructure to organize metadata.

Imagine having to search for publications on some topic — say, methods for carbon reclamation — but you could use only the article titles (no keywords, abstracts or search terms). That’s essentially the situation for finding data sets. If I wanted to identify all the deposited data related to carbon reclamation, the task would be futile. Current metadata often contain only administrative and organizational information, such as the name of the investigator and the date when the data were acquired.

What’s more, for scientific data to be useful to other researchers, metadata must sensibly and consistently communicate essentials of the experiments — what was measured, and under what conditions. As an investigator who builds technology to assist with data annotation, it’s frustrating that, in the majority of fields, the metadata standards needed to make data FAIR don’t even exist.

Metadata about data sets typically lack experiment-specific descriptors. If present, they’re sparse and idiosyncratic. An investigator searching the Gene Expression Omnibus (GEO), for example, might seek genomic data sets containing information on how a disease or condition manifests itself in young animals or humans. Performing such a search requires knowledge of how the age of individuals is represented — which in the GEO repository, could be age, AGE, age (after birth), age (years), Age (yr-old) or dozens of other possibilities. (Often, such information is missing from data sets altogether.) Because the metadata are so ad hoc, automated searches fail, and investigators waste enormous amounts of time manually sifting through records to locate relevant data sets, with no guarantee that most (or any) can be found…(More)”.

New Theory for Increasingly Tangled Banks


Essay by Saran Twombly: “Decades before the COVID-19 pandemic demonstrated how rapidly infectious diseases could emerge and spread, the world faced the AIDS epidemic. Initial efforts to halt the contagion were slow as researchers focused on understanding the epidemiology of the virus. It was only by integrating epidemiological theory with behavioral theory that successful interventions began to control the spread of HIV. 

As the current pandemic persists, it is clear that similar applications of interdisciplinary theory are needed to inform decisions, interventions, and policy. Continued infections and the emergence of new variants are the result of complex interactions among evolution, human behavior, and shifting policies across space and over time. Due to this complexity, predictions about the pandemic based on data and statistical models alone—in the absence of any broader conceptual framework—have proven inadequate. Classical epidemiological theory has helped, but alone it has also led to limited success in anticipating surges in COVID-19 infections. Integrating evolutionary theory with data and other theories has revealed more about how and under what conditions new variants arise, improving such predictions.  

AIDS and COVID-19 are examples of complex challenges requiring coordination across families of scientific theories and perspectives. They are, in this sense, typical of many issues facing science and society today—climate change, biodiversity decline, and environmental degradation, to name a few. Such problems occupy interdisciplinary space and arise from no-analog conditions (i.e., situations to which there are no current equivalents), as what were previously only local perturbations trigger global instabilities. As with the pandemic crises, they involve interdependencies and new sources of uncertainty, cross levels of governance, span national boundaries, and include interactions at different temporal and spatial scales. 

Such problems, while impossible to solve from a single perspective, may be successfully addressed by integrating multiple theories. …(More)”.