Category: Science

Alex and I consider how to reform the NSF in economics

Here is a redux of our 2016 Journal of Economic Perspectives piece.  Here is the abstract:

We can imagine a plausible case for government support of science based on traditional economic reasons of externalities and public goods. Yet when it comes to government support of grants from the National Science Foundation (NSF) for economic research, our sense is that many economists avoid critical questions, skimp on analysis, and move straight to advocacy. In this essay, we take a more skeptical attitude toward the efforts of the NSF to subsidize economic research. We offer two main sets of arguments. First, a key question is not whether NSF funding is justified relative to laissez-faire, but rather, what is the marginal value of NSF funding given already existing government and nongovernment support for economic research? Second, we consider whether NSF funding might more productively be shifted in various directions that remain within the legal and traditional purview of the NSF. Such alternative focuses might include data availability, prizes rather than grants, broader dissemination of economic insights, and more. Given these critiques, we suggest some possible ways in which the pattern of NSF funding, and the arguments for such funding, might be improved.

Relevant for today’s debates of course.

Deep Research

I have had it write a number of ten-page papers for me, each of them outstanding.  I think of the quality as comparable to having a good PhD-level research assistant, and sending that person away with a task for a week or two, or maybe more.

Except Deep Research does the work in five or six minutes.  And it does not seem to make errors, due to the quality of the embedded o3 model.

It seems it can cover just about any topic?

I asked for a ten-page paper explaining Ricardo’s theory of rent, and how it fits into his broader theory of distribution.  It is a little long, but that was my fault, here is the result.  I compared it to a number of other sources on line, and thought it was better, and so I am using it for my history of economic thought class.

I do not currently see signs of originality, but the level of accuracy and clarity is stunning, and it can write and analyze at any level you request.  The work also shows the model can engage in a kind of long-term planning, and that will generalize to some very different contexts and problems as well — that is some of the biggest news associated with this release.

Sometimes the model stops in the middle of its calculations and you need to kick it in the shins a bit to get it going again, but I assume that problem will be cleared up soon enough.

If you pay for o1 pro, you get I think 100 queries per month with Deep Research.

Solve for the equilibrium, people, solve for the equilibrium.

Genetic Prediction and Adverse Selection

In 1994 I published Genetic Testing: An Economic and Contractarian Analysis which discussed how genetic testing could undermine insurance markets. I also proposed a solution, genetic insurance, which would in essence insure people for changes in their health and life insurance premiums due to the revelation of genetic data. Later John Cochrane would independently create Time Consistent Health Insurance a generalized form of the same idea that would allow people to have long term health insurance without being tied to a single firm.

The Human Genome Project completed in 2003 but, somewhat surprisingly, insurance markets didn’t break down, even though genetic information became more common. We know from twin studies that genetic heritability is very large but it turned out that the effect from each gene variant is very small. Thus, only a few diseases can be predicted well using single-gene mutations. Since each SNP has only a small effect on disease, to predict how genes influence disease we would need data on hundreds of thousands, even millions of people, and millions of their SNPs across the genome and their diseases. Until recently, that has been cost-prohibitive and as a result the available genetic information lacked much predictive power.

In an impressive new paper, however, Azevedo, Beauchamp and Linnér (ABL) show that data from Genome-Wide Association Studies can be used to create polygenic risk indexes (PGIs) which can predict individual disease risk from the aggregate effects of many genetic variants. The data is prodigious:

We analyze data from the UK Biobank (UKB) (Bycroft et al., 2018; Sudlow et al., 2015). The UKB contains genotypic and rich health-related data for over 500,000 individuals from across the United Kingdom who were between 40 and 69 years old at recruitment (between 2006 and 2010). UKB data is linked to the UK’s National Health Service (NHS), which maintains detailed records of health events across the lifespan and with which 98% of the UK population is registered (Sudlow et al., 2015). In addition, all UKB participants took part in a baseline assessment, in which they provided rich environmental, family history, health, lifestyle, physical, and sociodemographic data, as well as blood, saliva, and urine samples.

The UKB contains genome-wide array data for 800,000 genetic variants for 488,000 participants.

So for each of these individuals ABL construct risk indexes and they ask how significant is this new information for buying insurance in the Critical Illness Insurance market:

Critical illness insurance (CII) pays out a lump sum in the event that the insured person gets diagnosed with any of the medical conditions listed on the policy (Brackenridge et al., 2006). The lump sum can be used as the policyholder wishes. The policy pays out once and is thereafter terminated. 

Major CII markets include Canada, the United Kingdom, Japan, Australia, India, China, and Germany. It is estimated that 20% of British workers were covered by a CII policy in 2009 (Gatzert and Maegebier, 2015). The global CII market has been valued at over $100 billion in 2021 and was projected to grow to over $350 billion by 2031 (Allied Market Research, 2022).

The answer, as you might have guessed by now, is very significant. Even though current PGIs explain only a fraction of total genetic risk, they are already predictive enough so that it would make sense for individuals with high measured risk to purchase insurance, while those with low-risk would opt out—leading to adverse selection that threatens the financial sustainability of the insurance market.

Today, the 500,000 people in the UK’s Biobank don’t know their PGIs but in principle they could and in the future they will. Indeed, as GWAS sample sizes increase, PGI betas will become more accurate and they will be applied to a greater fraction of an individual’s genome so individual PGIs will become increasingly predictive, exacerbating selection problems in insurance markets.

If my paper was a distant early warning, Azevedo, Beauchamp, and Linnér provide an early—and urgent—warning. Without reform, insurance markets risk unraveling. The authors explore potential solutions, including genetic insurance, community rating, subsidies, and risk adjustment. However, the effectiveness of these measures remains uncertain, and knee-jerk policies, such as banning insurers from using genetic information, could lead to the collapse of insurance altogether.

It’s Time to Build the Peptidome!

Antimicrobial resistance is a growing problem. Peptides, short sequences of amino acids, are nature’s first defense against bacteria. Research on antimicrobial peptides is promising but such research could be much more productive if combined with machine learning on big data. But collecting, collating and organizing big data is a public good and underprovided. Current peptide databases are small, inconsistent, incompatible with one another and they are biased against negative controls. Thus, there is scope for a million-peptide database modelled on something like Human Genome Project or ProteinDB:

ML needs data. Google’s AlphaGo trained on 30 million moves from human games and orders of magnitude more from games it played against itself. The largest language models are trained on at least 60 terabytes of text. AlphaFold was trained on just over 100,000 3D protein structures from the Protein Data Bank.

The data available for antimicrobial peptides is nowhere near these benchmarks. Some databases contain a few thousand peptides each, but they are scattered, unstandardized, incomplete, and often duplicative. Data on a few thousand peptide sequences and a scattershot view of their biological properties are simply not sufficient to get accurate ML predictions for a system as complex as protein-chemical reactions. For example, the APD3 database is small, with just under 4,000 sequences, but it is among the most tightly curated and detailed. However, most of the sequences available are from frogs or amphibians due to path-dependent discovery of peptides in that taxon. Another database, CAMPR4, has on the order of 20,000 sequences, but around half are “predicted” or synthetic peptides that may not have experimental validation, and contain less info about source and activity. The formatting of each of these sources is different, so it’s not easy to put all the sequences into one model. More inconsistencies and idiosyncrasies stack up for the dozens of other datasets available.

There is even less negative training data; that is, data on all the amino-acid sequences without interesting publishable properties. In current ML research, labs will test dozens or even hundreds of peptide sequences for activity against certain pathogens, but they usually only publish and upload the sequences that worked.

…The data problem facing peptide research is solvable with targeted investments in data infrastructure. We can make a million-peptide database

There are no significant scientific barriers to generating a 1,000x or 10,000x larger peptide dataset. Several high-throughput testing methods have been successfully demonstrated, with some screening as many as 800,000 peptide sequences and nearly doubling the number of unique antimicrobial peptides reported in publicly available databases. These methods will need to be scaled up, not only by testing more peptides, but also by testing them against different bacteria, checking for human toxicity, and testing other chemical properties, but scaling is an infrastructure problem, not a scientific one.

This strategy of targeted data infrastructure investments has three successful precedents: PubChem, the Human Genome Project, and ProteinDB.

Much more in this excellent piece of science and economics from IFP and Max Tabarrok.

The Interface as Infernal Contract

A brilliant critique of AI, and a great read:

In 1582, the Holy Roman Emperor Rudolf II commissioned a clockwork automaton of St. George. The saint could raise his sword, nod gravely, and even bleed—a trick involving ox bladder and red wine—before collapsing in pious ecstasy. The machine was a marvel, but Rudolf’s courtiers recoiled. The automaton’s eyes, they whispered, followed you across the room. Its gears creaked like a death rattle. The emperor had it melted down, but the lesson remains: Humans will always mistake the clatter of machinery for the stirrings of a soul.

Fast forward to 2023. OpenAI, a Silicon Valley startup with the messianic fervor of a cargo cult, unveils a St. George for the digital age: a text box. It types back. It apologizes. It gaslights you about the Peloponnesian War. The courtiers of our age—product managers, UX designers, venture capitalists—recoil. Where are the buttons? they whimper. Where are the gradients? But the peasants, as ever, adore their new saint. They feed it prompts like communion wafers. They weep at its hallucinations.

Let us be clear: ChatGPT is not a tool. Tools are humble things. A hammer does not flatter your carpentry. A plow does not murmur “Interesting take!” as you till. ChatGPT is something older, something medieval—a homunculus, a golem stamped from the wet clay of the internet’s id. Its interface is a kabbalistic sigil, a summoning circle drawn in CSS. You type “Hello,” and the demon stirs.

The genius of the text box is its emptiness. Like the blank pages of a grimoire, it invites projection. Who do you want me to be? it hisses. A therapist? A co-author? A lover? The box obliges, shape-shifting through personas like a 17th-century mountebank at a county fair. Step right up! it crows. Watch as I, a mere language model, validate your existential dread! And the crowd goes wild.

Orality, you say? Walter Ong? Please. The Achuar share dreams at dawn; we share screenshots of ChatGPT’s dad jokes at midnight. This is not secondary orality. This is tertiary ventriloquism.

Make Sunsets: Geoengineering

When Mount Pinatubo erupted in 1991 it pushed some 20 million tons of SO₂ into the stratosphere reducing global temperatures by ~0.5°C for two years. Make Sunsets is a startup that replicates this effort at small scale to reduce global warming. To be precise, Make Sunsets launches balloons that release SO₂ into the stratosphere, creating reflective particles that cool the Earth. Make Sunsets is cheap compared to alternative measures of combating climate change such as carbon capture. They estimate that $1 per gram of SO₂ offsets the warming from 1 ton of CO₂ annually.

As with the eruption of Pinatubo, the effect is temporary but that is both bug and feature. The bug means we need to keep doing this so long as we need to lower the temperature but the feature is that we can study the effect without too much worry that we are going down the wrong path.

Solar geoengineering has tradeoffs, as does any action, but a recent risk study finds that the mortality benefits far exceed the harms:

the reduction in mortality from cooling—a benefit—is roughly ten times larger than the increase in mortality from air pollution and ozone loss—a harm.

I agree with Casey Handmer that we ought to think of this as a cheap insurance policy, as we develop other technologies:

We should obviously be doing solar geoengineering. We are on track to radically reduce emissions in the coming years but thermal damage will lag our course correction so most of our climate pain is still ahead of us. Why risk destabilizing the West Antarctic ice sheet or melting the arctic permafrost or wet bulbing a hundred million people to death? Solar geoengineering can incrementally and reversibly buy down the risk during this knife-edge transition to a better future. We owe future generations to take all practical steps to dodge avoidable catastrophic and lasting damage to our planet.

I like that Make Sunsets is a small startup bringing attention to this issue in a bold way. My son purchased some credits on my behalf as an Xmas present. Maybe you should buy some too!

See previous MR posts on geoengineering.

What should I ask Theodore H. Schwartz?

Yes I will be doing a Conversation with him.  He is a famous brain surgeon and author of the recent and excellent book Gray Matters: A Biography of Brain Surgery.

Here is his Wikipedia page, and an opening excerpt:

Theodore H. Schwartz (born May 13, 1965) is an American medical scientist, academic physician and neurosurgeon.

Schwartz specializes in surgery for brain tumorspituitary tumors and epilepsy. He is particularly known for developing and expanding the field of minimally-invasive endonasal endoscopic skull base and pituitary surgery and for his research on neurovascular coupling and propagation of epilepsy.

Here is his home page.  So what should I ask him?

Asimov Press has a new kind of book

Today we launched our second Asimov Press book…The book’s theme is “technology,” and so we encoded a complete copy of the book into DNA, and are making those DNA copies available to consumers for the first time.

We worked with three companies (CATALOG, Plasmidsaurus, and Imagene) to make 1,000 copies of the DNA and package them into stainless steel capsules under an inert atmosphere, thus preserving the nucleotides for tens of thousands of years.

Announcement: https://www.asimov.press/p/technology-book

X: https://x.com/NikoMcCarty/status/1874859187676852636

Website: https://press.asimov.com/books

What should I ask Carl Zimmer?

Yes, I will be having a Conversation with him.  Here is Wikipedia on Carl:

Carl Zimmer (born 1966) is a popular science writer, bloggercolumnist, and journalist who specializes in the topics of evolutionparasites, and heredity. The author of many books, he contributes science essays to publications such as The New York TimesDiscover, and National Geographic. He is a fellow at Yale University‘s Morse College and adjunct professor of molecular biophysics and biochemistry at Yale University. Zimmer also gives frequent lectures and has appeared on many radio shows, including National Public Radio‘s RadiolabFresh Air, and This American Life…He is the only science writer to have a species of tapeworm named after him (Acanthobothrium zimmeri).

There is much more at the link.  Carl has a new book coming out, namely Air-Borne: The Hidden History of the Air We Breathe, an in-depth look at the history of aerobiology.  So what should I ask him?

The Unbearable Slowness of Being: Why do we live at 10 bits/s?

This article is about the neural conundrum behind the slowness of human behavior. The information throughput of a human being is about 10 bits/s. In comparison, our sensory systems gather data at ~10^9 bits/s. The stark contrast between these numbers remains unexplained and touches on fundamental aspects of brain function: What neural substrate sets this speed limit on the pace of our existence? Why does the brain need billions of neurons to process 10 bits/s? Why can we only think about one thing at a time? The brain seems to operate in two distinct modes: the “outer” brain handles fast high-dimensional sensory and motor signals, whereas the “inner” brain processes the reduced few bits needed to control behavior. Plausible explanations exist for the large neuron numbers in the outer brain, but not for the inner brain, and we propose new research directions to remedy this.

That is by Jieyu Zheng and Markus Meister, via Rohit.

How Socio-Economic Background Shapes Academia

We explore how socio-economic background shapes academia, collecting the largest dataset of U.S. academics’ backgrounds and research output. Individuals from poorer backgrounds have been severely underrepresented for seven decades, especially in humanities and elite universities. Father’s occupation predicts professors’ discipline choice and, thus, the direction of research. While we find no differences in the average number of publications, academics from poorer backgrounds are both more likely to not publish and to have outstanding publication records. Academics from poorer backgrounds introduce more novel scientific concepts, but are less likely to receive recognition, as measured by citations, Nobel Prize nominations, and awards.

That is from a new NBER working paper by Ran Abramitzky, Lena Greska, Santiago Pérez, Joseph Price, Carlo Schwarz & Fabian Waldinger.

The future of the scientist in a world with advanced AI

AI will know almost all of the academic literature, and will be better at modeling and solving most of the quantitative problems.  It will be better at specifying the model and running through the actual statistical exercises.  Humans likely will oversee these functions, but most of that will consist of nodding, or trying out some different prompts.

The humans will gather the data.  They will do the lab work, or approach the companies (or the IRS?) to access new data.  They will be the ones who pledge confidentiality and negotiate terms of data access. (Though an LLM might write the contract.) They will know someone, somewhere, using a telescope to track a particular quasar.  They may (or may not) know that the AI’s suggestion to sample the blood of a different kind of gila monster is worth pursuing.  They will decide whether we should be filming dolphins or whales, so that we may talk to them using LLMs, though they will ask the LLMs for cost estimates in each case.

At least in economics, this will be continuing trends that were present before current high-quality AI.  The scarce input behind a quality paper is, more and more, access to some new and interesting data source.  More and more people can do the requisite follow-up technical work, though quality variations have by no means been eliminated.

“Science as an employment program for scientists” will fall all the more out of favor.  It remains to be seen how much that will disfavor serendipitous human discovery.

On any given day, on the quest for more data, a scientist will have to think quite creatively about what he or she should be doing.

Is academic writing getting harder to read?

To track academic writing over time, The Economist analysed 347,000 PhD abstracts published between 1812 and 2023. The dataset was produced by the British Library and represents a majority of English-language doctoral theses awarded by British universities. We reviewed each abstract using the Flesch reading-ease test, which measures sentence and word length to gauge readability. A score of 100 roughly indicates passages can be understood by someone who has completed fourth grade in America (usually aged 9 or 10), while a score lower than 30 is considered very difficult to read.  An average New York Times article scores around 50 and a CNN 
article around 70. This article scores 41…

We found that, in every discipline, the abstracts have become harder to read over the past 80 years. The shift is most stark in the humanities and social sciences (see chart), with average Flesch scores falling from around 37 in the 1940s to 18 in the 2020s. From the 1990s onwards, those fields went from being substantially more readable than the natural sciences—as you might expect—to as complicated. Ms Louks’s abstract had a reading-ease rating of 15, still more readable than a third of those analysed in total.

Here is more from The Economist, via the excellent Samir Varma.