Month: April 2023

Saturday assorted links

1. Mary Lou Williams appreciation (NYT).

2. Is D.C. becoming a bagel city?

3. Arthur Allshire on whether LLMs will be an anti-censorship tool.

4. Filipinos nailed to crosses on Good Friday despite objection by Catholic Church.

5. The economy in Zambia.

6. “Pretty incredible that the Indian government has released a version of @OpenAI Whisper fine tuned on a Hindi data set.”  And AI and religion is a match made in heaven (NYT).  Really.  “The flood of papal imagery has been so voluminous that some people in online generative A.I. forums have begged for creators to use another inspiration.”  I really like this one.

Lessons from the Baruch Plan for Nuclear Weapons

The invention of atomic energy posed a novel global challenge: could the technology be controlled to avoid destructive uses and an existentially dangerous arms race while permitting the broad sharing of its benefits? From 1944 onwards, scientists, policymakers, and other technical specialists began to confront this challenge and explored policy options for dealing with the impact of nuclear technology. We focus on the years 1944 to 1951 and review this period for lessons for the governance of powerful technologies, and find the following: Radical schemes for international control can get broad support when confronted by existentially dangerous technologies, but this support can be tenuous and cynical. Secrecy is likely to play an important, and perhaps harmful, role. The public sphere may be an important source of influence, both in general and in particular in favor of cooperation, but also one that is manipulable and poorly informed. Technical experts may play a critical role, but need to be politically savvy. Overall, policymaking may look more like “muddling through” than clear-eyed grand strategy. Cooperation may be risky, and there may be many obstacles to success.

That is by Waqar Zaidi and Allan Dafoe, at the Centre for Governance of AI, exactly the kind of work people should be doing.

Matt Yglesias on movies vs. TV

But I’ve gotten really disgruntled with the “prestige TV” landscape and am trying to redirect my content consumption accordingly. One thing that makes movies really great in my view is that before they shoot a movie, they write a screenplay and the screenplay has an end. Both the screenwriter and other people have read that screenplay all the way from beginning to end and they’ve tweaked and changed it and gotten it into a position where they are ready to start production. Then after a movie is filmed, the editor and director work with the footage and come up with a complete movie that has a beginning, a middle, and an end. They then ship the movie out, and it’s screened by critics who watch the entire movie before writing their review.

This does not guarantee that every movie that comes out is good. But it does guarantee that if someone tells you “‘The Menu’ is good,” they are evaluating a completed product…

By contrast, TV shows have this quasi-improvisational quality where the showrunners are constantly needing to come up with new balls to toss into the air. In old-fashioned non-prestigious “adventure of the week”-type shows, this actually works fine because the writers are not building up tension or setting unexplored plots in motion. But as serialized TV storytelling has gotten more and more common, we’re more and more often asked to show patience through early episodes or to try to find things intriguing with no ability to know whether any of it will pay off. Creators often have no idea where they’re going with the story.

Back in HBO’s heyday, the tradeoff was that The Sopranos and The Wire got to paint on a giant canvas and tell stories that are just too capacious for the movie format. But eventually networks got tired of spending that kind of money and cut back the sizes of the casts to something more normal for television.

That is from his Friday mailbag ($).  The bottom line is that, like Matt, you should watch more movies and less TV.

Did Ottoman Sultans ban print?

Did printing transform the Ottoman Empire? And what took the Ottomans so long to print? Much of the scholarship surrounding the topic of Ottoman printing, or the occurrence of printing within the Ottoman Empire (1453–1922), is structured around these two related frameworks. In this essay, I argue that these frameworks are ahistorical because they predicate Ottoman printing on the European experience of print. To support this point, I examine the disproportionate role played by certain early modern European accounts of Ottoman printing within Western and Arabic historiography. In particular, I examine the life cycle of scholars’ belief that Ottoman sultans banned printing, which I contrast with extant documentation for the imperial Porte’s stance on printing. I argue that the sources available to scholars today do not support the notion that the sultans banned printing. Rather, they demonstrate that this claim arose from early modern European scholars’ search to articulate their sense of Ottoman inadequacy through explanations for why Ottomans did not print. The history of this particular line of inquiry is significant, I argue, because many scholars continue to probe the issue of why Ottomans did not print. In so doing, they maintain the expectation that print would revolutionize society, even though they have begun questioning the existence of the ban.

That is from Kathryn A. Schwartz, in Print History (jstor).  Via Benedikt A.

Friday assorted links

1. Gloria the young sloth.

2. ” Our theory sheds light on the mechanism of disastrous individual decisions in highly institutionalized authoritarian regimes.”  With reference to Putin.

3. White House issues update to cost-benefit guidelines.  And an EA analysis.  And is the best kimchi made in earthenware pots?

4. The gambler who beat roulette (Bloomberg).

5. “Inflation is conflict.”  New paper, not exactly my view, but it is not as if previous theories have been entirely successful.

6. Peter Sellars version of St. Matthew’s Passion, with Berlin, is now ungated for Easter.

The Nuclear Non-proliferation Treaty and existential AGI risk

The Nuclear Non-Proliferation Treaty, activated in 1970, has been relatively successful in limiting nuclear proliferation.  When it comes to nuclear weapons, it is hard to find good news, but the treaty has acted as one deterrent of many to nation-states acquiring nuclear arms.  Of course the treaty works, in large part, because the United States (working with allies) has lots of nuclear weapons, a powerful non-nuclear military, de facto control of SWIFT, and so on.  We strongly encourage nations not to go acquiring nuclear weapons — just look at the current sanctions on Iran, noting the policy does not always succeed.

One approach to AI risk is to treat it like nuclear weapons and also their delivery systems.  Let the United States get a lead, and then hope the U.S. can (in conjunction with others) enforce “OK enough” norms on the rest of the world.

Another approach to AI risk is to try to enforce a collusive agreement amongst all nations not to proceed with AI development, at least along certain dimensions, or perhaps altogether.

The first of these two options seems obviously better to me.  But I am not here to argue that point, at least not today.  Conditional on accepting the superiority of the first approach, all the arguments for AI safety are arguments for AI continuationism.  (And no, this doesn’t mean building a nuclear submarine without securing the hatch doors.)  At least for the United States.  In fact I do support a six-month AI pause — for China.  Yemen too.

It is a common mode of presentation in AGI circles to present wordy, swirling tomes of multiple concerns about AI risk.  If some outside party cannot sufficiently assuage all of those concerns, the writer is left with the intuition that so much is at stake, indeed the very survival of the world, and so we need to “play it safe,” and thus they are lead to measures such as AI pauses and moratoriums.

But that is a non sequitur.  The stronger the safety concerns, the stronger the arguments for the “America First” approach.  Because that is the better way of managing the risk.  Or if somehow you think it is not, that is the main argument you must make and persuade us of.

(Scott Alexander has a new post “Most technologies aren’t races,” but he doesn’t either choose one of the two approaches listed above, nor does he outline a third alternative.  Fine if you don’t want to call them “races,” you still have to choose.  As a side point, once you consider delivery systems, nuclear weapons are less of a yes/no thing than he suggests.  And this postulated take is a view that nobody holds, nor did we practice it with nuclear weapons: “But also, we can’t worry about alignment, because that would be an unacceptable delay when we need to “win” the AI “race”.”  On the terminology, Rohit is on target.  Furthermore, good points from Erusian.  And this claim of Scott’s shows how far apart we are in how we consider institutional and also physical and experimental constraints: “In a fast takeoff, it could be that you go to sleep with China six months ahead of the US, and wake up the next morning with China having fusion, nanotech, and starships.”)

Addendum:

As a side note, if the real issue in the safety debate is “America First” vs. “collusive international agreement to halt development,” who are the actual experts?  It is not in general “the AI experts,” rather it is people with experience in and study of:

1. Game theory and collective action

2. International agreements and international relations

3. National security issues and understanding of how government works

4. History, and so on.

There is a striking tendency, amongst AI experts, EA types, AGI writers, and “rationalists” to think they are the experts in this debate.  But they are only on some issues, and many of those issues (“new technologies can be quite risky”) are not so contested. And because these individuals do not frame the problem properly, they are doing relatively little to consult what the actual “all things considered” experts think.

*A New History of Greek Mathematics*

I have read only about 30 pp. so far, but this is clearly one of the best science books I have read, ever.  It is clear, always to the point, conceptual, connects advances in math to the broader history, explains the math, and full of interesting detail.  By Reviel Netz.  Here is a brief excerpt:

And this is how mathematics first emerges in the historical record: the simple, clever games accompanying the education of bureaucrats.

One of the best books of the year, highly recommended.

Thursday assorted links

1. “We find that using algorithmic responses changes language and social relationships. More specifically, it increases communication speed, use of positive emotional language, and conversation partners evaluate each other as closer and more cooperative. However, consistent with common assumptions about the adverse effects of AI, people are evaluated more negatively if they are suspected to be using algorithmic responses.”  Link here.

2. Living with non-alignment.  And a very sane take on AI risk.  Very good thread.  And should American VCs be funding Chinese AI?  And Leopold Aschenbrenner responds to me on AI risk, very good piece.

3. Why isn’t Europe doing worse?

4. “I hereby challenge professors from universities around the world to submit assignments that they believe are AI-immune.

5. Which individuals are most likely to believe that AI is likely to destroy society?

Is the Great Awokening a global phenomenon?

And perhaps it did not start in the United States?  Here is more from David Rozado, including a full research paper:

My excellent Conversation with Jessica Wade

Here is the audio, video, and transcript.  Here is part of the summary:

She joined Tyler to discuss if there are any useful gender stereotypes in science, distinguishing between productive and unproductive ways to encourage women in science, whether science Twitter is biased toward men, how AI will affect gender participation gaps, how Wikipedia should be improved, how she judges the effectiveness of her Wikipedia articles, how she’d improve science funding, her work on chiral materials and its near-term applications, whether writing a kid’s science book should be rewarded in academia, what she learned spending a year studying art in Florence, what she’ll do next, and more.

Here is the opening bit:

COWEN: Let’s start with women in science. We will get to your research, but your writings — why is it that women in history were so successful in astronomy so early on, compared to other fields?

WADE: Oh, that’s such a hard question [laughs] and a fascinating one. When you look back at who was allowed to be a scientist in the past, at which type of woman was allowed to be a scientist, you were probably quite wealthy, and you either had a husband who was a scientist or a father who was a scientist. And you were probably allowed to interact with science at home, potentially in things like polishing the lenses that you might use on a telescope, or something like that.

Caroline Herschel was quite big on polishing the lenses that Herschel used to go out and look at and identify comets, and was so successful in identifying these comets that she wanted to publish herself and really struggled, as a woman, to be allowed to do that at the end of the 1800s, beginning of the 1900s. I think, actually, it was just that possibility to be able to access and do that science from home, to be able to set up in your beautiful dark-sky environment without the bright lights of a city and do it alongside your quite successful husband or father.

After astronomy, women got quite big in crystallography. There were a few absolutely incredible women crystallographers throughout the 1900s. Dorothy HodgkinKathleen LonsdaleRosalind Franklin — people who really made that science possible. That was because they were provided entry into that, and the way that they were taught at school facilitated doing that kind of research. I find it fascinating they were allowed, but if only we’d had more, you could imagine what could have happened.

COWEN: So, household production you think is the key variable, plus the ability to be helped or trained by a father or husband?

The discussion of chirality and her science work is very interesting, though hard to summarize.  I very much like this part, when I asked her about her most successful unusual work habit:

But just writing the [Wikipedia] biography of the person I was going to work with meant that I was really prepped for going. And if I’m about to see someone speak, writing their biography before means I get this. That’s definitely my best work habit — write the Wikipedia page of what it is that you are working on.

I don’t agree with her on the environment/genes issue, but overall a very good CWT, with multiple distinct parts.

Do women disagree less in science?

This paper examines the authorship of post-publication criticisms in the scientific literature, with a focus on gender differences. Bibliometrics from journals in the natural and social sciences show that comments that criticize or correct a published study are 20-40% less likely than regular papers to have a female author. In preprints in the life sciences, prior to peer review, women are missing by 20-40% in failed replications compared to regular papers, but are not missing in successful replications. In an experiment, I then find large gender differences in willingness to point out and penalize a mistake in someone’s work.

That is from a new paper by David Klinowski.  Via the excellent Kevin Lewis.