Helen Toner is an AI policy researcher (at the Center for Security and Emerging Technology at Georgetown University and formerly on OpenAI’s board). The article gives multiple clear examples showing how current AI systems can have surprisingly poor performance in some tasks while being superhuman in others, unpredictably. I agree with her prediction of a turbulent period: the jaggedness won't go away any time soon. Some tasks are hard to verify, some are hard to fit into the context window of an LLM, some settings are adversarial.
So the bold claim in this talk is: maybe AI will keep getting better and maybe AI will keep sucking in important ways. I want to be really clear: I think expecting that jaggedness might continue is consistent with expecting that in the long run, AI can get better than humans at most or all things. And it’s also consistent with expecting that in the short run, AI will be very, very disruptive. So this is not a view that’s saying AI capabilities are jagged and therefore the future is going to be boring, it’s going to be similar, this is all a nothing-burger. I think this could still look very, very interesting, difficult to deal with, disruptive, confusing, risky.
Nice image illustrating the article: Toner uses the analogy of hold and cold liquids mixing from fluid dynamics to illustrate jaggedness (from this video of "Rayleigh-Benard Convection").
Rachel Coldicutt has good commentary Bluesky, although is frustrated that much of this material isn't already well known: "This is a really clear and impressive bit of communication. I'm somewhat baffled that it's needed."
The Times has uncovered nearly 50 cases of people having mental health crises during conversations with ChatGPT. Nine were hospitalized; three died. After Adam Raine’s parents filed a wrongful-death lawsuit in August, OpenAI acknowledged that its safety guardrails could “degrade” in long conversations. It also said it was working to make the chatbot “more supportive in moments of crisis".
Mental fact demonstrated: They have contemplated how their actions affected me and now understand why they shouldn’t behave similarly in the future.AI Harm to Costly Signaling: They may have used an LLM to write this. If so, they do not actually care enough about me to think through the negative effects their actions had on me.AI Harm to Proof of Knowledge: They may not actually understand why the actions harmed me or possess specific background knowledge necessary to avoid harming me in the future.
Consider, for example, a student in the developing world without access to a functional accreditation system. In the past, a well-crafted, thoughtful e-mail might well serve to open doors to informal networks of mentorship and training: such a gesture provided both proof of abilities and of the necessary internal motivation that would lead a busy, but sympathetic, professor to take note. Such avenues to advancement are closed, however, once equivalent texts can be produced with a minimal prompt. A student without social capital is hurt by this vitiation of mental proof, while another, who comes with institutional backing, has other avenues to establish their credibility.