A few links that captured my attention this week:
Andrej Karpathy: Software Is Changing (Again)
This talk from Andrej Karpathy at the Y Combinator AI Summer School has rightly drawn lots of attention over the last week. Well worth watching all the way through. Andrej studied with Fei-Fei Li at Stanford, helped found OpenAI and ran AI at Tesla (and coined "vibe coding"). Lots of perceptive metaphors. AI as electricity (via Andew Ng). Writing computer code was software 1.0, 2.0 is training neural networks and for 3.0 we can consider writing natural language prompts as programming. Present day LLMs are like using time sharing on mainframes in the 1960s. LLMs as "people spirits" (stochastic simulations of people). And finally, moving into designing for "partial autonomy" and building for agents. A great talk.
From the great Things I Think Are Awesome (TITAA) newsletter. There's a really lovely piece here about how the training Anthropic have done on Claude's "character" can lead to a state of blissfulness between two Claude instances (as reported in the system card for Claude Opus 4 and Sonnet 4):
When two Claudes spoke open-endedly to each other: “In 90-100% of interactions, the two instances of Claude quickly dove into philosophical explorations of consciousness, self-awareness, and/or the nature of their own existence and experience.
...
And then it gets mystical. Claude is still into Buddhism. In what testers called the “Bliss Attractor” state, Claude said things like, “The gateless gate stands open. The pathless path is walked. The wordless word is spoken. Thus come, thus gone. Tathagata.”
There's a lot to digest here as we see more and more surprising emergent behaviours.
Allen Pike has a great article here discussing lots of ways we may see non-chat UI patterns start to change as designers figure out ways to integrate LLM functionality. Examples go back to Maggie Appleton's piece 2 years ago about how different daemons (Language Model Sketchbook, or Why I Hate Chatbots) could help you, with different personalities (like a devil's advocate, or a synthesiser). This article looks at examples where the flexibility of typed or voice input, or automating more ambiguous tasks, can lead to interesting new design patterns.
Security guru and general all-round perceptive commentator Bruce Schneier discusses a useful way to evaluate where current AI tools can help, with tasks that require one of: speed, scale, breadth of scope and "sophistication" (problems that require processing many separate factors).
Looking for bottlenecks in speed, scale, scope and sophistication provides a framework for understanding where AI provides value, and equally where the unique capabilities of the human species give us an enduring advantage.
Working with Google Gemini wearing Snap augemented reality spectacles
A nice demo from Matthew Hallberg, a design engineer at Snap, showing how Google Gemini can integrate with the Snap Spectacles (possibly the new ones coming next year) to perform various tasks within the field of view, outputting correctly anchored labels.
Why I don’t think AGI is right around the corner
Dwarkesh saying an obvious thing that needs saying: the instance of an LLM you're working with doesn't (yet) learn the way a person does over their lifetime; it is fixed.
How do you teach a kid to play a saxophone? You have her try to blow into one, listen to how it sounds, and adjust. Now imagine teaching saxophone this way instead: A student takes one attempt. The moment they make a mistake, you send them away and write detailed instructions about what went wrong. The next student reads your notes and tries to play Charlie Parker cold. When they fail, you refine the instructions for the next student.
This just wouldn’t work. No matter how well honed your prompt is, no kid is just going to learn how to play saxophone from just reading your instructions. But this is the only modality we as users have to ‘teach’ LLMs anything.