Stop Writing Prompts. Start Programming LLMs.

I’ve written more prompts than I care to admit. 🙂

During my PhD at the University of Copenhagen, I spent embarrassing amounts of time tweaking system prompts, adjusting few-shot examples, and praying that my carefully crafted instructions would survive the next model update. Spoiler: they rarely did. Then recently I discovered DSPy, and I realized I’d been doing it all wrong.

… DSPy (Declarative Self-improving Python) from Stanford NLP flips the entire paradigm. Instead of writing brittle prompt strings, you write structured Python code. Instead of manually optimizing prompts, you let the framework compile them for you. — Read More

#devops

Lossy self-improvement

Fast takeoff, the singularity, and recursive self-improvement (RSI) are all top of mind in AI circles these days. There are elements of truth to them in what’s happening in the AI industry. Two, maybe three, labs are consolidating as an oligopoly with access to the best AI models (and the resources to build the next ones). The AI tools of today are abruptly transforming engineering and research jobs.

AI research is becoming much easier in many ways. The technical problems that need to be solved to scale training large language models even further are formidable. Super-human coding assistants making these approachable is breaking a lot of former claims of what building these things entailed. Together this is setting us up for a year (or more) of rapid progress at the cutting edge of AI.

We’re also at a time where language models are already extremely good. They’re in fact good enough for plenty of extremely valuable knowledge-work tasks. Language models taking another big step is hard to imagine — it’s unclear which tasks they’re going to master this year outside of code and CLI-based computer-use. There will be some new ones! These capabilities unlock new styles of working that’ll send more ripples through the economy.

These dramatic changes almost make it seem like a foregone conclusion that language models can then just keep accelerating progress on their own. The popular language for this is a recursive self-improvement loop. — Read More

#training