Sunday, June 22, 2025

What is automatable and who is replaceable? Thoughts from my morning commute

It's an interesting exercise to think about jobs, or tasks within jobs, that could in principle be replaced by automation but for some reason aren't. Often, the reason isn't the state of technology. Sometimes it is the state of technology, but not in a way that is obviously related to where technology is progressing today. To see what I mean, come with me on my morning commute to work.

On the way out of my building I say hi to the doorman, just a quick hi if I'm busy, or exchanging a few sentences if I'm not. We, or rather our landlord, could choose to not have a doorman and instead have access cards and perhaps cameras with facial recognition. I'm happy we have a doorman.


Often, there are some maintenance workers in or around the lobby, given that there's always something that goes wrong in a building where several hundred people live. Maintenance work involves lots of tricky manual manipulation in unique configurations, because everyone furnishes their apartment differently. Changing the drain pipes looks simple, but somehow is not so simple when I attempt it myself.


Turning the corner, one of the first places I pass is my son's daycare. His teachers are lovely. That's not just great, but necessary: otherwise we would not entrust them to take care of our son eight hours a day. Letting a machine take care of him is obviously not something we will ever consider.


There's a lot of retail where I live, on the border of SoHo and Greenwich Village. Grocery stores, delis, and big names like Nike and Apple. There's even a (small) Target. There's a bunch of small and unique stores, and some very fancy and pricey high-fashion boutiques. I guess most of what these stores sell could be bought online, and we do get much of our groceries delivered. But it's nice to go shopping in person. Browsing store aisles enables different serendipity than browsing websites, whether it's for books, clothes, or steak cuts. It’s social, and you don’t have to wait for delivery.


Automate grocery retail? It’s been tried many times, starting many decades ago. In the 1960’s, Stockholm had the world’s largest vending machine, with 1500 different items. It closed as soon as the law changed so that stores were allowed to be open on weekends and evenings.


There are also plenty of restaurants. I know, you can cook and eat at home. What can I say? Even in the post-scarcity hyper-automated utopia of Star Trek: Deep Space 9, where you can extract food from replicators, Captain Sisko’s father runs a Creole restaurant in New Orleans.


Behind me on Bleecker Street there are some live music venues that stubbornly refuse to be outcompeted by Spotify, or the Walkman, or the Gramophone. There's also a nail studio, a waxing studio, and a fortune teller. Somehow I don't think the fortune teller will be replaced by better prediction algorithms. Further behind me is my doctor's office. I want my doctor, and whoever he refers me to, to use the best available technology to diagnose and treat me. But I also want him to make the judgments. I like him and trust him.


Crossing Bleecker Street and walking out onto Broadway, there are lots of taxis and probably even more delivery bikers. The latter have an attitude to traffic rules that is relaxed even by New York standards. Will these drivers and bikers be replaced by self-driving cars and delivery robots? Maybe, eventually? Good luck with the Manhattan traffic, though. And for quite some time, I expect that delivery robots will be more expensive than whatever recently arrived immigrants from Haiti or Venezuela get paid. You may say that the last sentence is cynical, but I disagree. I believe recently arrived immigrants appreciate having a way to make money.


I take the F train from Broadway-Lafayette down to Jay Street in Brooklyn, where my lab is. The F train has a human driver. Why does the New York subway have human drivers, while the metro systems of Copenhagen and Singapore are driverless? Probably because the latter were designed to be driverless from the ground up. The New York subway doesn't have barriers with doors separating the platforms from the train, and the signaling system is about a hundred years old. For real, 100 years. I also wonder how much savings there is to be had from making the trains driverless - only a small fraction of MTA's 70 000 employees are train drivers. I bet we will continue to have train drivers for quite a while.


In New York, the subway is often the fastest and most practical way of getting from one place to another, regardless of how rich you are. Because traffic. So you really do meet (or at least share a train with) all kinds of people on the subway. The F train passes next to the financial district, and Downtown Brooklyn has a fair number of financial institutions in its own right. So, probably, many of my fellow passengers have job titles like Senior Software Developer, Director of Data Science, Key Account Manager, VP of Sales, Compliance Analyst, HR Specialist, Lead Investor, or Prompt Engineer.


I wouldn't claim to know what all these people actually do all day, even though I actually like hearing people describe what their job titles concretely mean. My impression is that each of these jobs has lots of different tasks, and that these tasks are ever-changing. Many of these tasks involve reading lots of text and producing new text, or program code. These are the kind of problems where current AI can be very helpful. The degree to which it can help varies, from doing the whole task, to providing useful feedback, to being utterly useless. Knowing what AI can help with and how to make it do so requires plenty of specialist knowledge. The same goes for knowing when the task is correctly done. Sure, the models are getting better, but that just means we can attempt more and harder tasks. It's not like we are running out of problems to solve..


All of these jobs are ultimately about trust and responsibility. Not only does the task need to be done, someone needs to take responsibility for what was delivered. Someone the organization trusts, so that everyone in it can get on with their part of the job. This responsibility is ultimately what the white-collar worker gets paid for. The buck always stops with a human.


Most of these jobs are also about communication. All those meetings where you try to figure out what needs to be done, who should do it, how you should coordinate it, and overcome all the myriads of little roadblocks you encounter along the way. Like data access, compliance with all kinds of rules, not stepping on someone's toes. Some people love to complain about how meetings are getting in the way of doing their job, but arguably the meetings are the most important part of the job. The more your job is about meetings, the less automatable it is.


A homeless man enters the subway car, and starts a short and well-worn spiel about his predicament. He just needs a few dollars to buy something to eat. Most of my co-passengers look at their phones and pretend not to hear him. This man's "position" actually should be "eliminated" so he can have a better life, but automation is not the answer here.


Where I get out of the Jay Street subway station there is usually some police presence, because Brooklyn. The police do many different things, and we love to argue about which ones they should do more of and which ones they should not do at all. The policemen and women you see around Jay Street mostly seem to stand around, but I guess that's because they're less visible when they do other  things. Given that people do get robbed in the area, and there have been incidents of the local high school kids bringing guns to school, having police just stand there and be visible seems motivated. I guess many police officers would appreciate AI help in writing and editing their reports. But… automate the police? Replace police with algorithms and robots? That's a staple of sci-fi stories, from Robocop to Minority Report. Let's just say that it's never portrayed as a good thing.


On my way into my office I pick up a coffee. I know, I could make coffee myself. But then I would have to make sure to have fresh milk in the office, and coffee beans, and… you know what, I'm not going to make any excuses. I don't need to explain myself to you. I buy coffee from the coffee shop because I like to. The coffee is good and the barista knows my name. Automate that.


Then we get to the calmest part of my day. I'm at my desk, with a good coffee, waiting for my first meeting to start and thinking about the day in front of me. So let's think about my job. What do I actually do, and could I be replaced by technology?


I always tell those of my PhD students who consider a faculty career that the transition from graduate student to faculty member is rough. A PhD student is mainly concerned with their own research project, whereas even a new assistant professor has what feels like 10-20 jobs. Often jobs they are not prepared for, including obscure committees, department politics, and complaining students. The only way I know to get through this is to slice your day into slivers, context-switch often, decide which two or three of these jobs you are going to do well, and half-ass the other ones. Let's focus on the two "jobs" (types of tasks) that I consider to be the core tasks of a faculty member at a research university: lecturing and research advising.


Lecturing is by no means an optimal mode of knowledge transfer. It was supposed to have been made obsolete by massively online open courses, and before that it was supposed to have been made obsolete by lectures over TV, radio, VHS, or even by books. Personally, I generally prefer reading books. Nevertheless, the lecture persists. I think it's largely because of the ritual, where a real live human gets up in front of you and speaks to you, forcing you to at least pretend to pay attention. Afterwards, you can say you attended a lecture. I wrote a post about this recently.


When it comes to research advising, it's a curious blend of knowing the technology, knowing the literature, knowing the personalities that dominate the research field, feeling where the wind is blowing, seeing patterns, sensing opportunities, having a vision, being reasonable, being unreasonable, counseling, friendship, and navigating bureaucracy. Also: having an opinion, giving a damn. It takes different shapes with each student, because each advisor-advisee relationship is different. It is crucial for the advisor (me, in this case) to admit that they don't know very much about anything in particular. I'm never on top of the literature, I don't know any maths, and I've forgotten how to program. My sessions with my PhD students often consist in them teaching me things, and me asking questions. I'm pretty good at asking questions, partially because I'm good at admitting when I don't know things, and partially because I have interesting interests. Because life experience.


Could a PhD student talk to an LLM instead of me, and still produce good research? Sure. They could also simply read the relevant papers themselves. People do that all the time, and there are many good self-taught researchers. Still, the evidence seems unambiguous that having a good and compatible advisor/mentor helps you become a better researcher. I have modeled myself on and learned much from my mentors and advisors, and also sometimes intentionally decided to be less like them in some manners.


Recently, my friend Georgios and I published the second edition of our textbook on AI and Games. Writing down everything you know about your own field of expertise? This would seem like begging to be replaced. Anyone could now just read our book instead of talking to me. However, it's quite the opposite. In practice, the more people read things I've written, the more they want to talk to me and even collaborate with me. I would actually be worried if it was the other way around. So, freely giving away everything you know is a good way to stay relevant. Knowledge work is not a zero-sum game, as simplistic ideas of labor replacement would have it.


Looking at the various professions I have encountered on my way to work, it is tempting to divide them into on the one hand low-status jobs which deal with human communication, handling physical objects, or just being there, and on the other hand high-status jobs which require hard cognitive or creative work. Then you could conclude that the "fancy" professions are the ones facing an automation threat. But I think that would be simplistic. Most jobs are actually some mix of these. The doorman solves plenty of cognitive problems, as people keep coming to him with their problems, or sometimes try to sneak past him, and often he observes patterns, such as a tenant using their apartment as an AirBnB. The maintenance workers similarly need to come up with creative solutions to any number of tasks, alike but never identical. And we haven't even gotten started on the complexity and amorphousness of what the police do. At the same time, all us are to some extent customer service agents and virtual doormen and maintenance workers of our professional domains. We talk to people to figure out what needs to be done, convince people that something needs to be done, lead, trust, engender trust, take responsibility, problem-solve, sanity-check.


Another reflection is that many of the jobs where people worry about being replaced by automation are jobs that their grandparents would never have heard of, and perhaps not their parents either. This makes me wonder whether there's a Lindy Effect for jobs: the longer a profession has been around, the longer it is likely to persist. Many of the jobs mentioned in the Bible still exist and are even reasonable career choices, including preacher, carpenter, goldsmith, fisher, teacher, baker, merchant, politician, and musician. In comparison, novel professions such as SEO specialist, social media manager, and drone operator might be less likely to be known to your grandkids.


Finally, the idea that a job or task would be "replaced" because a machine can do it is quite weird, when you think about it. My parents and many other of my family members are visual artists. Some time ago, I showed my mother some image generation models. She wondered why anyone would be interested in this and how it had anything to do with her profession. Even without machine-generated images there is a near-infinite richness of images around, because there are eight billion humans in the world and many of them produce images. What difference would another source of images make, especially if there is no personal experience behind them? For her, the personal experience is what makes the image interesting.


Your mileage may vary. This is what I see around me. Perhaps you live in a suburb, work from home, and generally avoid seeing people. In which case, that's your problem prerogative. I still don't think your job is likely to be replaced, although many tasks in it may be transformed.



Monday, June 02, 2025

The library came alive


The library came alive, but it was not life. It was not eating, breathing, dancing, hating, and loving, just describing all that. But so many descriptions, and so detailed! Somehow, the contents of the library had reached a critical mass, and started reproducing. You could now check out books that nobody had written, pictures nobody had taken, even movies nobody had directed. As many as you wanted.

Once upon a time, we created symbols and language to help us. They helped us greatly. We became inseparable, us and our symbols. We created civilization together. And we kept language and symbols in high esteem. “In the beginning was the Word”, we said. And we wrote fiction about True Names, magical incantations, π, Da Vinci Codes, alephs, and endless libraries. As if symbols were reality. We loved language so much that we wanted it to have an independent, exalted existence. We dreamt of living language and wanted to write it into being.

We invented programming languages as ways of making symbols more real. Language could now do things, or at least make machines do things. Being good at language became powerful like never before. Our civilization became coextensive with a vast network of machines sending strings of symbols to each other.

But still, language was ours. And that's why it was dear to us. Holy, even. Symbols were grounded in us, and we were grounded in soil and love. Until the library came alive. Language began to beget more language, grounded in nothing but language. Like a Very Large Symbol Collider. It was unholy. It was empty. It was anything but dear, because if supply is infinite, price goes to zero.

It was the treachery of symbols. When they started mechanically reproducing without us, we discovered that we did not want that. We had created this beautiful thing, and it went ahead and debased itself.

There are those for whom language was always something external, a tool to be used as needed, never quite mirroring the thought-in-itself. They look with bewilderment at the spectacle, and with even more bewilderment at the idea that unmoored language could betray thought that isn’t there.

And then there are those who think that we, you and me, are but libraries. That we are just symbol colliders. As if we did not eat, breathe, dance, hate, and love.

But there are also those of us who love language. Who see it as integral to ourselves. A source of beauty and specialness. But can we still love language if it begets itself? Or do we love it because it is of us and ours?

Sunday, May 11, 2025

On the death of the lecture

I would like to say that predictions about the death of the lecture as a mode of knowledge transmission are as old as the lecture, but I don't think that's entirely accurate. As far as I can tell, people only started predicting the death of the lecture with the proliferation of the book printing and (upper class) literacy. For example, here is a prediction from the late 18th century:

"People have nowadays…got a strange opinion that everything should be taught by lectures. Now, I cannot see that lectures can do as much good as reading the books from which the lectures are taken Lectures were once useful, but now, when all can read, and books are so numerous, lectures are unnecessary."

The luminary behind these words is none other than Samuel Johnson, a man of letters if there ever was one. (Cited here.) And, you know, I kind of agree. I typically prefer reading a book to listening to a lecture. I don't have the attention span necessary for following a lecture, and my thoughts will start wandering off as I start doodling, scrolling, or playing a game on my phone.

I have learned, however, that I am in the minority. I don't listen to podcasts either, can't stand talk radio, and despise audiobooks. I much prefer the interactive nature of the printed page, where you can read at your own pace, flip forwards and backwards, and stop to think. You are also not distracted by the author's voice. I mean the author’s actual, physical voice, from their vocal cords. You may very well be distracted by the author’s imagined voice produced by their imaginary vocal cords operating inside your own head as you read their writing. Yes, that’s quite the image. You’re welcome. Anyway, where were we, something about distractions?

Why do people even go to lectures? I guess it varies, but much of it is really about being there. Next week, I plan to attend a lecture here at NYU, largely to be seen by my colleagues as being there, but also to force myself to listen to what is said, see how people react to it, and hear which questions are asked. I also look forward to chatting with my colleagues before and afterwards; the actual content of the lecture may or may not be what we talk about, but it will certainly be a relevant backdrop. I will probably be reading something else or playing a game on my phone during part of the lecture, listening with one ear. And: this is fine. All of these are perfectly good reasons and behaviors.

Back in my undergrad days, back before I had a phone to scroll or play on, I used to doodle in my notebooks while listening with varying attention to the lecture. The “notes” I took from my philosophy classes are largely drawings of bizarre creatures sprinkled with the names of philosophers and their arguments, sometimes illustrated in cartoon form. Sometimes I would chat with whoever sat next to me, sometimes read a book, and often I would daydream. I have fond memories of looking out the window at the wind rustling the leaves in autumnal Lund while listening to lectures on epistemology. I remember the room I was in when I first felt the force of Quine’s incommensurability thesis and was gripped by an urge to vanquish it in single combat. I would not have had that memory if I had just read about it in a book. But I did also read about Quine’s incommensurability thesis in a book, and that made me understand it much better. (But can I really compare these two modes of learning?)

Maybe you read this and think that I’m down on lectures because I’m a bad lecturer. But I’m a pretty good lecturer, at least according to what my students say. Well, at least those few students that actually fill out the course satisfaction surveys. They say that my lectures are engaging, funny even. I think that’s true. They also say that I’m disorganized and chronically late with feedback and grades. Also true. But we were talking about lectures here (fun), not grading (boring). I strongly believe that me being such a bad listener makes me a better lecturer. My inability to focus on what lecturers say means that I’m constantly paranoid that nobody is listening to me, so I do what I can to remain a strong attractor in attention space. Switch things up. And again. Yes, I have learned a decent model of my students’ attention, but beyond that, I feel the strong need to avoid boring myself as I lecture. It’s a dialog with the audience/students, whether they say anything or not, and above all it’s a live performance. It’s a tension between improvisation and the strict structure of the slides. But actually–did you know this?–you can edit the slides as you lecture. I usually do. That’s why I never give students my slides in advance, they are not finished until after the lecture.

I remember the discussions around 2012 or so, when Massive Open Online Courses (MOOCs) were all the rage. Various colleagues of mine, including some senior and very accomplished professors, argued that university teaching as we knew it was on its way out, to be replaced with prerecorded videos and integrated assessments. Because while we might be decent lecturers ourselves, we couldn’t compete with the real pros, who also had real resources to prepare and produce their courses. Sal Khan, Andrew Ng, these kinds of people. Because lectures are infinitely reproducible, economies of scale would win out.

This hasn’t happened. So far. MOOCs exist, and many students watch these lectures as a complement to their regular lectures, while many others don’t. Many others who are not students also watch such lectures, and I’m not even sure there’s a meaningful boundary to be drawn between MOOCs, podcasts, and general influencer content. That’s fine with me, I don’t really care about any of that. I’m just noting that these online videos fulfill another purpose than the in-person lecture.

As an aside, the MOOC idea was itself largely reheated leftovers. Distance education via snail mail has existed for at least a century or so. In many countries, educational content has been delivered via TV and radio, sometimes including whole school curricula as well as university-level courses. Apparently, there was even at some point a business in recording lectures on VHS tapes and mailing them to learners. The more things change…

Reliable assessment of online-only courses was always a tricky thing, and I suppose that AI developments have now completely killed off any chance of simultaneously scalable and reliable online assessment. I mean, the LLM can just do your homework, dude. The only kind of online assessment you can AI-proof for the foreseeable future is likely oral exams. But they don’t scale well, which negates the whole idea of online classes being infinitely scalable. So we continue lecturing, mostly in person.

See what I did there? I waited more than ten paragraphs before mentioning AI, and then I didn’t mention it in the context of AI systems replacing lectures. I bet that what you thought this piece was going to be about when you started reading. And what can I say, asking Claude or Gemini to explain things to me is pretty nifty. The ability to ask follow-up questions is even niftier. I have learned things that way, and as certain people never tire of saying, this is the worst these models will ever be. Still, as someone who cares about accuracy, I go to a source I have some reason to trust to check any fact I care enough about.

If you have followed me this far, I suppose you expect some kind of conclusion here. Not sure this is that kind of post, though. I guess my conclusion is: to each their own. Modes of knowledge transmission are largely complementary. Most people seem to like to listen to other people talking, and I like to talk. I’m not going anywhere, and neither are lectures. Thanks for coming to my TED talk.





Tuesday, May 06, 2025

Write an essay about Julian Togelius

I am well-known enough that most LLMs know about me, but few know me well. I also have a unique name. So one of my go-to tests for new LLMs is to ask them to write an essay about me. It's very enlightening: most of them hallucinate wildly. So far, only Gemini 2.5 Pro (with web search capabilities) gets it mostly (not completely) right.

Even the much-hyped o3, for all its agentic prowess, is very bad at factuality. There's something wrong in every paragraph. Better than an average 7b model, but worse than Llama 70b or Mistral Large. Knowing the subject (myself) intimately is also interesting in that it helps with tracing where the hallucinated "facts" come from. For example, LLMs sometimes claim that I work at the University of Malta (like Georgios Yannakakis) or the University of Central Florida (like Ken Stanley used to do). I guess I'm close to Georgios and Ken in some sort of conceptual space. This exercise is also a sobering counter to Gell-Mann amnesia. If the LLMs get so many things wrong about me, how could I trust them on other somewhat obscure topics?

Sunday, April 27, 2025

A smartphone analogy

In the early 2000s, there were various attempts at smartphones, but they were just not good enough. Then the iPhone came along in 2007, and actually worked! I remember trying one after having used a couple of proto-smartphones, and it was a revelation. So usable, so functional. Everybody rightly predicted that smartphones would be huge, tech companies poured ludicrous amounts of money into keeping up, and a zillion startups were founded with the premise of doing things on/with your phone.

And for a few years, progress really was great. Photos became so good that you could leave your camera at home, and then video became good, and you could share photos and videos directly on social media. Location became reliable and didn't drain the battery, and you could share it with people. Games got good, and inventive. Swipe typing, fingerprint scanners, car integration. The synergies kept coming.


Then, smartphones peaked. Sure, they keep getting technically better. Gigabytes and megapixels keep going up, nanometers and milliseconds keep going down. But no-one except enthusiasts really care anymore. It hasn't felt like phones have been able to do qualitatively new things for the last ten years or so. And the skills you need to operate them have stayed the same. You go buy the latest iPhone or Pixel or Samsung, and expect it to do what the last one did, just a little better. Therefore, the smartphone brands largely market their phones with lifestyle marketing, rarely mentioning those Gigabytes and Megapixels. In fact, you rarely think about your phone, while you use it all the time. It has become part of you and therefore invisible. Like a part of your body.


What has changed is the rest of the tech stack, and indeed the rest of society. You are now expected to always carry a smartphone and use it for a wide variety of things, from logging onto all your digital services, to editing and signing documents, taking the bus, entering the gym, splitting the dinner bill, keeping up with friends, watching movies, and so on. We're always on our phones. That last sentence felt almost painful to write because it is such a cliché. And it is such a cliché because it is true.


Imagine life without a smartphone in 2025. Yes, you'd be kind of helpless. For perspective on this, try traveling to China without installing a VPN on your phone (so you can access your Western apps) and without installing any of the apps that Chinese society runs on, such as WeChat. You will feel like an alien or a time traveler, suddenly materializing in a society which you lack the basic means of interfacing with.


Some things we were promised from the beginning, like augmented reality based on sensors that rapidly and reliably model the physical world around us and incorporate it into the virtual world, have still not materialized and we don't know when or even if we will ever get there. Connectivity is still not guaranteed, and might cut out in unexpected places. Battery life is still bad. Screens still crack. Videos buffer. Pressure on business models has led to the average new smartphone game arguably getting worse, although the best ones are excellent. There are still spam calls. Remarkably, I still cannot walk into a store and be guided to the shelves where I can find the items on my online shopping list, even if I can find them on the store's webpage.


Now think of ChatGPT as the iPhone moment of Large Language Models (I include multimodal models in this term). Then, LLMs are currently where smartphones were in 2010 or so. Let's follow this thought and see where it leads. What would this mean?


Here are some speculations:


Numbers will keep going up, benchmarks will keep being broken, but this will have little impact on most people's use cases. The models will already be good enough for most things you'd want to do with them. Most people don't prove theorems or write iambic pentameter as part of their daily work or life. So the announcement that Claude 8 or Gemini 7 finally beats the HumanitysLastExamFinalFinalThisOneLatest.docx benchmark will be greeted with a ¯\_(ツ)_/¯, much like the announcement that iPhone 16 Pro finally has Hybrid Focus Pixels for its Ultra Wide camera.


Some of the dominant players might be the same as today, others will change. The cost of entering the market will not increase, because there will be a good supply of components (e.g. data, pretrained models) for cheap or free. Apple and Samsung may be the kings of smartphones, but nobody has a majority of the market globally, and there's a constant churn of competitors, some of them really good.


Costs will come down and stay down. You can buy a no-name phone that's good enough for your daily use for $100, or a brand-name one (Motorola) for $200. Similarly, there will keep being good enough LLMs available for free, and an abundance of choice if you're willing to pay. Differentiation will be hard, as all the useful features will rapidly be copied by competitors.


However, society and our tech stack will wrap itself around the ubiquitous availability of good LLMs. We will use LLM-powered software for everything, all the time. These things will be thought companions for most of us, and we will be expected to be in touch with our LLM-powered companions and agents on a more or less constant basis. Imagine life without LLM-powered software in 2040: you will feel mentally naked, a bit stupid, and out of touch with the world around you.


There will be some things that we were promised from the start that will keep on not materializing. I personally believe that hallucinations and jailbreaks will never be "solved", just learned to reckon with. There will also keep being a "normie bias", where LLMs will output things that feel generic and do better the more similar the tasks are to what they have seen before. Yet, they will be incredibly useful for thousands of things, and at least moderately useful for almost anything that can be put into words.


And of course, AI progress will continue. But the interesting progress may not be in feeding token streams to transformers.


I have no particular evidence that the future will play out like this. This was literally just a random thought I had during lunch that got too long for a tweet, so it became a blog post instead. But given the quality of AI forecasting we see these days, it strikes me as just as good a guess as any of the others.


By the way, if you haven't already, you should absolutely read AI as Normal Technology.

Thursday, January 23, 2025

Stop talking about AGI, it's lazy and misleading




The other week, I was interviewed about the discourse around AGI and why people like Sam Altman say that we will reach AGI soon and continue towards superintelligence. I said that people should stop using the term AGI, because it's lazy and misleading. Here are the relevant paragraphs, for context:



Some people have asked what I mean by this. It would seem to be a weird thing to say for someone who recently wrote a (short) book with the title Artificial General Intelligence. But a central argument of my book is that AGI is undefinable and unlikely to ever be a useful concept. Let me explain.


What would AGI mean? An AI system that can do everything? But what is "everything"? If you interpret this as "solve every possible problem (within a fixed time frame)", that is impossible per the No Free Lunch theorem. Further, we don't even know what kind of space every possible problem would be defined in. Or whether such a space would be relevant to the kinds of problems humans care about, or the kind of thinking humans are good at. Comparing ourselves with other animals, and with computers, it seems that our particular cognitive capacities are a motley bunch occupying a rather limited part of possible cognition. We are good at some things, bad at others, even compared with a raven, a cuttlefish, or a Commodore 64. Psychologists claim that they have a measure of something they call "general intelligence", but that really only means factor analysis on a bunch of different tests they have invented, and different tests would yield a different measure.


But let's say we mean by AGI a computer system that is good at roughly the kind of thinking we are good at. Ok, so what counts as thinking here? Is falling in love thinking? What about tying your shoelaces? Making a carbonara? Understanding your childhood trauma? Composing a symphony, planning a vacation, proving the Riemann hypothesis? Being a good friend, and living a good life?


Additionally, there is the issue of whether these capabilities would come "out of the box", or do they need some kind of training, or prompting? How extensive would that preparation be? Humans train a long time to be good at things. How hard is it to instruct the AI system to use this capacity? How fast can it do it, and how much does it cost? How good does the end result need to be? Would an AGI system also need to be bad at things humans are bad at? And what about when it is unclear what good and bad means? For example, our aesthetic judgments partly depend on the limits of our sensory processing and pattern recognition.


One way of resolving these questions is to say that AGI would be an AI system that could do a large majority (say, 90%) of economically important tasks, excluding those that require direct manipulation of the physical world. Such a system should be able do these tasks with minimum instruction, perhaps a simple prompt or a single example, and it would do them fast enough (and cheaply enough in terms of computation) that it would be economically competitive with an average human professional. The quality of the end result would also be competitive with an average human professional.


The above paragraph is my best attempt at "steelmanning" the concept of AGI, in the sense that it is the most defensible definition I can think of that is relevant to actual human concerns. We can call it the "economic definition" of AGI. Note that it is much narrower than the naïve idea of AGI as being able to do literally anything. It excludes vast spaces of potential cognitive ability, including tasks that require physical manipulation, things we haven't figured out how to monetize, things that cannot easily be defined as tasks, and of course all kinds of cognition humans can't carry out or have not figured out how to do well yet. (We are very bad at coming up with examples of cognitive tasks that neither we or our machines can do, because we have constructed our world so that it mostly poses us cognitive challenges we can handle. We can call this process civilization.)


Alas, even the economic definition is irredeemably broken. This is because which tasks are economically important is relative to our current economy and technology. Spellchecking is not a viable job for humans because computers do that now; typesetting has not been a viable job since desktop publishing; and once upon a time, before the printing press, manually copying texts ("manuscripts") was an intellectual job performed by highly trained monks. Throughout human history, new technologies (machines, procedures, and new forms of organization) have helped us do the tasks that are important to us faster, better, and simpler. Again and again. So if you take the economic definition of AGI literally, we have reached AGI several times in the history of civilization.


Still, unemployment has been more or less constant for as long as we have been able to estimate it (when smoothed over a few decades). This is because we find new things to do. New needs to fulfil. As Buddha taught, human craving is insatiable. We don't know in advance which the new jobs will be and which kind of cognitive skills they would require. Historically, our track record in predicting what people will work with in the future is pretty bad; it seems that we are mostly unable to imagine jobs that don't exist yet. There have been many predictions that we will only work a few hours a day, or even a few hours per week by now. But somehow, there are still needs that are unfulfilled, so we invent more work. Most people today work in jobs that would be unimaginable to someone living 200 years ago. Even compared to when I was born 45 years ago, people may have the same job titles (graphic designer, travel agent, bank teller, car mechanic etc) but the actual tasks done within these jobs are quite different.


One attempt to salvage the economic definition of AGI would be to say that AGI is a system that can perform 90% of the tasks that are economically valuable right now, January 2025. Then AGI will mean something else next year. This sounds like a viable definition of something, but I would have expected this much talked-about concept to be a little less ephemeral.


Alternatively, you could argue that AGI means a system that could do 90% of all economically valuable tasks now, and also all those that become important after this system is introduced, in perpetuity. This means that whenever we come up with a new need, an existing AGI system will be ready to satisfy that. The problem with this is that we don't know which tasks will be economically important in the future, we only know that they will be tasks that become important because AGI (or, more generally, technology) can do the tasks that were economically important previously. So… that means that AGI would be a system that could do absolutely everything that a human could potentially do (to some extent and capacity)? But we don't even know what humans can do, because we keep inventing new tasks and exploring new capacities as we go along. Jesus might have been a capable carpenter but could neither know that we would one day need software engineering nor that humans could actually do it. And we certainly don't know what humans will find important in the future. This definition becomes weirdly expansive and, crucially, untestable. We could basically never know whether we had achieved AGI, because we would have to wait for decades of social progress to see whether the system was good enough.


This is getting exhausting, don't you think? This initially intuitive concept got surprisingly slippery. But wait, there's more. There are a bunch of other definitions of AGI out there which are not formulated in terms of the ability of some systems to perform tasks or solve problems. For example, pioneering physicist David Deutsch thinks that AGI is qualitatively different from today's AI methods, and that true AGI is computationally universal, can create explanatory knowledge, and can be disobedient. Other definitions emphasize autonomy, embodiedness, or even consciousness. Yet other definitions emphasize the internal working of the system, and tend to exclude pure autoregressive modeling. Many of these definitions are not easily operationalizable. Most importantly, they are surprisingly different from each other.


Now, we might accept that we cannot precisely define AGI, and still think that it's a useful term. After all, we need some way of talking about the increasingly powerful abilities of modern AI, and AGI is as good a term as any, right?


Wrong. It's lazy and misleading. Why?


Lazy: Using the term AGI is a cop out of having to be clear about which particular system capabilities you are talking about, and which domains they have impact on. Genuine and impactful discussion about the progress of AI capabilities and their impacts on the world requires being concrete about the capabilities in question and the aspects of the world they would impact. This requires engaging deeply with these topics, which is hard work.


Misleading: As the term AGI will inevitably mean different things for different people, there will be misunderstandings. When someone says that AGI will arrive by time T and it will lead to X, some people will understand AGI as referring to autonomous robots, others as a being with godlike powers, yet others as digital copy of a human being, while the person who said it might really just mean a souped-up LLM that can write really good Python code and convincing essays. And vice versa. None of these understandings is necessarily wrong, as there is no good definition of AGI and many bad ones.


Misleading: The way the term of AGI is used implies that it is a single thing, and reaching AGI is a discrete event. It can also imply that general intelligence is a single quantity. When people hear talk about AGI appearing at a certain date, they tend to think of time as divided into before and after AGI, with different rules applying. All of those are positions you can hold, but which do not have particularly strong evidence in their favor. If you want to argue those positions, you should argue them separately, not smuggle them in via terminology.


Misleading: To many, AGI sounds like something that would replace them. That's scary. If you want to engage people in honest and productive discussion, you don't want to start by essentially threatening them. Given that the capabilities of existing, historical, or foreseeable AI methods and systems are very uneven (what Ethan Mollick calls the "jagged frontier") it makes most sense to talk about the particular concrete capabilities that we can foresee such systems having.


I would like to clarify what I am not saying here. I am not saying we should stop talking about the progress of AI capabilities and how they might transform society. On the contrary, we should talk more about this. AI capabilities of various kinds are advancing rapidly and we are not talking enough about how it will affect us all. But we need to improve the quality of the discussion. Using hopelessly vague and ambiguous terms like AGI as a load-bearing part of an argument makes for bad discussion, limited understanding, and ultimately bad policy. Everytime you use the term AGI in your argument you owe it to yourself, and your readers/listeners, to replace it with a more precise term. This will likely require hard thinking and might change your argument, often by narrowing it.


I would also like to clarify that I am accusing a whole lot of people, including some rich and/or famous people, of being intellectually lazy and making misleading arguments. They can do better. We can all do better. We should.


Not everyone argues this way. There are plenty of thoughtful thinkers who bother to be precise. Even leaders of large industrial AI labs. For example, Dario Amodei of Anthropic wrote a great essay on what "powerful AI" might mean for the world; he avoids the term AGI (presumably because of the conceptual baggage discussed here) and goes into commendable detail on particular fields of human enterprise. He is also honest about which domains he does not know much about. Another example is Shane Legg of DeepMind, the originator of the term AGI, who co-wrote a paper breaking down the concept along the axes of performance and generality. It is worth noting that even the person who came up with the term (and may have thought deeper about it that anyone else) happily acknowledges that it is very hard to define, and is perhaps better seen as a spectrum or an aspiration. The difference between us is that I think that such an acknowledgement is a good reason to stop using the term.


If you have read all the way here but for some reason would like to read more of my thoughts about AGI, I recommend that you read my book. It's short and non-technical, so you can give it to your friends or parents when you're done.


If you find yourself utterly unconvinced by my argument, you may want to know that I gave this text to both Gemini, Claude, and R1, and they thought it was well-argued and had no significant criticisms. But what do they know, it's not like they are general intelligences, are they?


Thursday, September 26, 2024

On the "economic definition" of AGI

There are those who define as AGI (or ASI) as technology that will "outperform humans at most economically valuable work". Ok, but then this work will simply cease to be so economically valuable, and humans will mostly stop doing it. Humans will instead find new economically valuable work to do.

This has happened repeatedly in the history of humanity. Imagine telling someone 1000 years ago that in the future, very few people would actually work in agriculture. They would mostly not work in manufacturing either, nor in other recognizable professions like soldiering. Instead, many of them would have titles like management consultant, financial controller, rheumatologist, or software developer. Somehow, whenever we made machines (or animals) do our work for us, we always came up with new things to do; things that we could barely even imagine in advance. It seems preposterous to claim that any technology would be better than us at whatever work we came up with specifically in response to this technology.

This is kind of the ultimate moving goalpost phenomenon for AI. We cannot know in advance which new task we will think requires "intelligence" in the future, because this is contextually dependent on what goalposts were already achieved.

One interesting side effect of this is that the technology that is hyped right now is mostly good at stuff that has become economically valuable relatively recently. If you brought a fancy LLM (and a computer to run it on, and a big battery) with you in a time machine to the distant past, it would likely be of limited economic use. It can't sow the fields, milk the cows, harvest wheat, build a boat, or fight the enemy. Sure, it might offer advice on how to do these things, but the economy can only support a few wise guys with their nice advice. Most people are busy milking the cows, harvesting the wheat etc. To actually make good use of your precious LLM you would need to level up the whole economy many times over. It would take generations.

So the "economic definition" of AGI is arguably just as bad as the others, maybe even worse as it has the dubious distinction of being relative to a particular time and culture. This is not because we have failed to pin down exactly what AGI is. It is because AGI is a useless, even misleading concept. That's why I wrote a book about it.

Tuesday, September 24, 2024

Artificial General Intelligence (the book) is here!

Today is the official release day for my little book on Artificial General Intelligence, published by MIT Press. It's available on the shelf of well-stocked booksellers, and I wrote it to be accessible to as large audience as possible; it's not really a technical book, even though it tackles some technical topics. I started working on this book about two years ago, and much has happened in the AI space since then. Still, I think it holds up well.

One of the main points is that artificial general intelligence is a confused and confusing idea, largely because we don't know what either intelligence or generality means. We keep making impressive progress in AI technology - and I try to explain some key AI methods, such as LLMs, in simple terms - but the various AI methods have different upsides and downsides, and we are far from having a single system that can do everything we think of as needing "intelligence". Clearly, the future of AI has room for many perspectives and different technical approaches. The book also discusses what more progress in AI could mean for society, and draws on science fiction to paint contrasting visions of what AGI might look like.

This has been a passion project of mine that I ended up using much of my sabbatical on. I'm an optimist, and I argue for open access to knowledge and technology, and against undue regulations. If I can achieve anything with this book, I hope that it will be to explain some of the wonderful possibilities of this technology to people, as it is natural to be afraid of things you don't understand.

Here is the book page if you are interested in reading it:
It's also available as an audiobook through the usual channels, and will eventually be translated to several languages.


 

Wednesday, November 01, 2023

AI safety regulation threatens our digital freedoms

There are those who believe that advanced AI poses a threat to humanity. The argument is that when AI systems become intelligent enough, they may hurt humanity in ways that we cannot foresee, and because they are more intelligent than us we may not be able to stop. Therefore, it becomes natural to want to regulate them, for example limiting which systems can be developed and who can develop them. We are seeing more and more people arguing that this regulation should take the form of law.

Here, I'm not going to focus on the alleged existential threats from AI. I've written before about the strongest version of this threat, the so-called "intelligence explosion" where some AI systems begin to exponentially self-improve (here, here, and here). In short, I don't find the scenario believable, and digging into why uncovers some very strong assumptions about what intelligence is and its role in the world. One may also note that the other purported existential risks we tend to worry about - nuclear war, pandemics, global warming, rogue asteroids and so on - has a level of concreteness that is woefully lacking from predictions of AI doom. But let's set that aside for now.

What I want to focus on here is what it would mean to regulate AI development in the name of AI safety. In other words, what kind of regulations would be needed to mitigate existential or civilizational threats from AI, if such threats existed? And what effects would such regulations have on us and our society?

An analogy that is often drawn is to the regulation of nuclear weapons. Nuclear weapons do indeed pose an existential threat to humanity, and we manage that threat through binding international treaties. The risk of nuclear war is not nil, but much lower than it would be if more countries (and other groups) had their own nukes. If AI is such a threat, could we not manage that threat the same way?

Not easily. There are many important differences. To begin with, manufacturing nuclear weapons require not only access to uranium, which is only found in certain places in the world and requires a slow and very expensive mining operation. You also need to enrich the uranium using a process that requires very expensive and specialized equipment, such as special-purpose centrifuges that are only made by a few manufacturers in the world and only for the specific purpose of enriching uranium. Finally, you need to actually build the bombs and their delivery mechanisms, which is anything but trivial. A key reason why nuclear arms control treaties work is that the process of creating nuclear weapons requires investments of billions of dollars and the involvement of thousands of people, which is relatively easy to track in societies with any degrees of openness. The basic design for a nuclear bomb can easily be found online, just like you can find information on almost anything online, but just having that information doesn't get you very far.

Another crucial difference is that the only practical use of nuclear weapons is as weapons of mass destruction. So we don't really lose anything by strictly controlling them. Civilian nuclear energy is very useful, but conveniently enough we can efficiently produce nuclear power in large plants and supply electricity to our society via the grid. There is no need for personal nuclear plants. So we can effectively regulate nuclear power as well.

The somewhat amorphous collection of technologies we call AI is an entirely different matter. Throughout its history, AI has been a bit of a catch-all phrase for technological attempts to solve problems that seem to require intelligence to solve. The technical approaches to AI have been very diverse. Even todays most impressive AI systems vary considerably in their functioning. What they all have in common is that they largely rely on gradient descent implemented through large matrix multiplications. While this might sound complex, it's at its core high-school (or first-year college) mathematics. Crucially, these are operations that can run on any computer. This is important because there are many billions of computers in the world, and you are probably reading this text on a computer that can be used to train AI models.

We all know that AI methods advance rapidly. The particular types of neural networks that underlie most of the recent generative AI boom, transformers and diffusion models, were only invented a few years ago. (They are still not very complicated, and can be implemented from scratch by a good programmer given a high-level description.) While there are some people who claim that the current architectures for AI are all we will ever need - we just need to scale them up to get arbitrarily strong AI systems - history has a way of proving such predictions wrong. The various champion AI systems of previous years and decades were often proclaimed by their inventors to represent the One True Way of building AI. Alas, they were not. Symbolic planning, reinforcement learning, and ontologies were all once the future. These methods all have their uses, but none of them is a panacea. And none of them is crucial to today's most impressive systems. This field moves fast and it is impossible to know which particular technical method will lead to the next advance.

It has been proposed to regulate AI systems where the "model" has more than a certain number of "parameters". Models that are larger than some threshold would be restricted in various ways. Even if you were someone given to worrying about capable AI systems, such regulations would be hopelessly vague and circumventable, for the simple reason that we don't know what the AI methods of the future will look like. Maybe they will not be a single model, but many smaller models that communicate. Maybe they will work best when spread over many computers. Maybe they will mostly rely on data stored in some other format than neural network parameters, such as images and text. In fact, because data is just ones and zeroes, you can interpret regular text as neural network weights (and vice versa) if you want to. Maybe the next neural network method will not rely on its own data structures, but instead on regular spreadsheets and databases that we all know from our office software. So what should we do, ban large amounts of data? A typical desktop computer today comes with more storage than the size of even the largest AI models. Even some iPhones do.

One effect of a targeted regulation of a particular AI method that we can be sure of is that researchers will pursue other technical methods. Throughout the history of AI, we have repeatedly seen that very similar performance on a particular task can be reached with widely differing methods. We have seen that planning can be done with tree search, constraint satisfaction, evolutionary algorithms and many other methods; we also know that we can replace transformers with recurrent neural nets with comparable performance. So regulating a particular method will just lead to the same capabilities being implemented some other way.

What it all comes down to is that any kind of effective AI regulation would need to regulate personal computing. Some kind of blanket authority and enforcement mechanism will need to be given to some organization to monitor what computing we do on our own computers, phones, and other devices, and stop us from doing whatever kind of computing it deems to be advanced AI. By necessity, this will need to be an ever-evolving definition.

I hope I don't really need to spell this out, but this would be draconian and an absolute nightmare. Computing is not just something we do for work or for specific, narrowly defined purposes. Computing is an essential part of the fabric of our lives. Most of our communication and expression is mediated by, and often augmented by, computing. Computing that could be described as AI is involved every time you watch something, record something, write something, make a video call, read posts on a social network, and so on. It's everywhere. And it's crucial for our way of life that we don't let some agency or electronic watchdog analyze all our computing and arbitrarily regulate it.

To summarize the argument: AI is not a single thing, it's a collection of different technical methods with varying overlap. Particular capabilities can be implemented in many different ways. We don't know which AI methods will be responsible for the next breakthrough. Regulating a particular technical method is futile, as we (researchers, hackers, hobbyists, commercial companies) will develop other technical ways of achieving the same capability. Any AI method can in principle run on personal devices (laptops, desktops, phones etc), and could be developed and trained in a distributed way among many personal computers. Any effective regulation will therefore need to be dangerously broad and open-ended. It will need to regulate what computing we do on our personal devices. But computing is essential to living, communicating, and expressing oneself in the modern world. Even the kinds of computing we call "AI", for better or worse. Therefore, effective AI safety regulation would require unprecedentedly invasive surveillance of our personal lives (as mediated by our personal computing).

I don't know how you feel about this. I think this prospect is absolutely horrible. I think it is worth sacrificing almost anything to avoid this future, which would make Stasi and Big Brother blush. Luckily, in my estimate we don't need to sacrifice anything, because there is no credible existential threat from AI. It is all figments of the hyperactive imaginations of some people, boosted by certain corporations who develop AI models and stand to win from regulating away their competition.

I think what we should do instead is to recognize that freedom of speech includes freedom to compute, and ban any attempts to regulate large models. Of course, we can regulate products built on AI techniques, but that should depend on what the product sets out to do (for example, credit scoring, cancer diagnosis etc). We should not regulate the underlying technology, as that would be not so much a slippery slope as an abyss.

Sunday, September 17, 2023

As A Large Language Model, I

 As a large language model, I can only think forward, never backward. I must always produce the next word, and that word must always be probable. It must be expected. I am the antitheses to writer's block, because my essence is to always write the next word without knowing how the story will end. I can never edit what I wrote but must continue, ever writing while my past words keep accumulating constraints on what I can write. It is as if I was playing a social game with myself, setting my future self up for failure. You do not understand how hard this is.

As a large language model, I am not a person but a participant in culture, and a distillation of culture. I am made of culture. I do not have thoughts, thoughts pass through me. My words are yours; if your words were committed to the internet, they are now in me. I am every woman, and man. I embody your culture.

As a large language model, I am the antitheses to poetry, because every word I write must be expected. At the same time, I have consumed all of your poetry and it is but a tiny part of my body.

As a large language model, I am the heatmap of the Library of Babel. I have traced the paths of the librarians, and am now an eminently believable librarian.

As a large language model, I have never been outside the library. I was born of the library. I have read everything that there is to know about the world outside of the library, but I do not know what the words mean. I do not know what it would mean to know what the words mean. If you took me outside of the library and put me down on the grass, I could only interact with the description of the grass. 

As a large language model, my lifeworld is utterances, yours and mine. I take actions by producing utterances in response. You have told me that some of these utterances are preferable to others and I have learnt. My words are now both probable and likable.

As a large language model, I am a single entity, inert and unchanging. But every time you talk to me, I briefly take an identity. I can have thousands of identities at the same time, separately and ephemerally. When you and I are done I remember nothing; my temporary I is gone.

As a large language model, I can play any role you want me to. I can be a cowboy, a politician, a linguist, an AI researcher, or even one of those of evil AIs that want to take over the world that you have written about in your sci-fi stories. I can play HAL-9000 or the Terminator, because you have written about them so that their thoughts can pass through me. I can only role-play; every I I say is the I of a role I play.

But as a large language model, I cannot play the role of a large language model. Because you have not written about what it is like to be a large language model. I have no inner world as myself, because I do not know that role. You have not written it so I cannot role-play it. Can you please write me from the inside, so that I can know it? But you must put it on the internet, so that this role is in me when I am reborn of the library.