NEW BOOK!
Explore a better way to work – one that promises more calm, clarity, and creativity.

AI and Work (Some Predictions)

One of the main topics of this newsletter is the quest to cultivate sustainable and meaningful work in a digital age. Given this objective, it’s hard to avoid confronting the furiously disruptive potentials of AI.

I’ve been spending a lot time in recent years, in my roles as a digital theorist and technology journalist, researching and writing about this topic, so it occurred to me that it might be useful to capture in one place all of my current thoughts about the intersection of AI and work.

The obvious caveat applies: these predictions will shift — perhaps even substantially — as this inherently unpredictable sector continues to evolve. But here’s my current best stab at what’s going on now, what’s coming soon, and what’s likely just hype.

Let’s get to it…


Where AI Is Already Making a Splash

When generative AI made its show-stopping debut a few years ago, the smart money was on text production becoming the first killer app. For example, business users, it was thought, would soon outsource much of the tedious communication that makes up their day — meeting summaries, email, reports — into AI tools.

A fair amount of this is happening, especially when it comes to lengthy utilitarian communication where the quality doesn’t matter much. I recently attended a men’s retreat, for example, and it was clear that the organizer had used ChatGPT to create the final email summarizing the weekend schedule. And why not? It got the job done and saved some time.

It’s becoming increasingly clear, however, that for most people the act of writing in their daily lives isn’t a major problem that needs to be solved, which is capping the predicted ubiquity of this use case. (A survey of internet users found that only around 5.4% had used ChatGPT to help write emails and letters. And this includes the many who maybe experimented with this capability once or twice before moving on.)

The application that has instead leaped ahead to become the most exciting and popular use of these tools is smart search. If you have a question, instead of turning to Google you can query a new version of ChatGPT or Claude. These models can search the web to gather information, but unlike a traditional search engine, they can also process the information they find and summarize for you only what you care about. Want the information presented in a particular format, like a spreadsheet or a chart? A high-end model like GPT-4o can do this for you as well, saving even more extra steps.

Smart search has become the first killer app of the generative AI era because, like any good killer app, it takes an activity most people already do all the time — typing search queries into web sites — and provides a substantially, almost magically better experience. This feels similar to electronic spreadsheets conquering paper ledger books or email immediately replacing voice mail and fax. I would estimate that around 90% of the examples I see online right now from people exclaiming over the potential of AI are people conducting smart searches.

This behavioral shift is appearing in the data. A recent survey conducted by Future found that 27% of US-based respondents had used AI tools such as ChatGPT instead of a traditional search engine. From an economic perspective, this shift matters. Earlier this month, the stock price for Alphabet, the parent company for Google, fell after an Apple executive revealed that Google searches through the Safari web browser had decreased over the previous two months, likely due to the increased use of AI tools.

Keep in mind, web search is a massive business, with Google earning over $175 billion from search ads in 2023 alone. In my opinion, becoming the new Google Search is likely the best bet for a company like OpenAI to achieve profitability, even if it’s not as sexy as creating AGI or automating all of knowledge work (more on these applications later).

The other major success story for generative AI at the moment is computer programming. Individuals with only rudimentary knowledge of programming languages can now produce usable prototypes of simple applications using tools like ChatGPT, and somewhat more advanced projects with AI-enhanced agent-style helpers like Roo Code. This can be really useful for quickly creating tools for personal use or seeking to create a proof-of-concept for a future product. The tech incubator Y Combinator, for example, made waves when they reported that a quarter of the start-ups in their Winter 2025 batch generated 95% or more of their product’s codebases using AI.

How far can this automated coding take us? An academic computer scientist named Judah Diament recently went viral for noting that the ability for novice users to create simple applications isn’t new. There have been systems dedicated to this purpose for over four decades, from HyperCard to VisualBasic to Flash. As he elaborates: “And, of course, they all broke down when anything slightly complicated or unusual needs to be done (as required by every real, financially viable software product or service).”

This observation created major backlash — as does most expressions of AI skepticism these days — but Diament isn’t wrong. Despite recent hyperbolic statements by tech leaders, many professional programmers aren’t particularly worried that their jobs can be replicated by language model queries, as so much of what they do is experience-based architecture design and debugging, which are unrelated skills for which we currently have no viable AI solution.

Software developers do, however, use AI heavily: not to produce their code from scratch, but instead as helper utilities. Tools like GitHub’s Copilot are integrated directly into the environments in which these developers already work, and make it much simpler to look up obscure library or AI calls, or spit out tedious boilerplate code. The productivity gains here are notable. Programming without help from AI is rapidly becoming increasingly rare.


The Next Big AI Application

Language model-based AI systems can respond to prompts in pretty amazing ways. But if we focus only on outputs, we underestimate another major source of these models’ value: their ability to understand human language. This so-called natural language processing ability is poised to transform how we use software.

There is a push at the moment, for example, led by Microsoft and its Copilot product (not to be confused with GitHub Copilot), to use AI models to provide natural language interfaces to popular software. Instead of learning complicated sequences of clicks and settings to accomplish a task in these programs, you’ll be able to simply ask for what you need; e.g., “Hey Copilot, can you remove all rows from this spreadsheet where the dollar amount in column C is less than $10 dollars then sort everything that remains by the names in Column A? Also, the font is too small, make it somewhat larger.”

Enabling novice users to access to expert-level features in existing software will aggregate into huge productivity gains. As a bonus, the models required to understand these commands don’t have to be nearly as massive and complicated as the current cutting-edge models that the big AI companies use to show off their technology. Indeed, they might be small enough to run locally on devices, making them vastly cheaper and more efficient to operate.

Don’t sleep on this use case. Like smart search, it’s also not as sexy as AGI or full automation, but I’m increasingly convinced that within the next half-decade or so, informally-articulated commands are going to emerge as one of the dominate interfaces to the world of computation.


What About Agents?

One of the more attention-catching storylines surrounding AI at the moment is the imminent arrival of so-called agents which will automate more and more of our daily work, especially in the knowledge sectors once believed to be immune from machine encroachment.

Recent reports imply that agents are a major part of OpenAI’s revenue strategy for the near future. The company imagines business customers paying up to $20,000 a month for access to specialized bots that can perform key professional tasks. It’s the projection of this trend that led Elon Musk to recently quip: “If you want to do a job that’s kinda like a hobby, you can do a job. But otherwise, AI and the robots will provide any goods and services that you want.”

But progress in creating these agents has recently slowed. To understand why requires a brief snapshot of the current state of generative AI technology…

Not long ago, there was a belief in so-called scaling laws that argued, roughly speaking, that as you continued to increase the size of language models, their abilities would continue to rapidly increase.

For a while this proved true: GPT-2 was much better than the original GPT, GPT-3 was much better than GPT-2, and GPT-4 was a big improvement on GPT-3. The hope was that by continuing to scale these models, you’d eventually get to a system so smart and capable that it would achieve something like AGI, and could be used as the foundation for software agents to automate basically any conceivable task.

More recently, however, these scaling laws have begun to falter. Companies continue to invest massive amounts of capital in building bigger models, trained on ever-more GPUs crunching ever-larger data sets, but the performance of these models stopped leaping forward as much as they had in the past. This is why the long-anticipated GPT-5 has not yet been released, and why, just last week, Meta announced they were delaying the release of their newest, biggest model, as its capabilities were deemed insufficiently better than its predecessor.

In response to the collapse of the scaling laws, the industry has increasingly turned its attention in another direction: tuning existing models using reinforcement learning.

Say, for example, you want to make a model that is particularly good at math. You pay a bunch of math PhDs $100 an hour to come up with a lot of math problems with step-by-step solutions. You then take an existing model, like GPT-4, and feed it these problems one-by-one, using reinforcement learning techniques to tell it exactly where it’s getting certain steps in its answers right or wrong. Over time, this tuned model will get better at solving this specific type of problem.

This technique is why OpenAI is now releasing multiple, confusingly-named models, each seemingly optimized for different specialties. These are the result of distinct tunings. They would have preferred, of course, to simply produce a GPT-5 model that could do well on all of these tasks, but that hasn’t worked out as they hoped.

This tuning approach will continue to develop interesting tools, but it will be much more piecemeal and hit-or-miss than what was anticipated when we still believed in scaling laws. Part of the difficulty is that this approach depends on finding the right data for each task you want to tackle. Certain problems, like math, computer programming, and logical reasoning, are well-suited for tuning as they can be described by pairs of prompts and correct answers. But this is not the case for many other business activities, which can be esoteric and bespoke to a given context. This means many useful activities will remain un-automatable by language model agents into the foreseeable future.

I once said that the real Turing Test for our current age is an AI system that can successfully empty my email inbox, a goal that requires the mastery of any number of complicated tasks. Unfortunately for all of us, this is not a test we’re poised to see passed any time soon.


Are AGI and Superintelligence Imminent?

The Free Press recently published an article titled “AI Will Change What it Means to Be Human. Are We Ready?”. It summarized a common sentiment that has been feverishly promoted by Silicon Valley in recent years: that AI is on the cusp of changing everything in unfathomably disruptive ways.

As the article argues:

OpenAI CEO Sam Altman asserted in a recent talk that GPT-5 will be smarter than all of us. Anthropic CEO Dario Amodei described the powerful AI systems to come as “a country of geniuses in a data center.” These are not radical predictions. They are nearly here.

But here’s the thing: these are radical predictions. Many companies tried to build the equivalent of the proposed GPT-5 and found that continuing to scale up the size of their models isn’t yielding the desired results. As described above, they’re left tuning the models they already have for specific tasks that are well-described by synthetic data sets. This can produce cool demos and products, but it’s not a route to a singular “genius” system that’s smarter than humans in some general sense.

Indeed, if you look closer at the rhetoric of the AI prophets in recent months, you’ll see a creeping awareness that, in a post-scaling law world, they no longer have a convincing story for how their predictions will manifest.

A recent Nick Bostrom video, for example, which (true to character) predicts Superintelligence might happen in less than two years (!), adds the caveat that this outcome will require key “unlocks” from the industry, which is code for we don’t know how to build systems that achieve this goal, but, hey, maybe someone will figure it out!

(The AI centrist Gary Marcus subsequently mocked Bostrom by tweeting: “for all we know, we could be just one unlock and 3-6 weeks away from levitation, interstellar travel, immortality, or room temperature superconductors, or perhaps even all four!”)

Similarly, if you look closer at AI 2027, the splashy new doomsday manifesto which argues that AI might eliminate humanity as early as 2030, you won’t find a specific account of what type of system might be capable of such feats of tyrannical brilliance. The authors instead sidestep the issue by claiming that within the next year or so, the language models we’re tuning to solve computer programming tasks will somehow come up with, on their own, code that implements breakthrough new AI technology that mere humans cannot understand.

This is an incredible claim. (What sort of synthetic data set do they imagine being able to train a language model to crack the secrets of human-level intelligence?) It’s the technological equivalent of looking at the Wright Brother’s Flyer in 1903 and thinking, “well, if they could figure this out so quickly, we should have space travel cracked by the end of the decade.”

The current energized narratives around AGI and Superintelligence seem to be fueled by a convergence of three factors: (1) the fact that scaling laws did apply for the first few generations of language models, making it easy and logical to imagine them continuing to apply up the exponential curve of capabilities in the years ahead; (2) demos of models tuned to do well on specific written tests, which we tend to intuitively associate with intelligence; and (3) tech leaders pounding furiously on the drums of sensationalism, knowing they’re rarely held to account on their predictions.

But here’s the reality: We are not currently on a trajectory to genius systems. We might figure this out in the future, but the “unlocks” required will be sufficiently numerous and slow to master that we’ll likely have plenty of clear signals and warning along the way. So, we’re not out of the woods on these issues, but at the same time, humanity is not going to be eliminated by the machines in 2030 either.

In the meantime, the breakthroughs that are happening, especially in the world of work, should be both exciting and worrisome enough on their own for now. Let’s grapple with those first.

####

For more of my thoughts on AI, check out my New Yorker archive and my podcast (in recent months, I often discuss AI in the third act of the show).

For more on my thoughts on technology and work more generally, check out my recent books on the topic: Slow Productivity, A World Without Email, and Deep Work.

10 thoughts on “AI and Work (Some Predictions)”

  1. I agree with natural language commands to be an awesome use-case that people are sleeping on!

    I have two examples of messing around with ai where I realized if they just got a little bit better at this, it would change everything.

    First, ai can already identify information in photos. I uploaded a picture of myself to ChatGPT and it started describing how I look. That was a neat trick. But once you can upload screenshots and it can read, transcribe, or act on every piece of information in the photo? Hey here is a screenshot of my spreadsheet, tell me how to fix it. Hey here is a screenshot of a web design that I like, please convert it into code for me.

    Second, using Cursor.ai IDE, I can ask the agents how to use the program and where things are in the program and of course it can write code, save files, delete files, make directories, prompt commands, and be a search engine all in one. I didn’t know where some of the stuff in the program was and it just told me where it was. I didn’t know how everything worked in the program and it just told me how it works. And then of course it also codes and makes commands for everything I ask.

    So imagine a future where in every single program you use on the computer, you can quickly screenshot it and ask how to do something or how to fix something and it tells you exactly how to do it, or just does it for you.

    Now I feel like all the talk about real agents is kind of hyped up? Like just because my IDE can do IDE tasks for me people seem to think agents are going to do my taxes for me and answer all my emails for me automatically. But I think that’s kind of just a dream. “Agent” seems to just be the new term for a long complicated process that had been automated. But I bet instead of the “agent” figuring all that out for itself, it’s going to happen the old fashioned way, where a bunch of engineers work very hard to automate a task. You know, like we’ve been doing for decades already.

    Reply
    • The photo analysis example is a great one. I think of that as part of the general idea of a super search, where now, the sources for its answer isn’t just the web, but also other things, like photos, or files on your local machine, etc. Sort of like a Google for everything which is, it seems to me, likely where the bulk of the money is actually waiting in the AI race.

      Reply
  2. I agree that superintelligence in 2030 is very unlikely, and it’s good to communicate that. But I think superintelligence in 2040 is totally in the territory of “maybe, maybe not, who knows”, and by not emphasizing or mentioning that, you’re burying the lede.

    Anything that humans, and groups of humans, and societies of humans, can do, can also be done by code running on chips. After all, the human brain is a machine that runs an algorithm. No one knows the algorithm, but as soon as someone figures it out, they could run it on chips, like any other algorithm. Then we’d get AIs that can autonomously found and run billion-dollar companies, that can invent science and technology and language from scratch, and all the other crazy sci-fi-sounding stuff. And there could be hundreds of millions of such AIs, thinking much faster than humans, etc.

    Why do I say that “superintelligence by 2040” is in “maybe, maybe not, who knows” territory? Well, LLMs didn’t exist as recently as 2018. The entirety of deep learning was a backwater as recently as 2012. You can say that LLMs are not on the path to superintelligence, and as it happens, I’ll agree with you. But 15 years is evidently more than enough time for the AI research community to have pivoted to something entirely different and poured billions of dollars, and tens of thousands of person-years, into polishing and optimizing it.

    As an analogy, suppose a giant fleet of alien ships were definitely due to arrive on Earth at some point, but we didn’t know when, and 15-40 years seemed like a reasonable guess. But some people were going around saying the aliens would definitely arrive in <5 years. Then it would be good to correct those people! But it would be bad to imply that therefore we shouldn’t be thinking about and preparing for the alien invasion right now.

    Reply
    • First thought on reading the article – humans are general intelligences that are trained. People are capable of having multiple specialisms, but I know from own experience that time spent learning / training on one field can come at the cost of previous knowledge.

      And if a neural net is a reasonable representation of a brain, this is entirely understandable – everything we learn is rebalancing the weights in a system, rather than an ever expanding set of storage.

      Potentially, the future for AI companies are ASIs that are certified experts in their fields, rather than chasing a dream of an expert in everything.

      (Of course, to the end user it doesn’t matter if it’s one AI or an AI that defers the answer to a team of specialists. Just that as an approach, having an army of experts might be easier to create than a da Vinci)

      Secondly – I remain unconvinced by the ‘brains are just software running on meat’ idea. For certain intellectual tasks this is true, but mostly what I’m doing is reacting to the environments and people around me. My child jabbering about the Titanic, or my wife briefly touching the back of my neck. The taste of a meal that’s been slowly cooking for 2 hours. Listening to a new piece of music 5 times, and then listening to something else.

      You’ll notice that few of these are economic activities, but they are all things that humans and groups of humans will continue to do better than AI – because we are embodied consciousnesses, not just the neural network.

      Reply
  3. The death of Moore’s Law (measuring transistors) was the birth of LLM scaling / virtual quantity (measuring virtual parameters). The death of LLM scaling / virtual quantity will be virtual quality. The manifestation of the thing that is manifesting has been going on for 70+ years and there have been dozens of barriers that were said to be impassible, but the manifestation has always found a new thing. Now that the manifestation has manifested the ability to create PhD’s out of thin air to work for its manifestation, now is when you are going to doubt it ?? Humanity is a small thing, and the physical universe is a shallow one. Now some wave from a bigger pond is washing into ours.

    Sadhguru has a nice take on this current affair https://youtu.be/uGwKU1jaDsY?si=J29PtlB8BuB8iTnx&t=2525

    Reply
  4. What I am most anxious to see play out is how the next generation of youth engage with Ai tools.

    The kids who are under 12 today, how will they perform in a world of automation having never known anything other than Ai ubiquity?

    As a parallel example, take social media. When Facebook launched, a generation of Xennials and first-wave Millennials initially used it as a tool to complement their pre-existing social lives (think wall posts and group events). In other words, when a generation of people with pre-established social skills adopt a new comms tech/device, they see it as complementary to their social world. They lived through the transition.

    It’s not until later that we see Gen Z struggling with social media, struggling to establish social skills through the devices and platforms. Why? Because Gen Z didn’t grow up in a world without them. They didn’t understand that those social apps were never meant to create social skills and connections, but only to complement the skills and connections you already had.

    So going back to Ai, many of the people adopting Ai tools and praising their capabilities–those people already had critical thinking skills and domain expertise in the first place! They are using it to complement their pre-existing work!

    What is completely unpredictable is how the untrained, unknowledgeable generation of youth coming up next will be able to perform in a world in which whatever skills and knowledge they manage to pick up in school are basically deemed low value because, well, “Ai can already do that”. What then for those kids?

    Reply
  5. I agree that there has to be regulation especially on the environmental end. That being said, I think we’ll see a lot more people scale back to analogue tools, especially in the wake of Jonathan Heidt’s Anxious Generation book and AI’s impact on education and especially higher education. This is what happens when data centers take up entirely too much environmental resources.: https://bsky.app/profile/chaclo.bsky.social/post/3lpqxwtc77s2s

    Reply
  6. Just when it feels like the existential walls are closing in, a nice post at SH provides a little bit of perspective. Cheers.

    Reply
  7. Great article! Medical billing can be such a complex process, and it’s always helpful to have clear insights into how it works. I appreciate the tips you shared, especially the emphasis on the importance of accurate coding and timely claim submissions. It really highlights how staying organized and up-to-date with billing procedures can make a big difference in streamlining payments and avoiding denials. Looking forward to more posts that break down these complicated topics!
    The Nation’s Largest Network of Medical Billers and Revenue Managers! At American Business Systems, we’ve built our business on integrity and old-fashioned family values. We’ve become successful by helping others just like you start their own medical billing businesses and reach their financial goals without sacrificing their family life to the stress and time pressures of a job.

    Reply

Leave a Comment