Tom Wigg and Stephen Byrd discuss the accelerating pace of AI breakthroughs, the forces driving them and why the next phase of development may look very different from anything we’ve seen so far.
Read more insights from Morgan Stanley.
----- Transcript -----
Tom Wigg: Welcome to Thoughts on the Market. I’m Tom Wigg, Head of Specialty Sales in the Americas at Morgan Stanley, and a sector specialist in Technology, Media and Telecom.
We wake up every day to new AI product releases, so it’s easy to lose sight of the unprecedented non-linear improvement in AI capabilities. But things are about to get weird.
It’s Tuesday, April 28th at 8am in New York.
The market has been thinking about AI in linear terms. But we need to reframe that assumption of only incremental improvement and think about exponential improvement.
That was my takeaway from a conversation with Stephen Byrd, Global Head of Thematic and Sustainability Research at Morgan Stanley. In our conversation, we zeroed in on Stephen’s bull case for broader AI model improvements.
Tom Wigg: First, I want to talk about one obsession that you’ve been writing about for the last several months – is this idea that we’re going to see nonlinear improvements in the frontier models coming out this spring.
Stephen Byrd: Yes.
Tom Wigg: There’s been, you know, some big headlines around new models, benchmarks coming out publicly. Is this, you know, your bull case playing out on these models? And what are the implications?
Stephen Byrd: Yes! Absolutely, Tom. So we have, to your point, we are obsessed. And I know I’m not shy about that – with the nonlinear rate of AI improvement. It is the most important impact to so many stocks that I can think of in the sense that it can impact all industries, all business models. So, what we’ve been saying for some time is, if you look back over the last couple of years at the relationship between the amount of compute used to train these LLMs and the capabilities, we have a very clear scaling law.
And approximately the law is, if you increase the training compute by 10x, the capabilities of the models go up by 2x. Now, as you and I’ve talked about this a lot; just meditate on that for a moment. I think things are about to get weird in the sense that on the positive side, we’re going to see all kinds of underappreciated capabilities across many industries. So this disruption discussion, I think, is going to spread, but it’s also going to require investors to, kind of, be more thoughtful about what they do with that concept. Meaning you can’t sell everything. In the sense that AI will disrupt some businesses.
I actually think this is healthy in some ways because now it forces investors to really look at each business model and assess which is going to get disrupted, which can get supported and enabled by AI, which are immune. Because there are some business models that actually are immune.
But essentially from here, Tom, I’d say we are expecting through the spring and summer to see multiple models that are able to perform a much greater percentage of the economy at better levels of accuracy at incredibly low cost. Which I know you and I have talked a lot about the cost of actually doing this work from the LLMs.
This is massive. This is going to impact so many industries. I think this is all to the good for the AI infrastructure plays because it shows the importance of getting more intelligence out into the world.
Tom Wigg: So, you mentioned the constraints we’re seeing across compute, memory and power. It seems like most of the CEOs of the labs and hyperscalers are talking about this. Investors are bullish in terms of the ownership in, you know, memory, optical, semi-cap, et cetera. But the question I’m getting more recently is around what’s the ROI on all this spending. And does the market action in these hyperscalers, which have been pretty bearish year-to-date, force a cut on CapEx? So, maybe if you can marry that with what you’re picking up on the ground in terms of compute spend and whether the frenzy still continues, you know, versus the ROI? And, like, what could happen?
Stephen Byrd: Yeah. The short answer – I’m going to go through detail – is I think the bullishness is going to get more bullish over the coming months. And let me walk you through a couple of the mathematics and then just what I’m seeing on the ground to your point, Tom.
So the mathematics. We have a token economics model that looks from the perspective of a hyperscaler or an LLM developer in terms of – if they sell their token at a certain price and you fully load the cost of a data center and all associated costs, financing, you name it – in what are the returns? And the bottom line is the returns are excellent.
The other element we spend a lot of work on, and you and I talk a lot about, is the demand for compute. In this world where the LLMs are increasing in capability and the token usage goes way up with agentic AI, video world models, all that stuff, we think that there is a massive shortage of compute. So, if you’re lucky enough to be a hyperscaler with the compute, with the power, we think that they will have a lot of pricing power on the tokens.
Let me explain why we see price power on the tokens. Now I’m going to flip to the perspective of an adopter. Let me give you just rough mathematics. There was a study last year from one of the big labs showing that on average, an enterprise user using an LLM might be able to replace work that would take about one and a half hours from a human. That would save about $55 of cost. A million tokens, depends on whether you’re looking at input or output – but let’s just call it $5 for a million tokens.
The average usage case today for a fairly complex agentic task in an enterprise setting is in the tens of thousands of tokens. Okay? So let’s just do that math again. $55 of savings. A million tokens cost $5, and a typical agentic usage is far less than the million tokens today, though that will accelerate. The economics are a home run for adopters.
So, we’re in a situation where compute is very scarce. I see pricing power all over the place for those who have the compute and have the power.
Tom Wigg: So, when you put it like that, Stephen, it seems so inevitable and obvious. But I wonder why the hyperscalers are trading the way they are? And when do they see the revenue inflection you’re talking about? Is this like a stay tuned kinda 2026 event? Is this something we have to wait for for 2027-2028?
Like, how do you think this flows through to the extent that the market will get more comfortable that all this free cash flow pressure is worth it on the other side?
Stephen Byrd: Yeah. This is, in short, I think this is a 2026 event. But let me dive into that because what you just asked is so important for so many stocks.
So, let’s talk through this. The capabilities of the models are advancing so fast that the average corporate user is not yet keeping up. There is this gap. But that will happen quickly, and we’re seeing signs from these labs of revenue at the lab level that is accelerating. So that’s a good sign.
What we’re seeing, though, among fast adopters is those adopters who really understand the capabilities are quickly realizing just how economically beneficial there is. An example, one of my best friends founded a software company many years ago. Last month was – that was the last month in which his programmers wrote code. They’re done with writing code.
The efficiency benefits for his business are absolutely massive. But he feels like he’s just scratching the surface, and he’s about as technically capable as anyone I know. He has two PhDs in the subject matter. He’s very, very good.
So long way to say that we’re living in almost two worlds where the fast adopters will show what’s possible. The average utilization for enterprises will still take some time. But I do think that the market will react to what they see from the fast adopters in the sense of – the tangible economic benefits are so big.
Now, on the ground, what I’m seeing on the infrastructure side, my friends in power tell me that a couple months ago is when they saw the sense of urgency from the AI community go up a couple of notches for them to get the infrastructure they need. So they saw this explosion in compute coming. In the last two months, the weekly usage of tokens according to OpenRadar is up a couple hundred percent in a couple months.
So, I do think we’re seeing this. So, this is; it’s happening quickly. What I would say is the market will have these signposts in every industry of early adopters showing this benefit. I think that’s enough for us to start to get bullish. We also… I just think when you look at the demand for compute, the compute numbers need to go up. And with that, you know, everything in the AI value chain, infrastructure value chain, the volumes need to go up.
Tom Wigg: One bear case that I wanted to interrogate was – there’s one view that, yes, there’s a token explosion right now. But it’s because the first use case is coding. Which is inherently, you know, very developer-friendly and token-intensive relative to other knowledge work.
Can you talk about, you know, whether you subscribe to that? Or whether the token intensity will be as high or lower as this expands to other areas of knowledge work in the next several years?
Stephen Byrd: Yeah, it’s a great question. The short version is that, yes, it’s true that software usage is more token intensive. However, what we’re going to be seeing – we’re starting to see it – is in almost every knowledge-based job, we’re going to move to agentic AI. And when we do that, you tend to see an explosion in compute.
Let me walk you through the numbers. There are a couple studies that show essentially when you go from a query-based usage of LLMs to an agentic use for any occupation, you see about a 10x increase in token usage per use of those models. And you can see why.
I’ve anecdotes of some of my friends who are newer to this – who set their agents loose overnight to do non-coding work. And in the morning they get some pretty amazing results. But they also used a lot more tokens than they’d expected … (laughs)
Tom Wigg: And a five grand credit card bill?
Stephen Byrd: Exactly. It’s like maybe next time you put a few parameters around that. But long way to say, it’s agentic across every workflow that I can think of that will still result in an explosion in token demand.
Tom Wigg: It’s definitely a good idea to put some parameters around your agentic workflow.
My thanks to Stephen for that conversation. And thank you for listening. Let us know what you think of the show by leaving us a review where you listen. And if you find Thoughts on the Market worthwhile, tell a friend or a colleague about us today.