Stratechery Plus Update

  • 2025.47: Gemini! At The Disco

    (Generated by Gemini 3)

    Welcome back to This Week in Stratechery!

    As a reminder, each week, every Friday, we’re sending out this overview of content in the Stratechery bundle; highlighted links are free for everyone. Additionally, you have complete control over what we send to you. If you don’t want to receive This Week in Stratechery emails (there is no podcast), please uncheck the box in your delivery settings.

    On that note, here were a few of our favorites this week.

    1. Gemini 3 Arrives. There was about a week of high hopes preceding the Gemini 3 release, and sure enough, Google’s new model is apparently state of the art across the board (although Anthropic maintains a lead in one coding benchmark). So what does that mean? Wednesday’s Daily Update was exactly what I wanted to read in the Gemini aftermath, as Ben explored the implications for the rest AI ecosystem; this week’s Sharp Tech episode builds on that analysis with an extended discussion of winners and losers from Gemini week. And there is good news: contrary to the claims of many anon accounts on X this week, the Gemini gains aren’t necessarily a death knell for Nvidia or OpenAI, even if, with apologies to Satya Nadella, it appears that Google very much knows its way around the dance floor. Andrew Sharp
    1. The Most Takeable Companies in Tech. Our goal here with Stratechery Plus is to provide rock solid analysis of tech first-and-foremost, but also China, the NBA, etc. Andrew is in the middle of that as the host of Sharp Tech, Sharp China, and Greatest of All Talk. However, at the end of the day, takes still matter: people get bored with spreadsheets and spreadsheet writing; it’s the opinions on the unknown that are the most entertaining and attention-grabbing. To that end, Andrew, an expert takesman, has produced the most important ranking of the year: the most takeable tech companies, from 15 to 1. My take on this very entertaining Article is that the first 14 companies were a preamble to explain why we just can’t stop talking about OpenAI. Ben Thompson

    2. China Is Very Unhappy with Japan’s New Prime Minister. In the past two weeks, Chinese diplomats and pundits have called new Japanese Prime Minister Sanae Takaichi an “evil witch,” an “American running dog,” and warned that her “dirty neck” will be “cut off” if she sticks her head into China’s internal affairs. This week’s episode of Sharp China explores why Takaichi’s comments on Japan’s options in the event of Chinese invasion of Taiwan inspired this reaction, including the history of recent tensions with Japan, and why the Chinese rhetoric is as much a signal to the rest of Asia as it is to Japan. Also discussed: Hasan Piker’s viral travels through China, and thoughts on a Useful Idiot Industrial Complex that continues to thrive in the digital age.  AS

    Stratechery Articles and Updates

    Sharp Text by Andrew Sharp

    Dithering with Ben Thompson and Daring Fireball’s John Gruber

    Asianometry with Jon Yu

    Sharp China with Andrew Sharp and Sinocism’s Bill Bishop

    Greatest of All Talk with Andrew Sharp and WaPo’s Ben Golliver

    Sharp Tech with Andrew Sharp and Ben Thompson

    This week’s Sharp Tech video is on how Apple commoditized mobile carriers.


    Get notified about new Articles


  • Robotaxis and Suburbia

    Listen to this post:

    It was difficult in the beginning to answer the question I got from everyone: what’s it like living in America again? After all, I had been coming back to Wisconsin in the summer for years, and my move back happened in the summer; things mostly felt like more of the same. Then, the leaves started turning colors, the air became chillier, and, as daylight grew shorter I came to relish one decision in particular: living in the suburbs.

    There is much to be said for urban life, and I was certainly spoiled in that regard living in Taipei. It always seemed odd to answer the question “What is the best part about living in Taiwan?” with the word “Convenient”, but that’s the honest truth: everything you needed was within walking distance, the subway was extensive, clean, and reliable, and, once you understood that traffic was governed by the rule of rivers (the bigger you are the more right of way you have), driving really wasn’t that bad either.

    When my parents first moved away from Wisconsin I needed a new place to stay for the summers and, concerned about upkeep of an empty house in the harsh winter, I opted for a downtown condo; it helped that downtown Madison was a beehive of activity with the university and state government, and I liked the idea of walking everywhere. Then came COVID and the summer of 2020, and suddenly downtown wasn’t so busy anymore; I found myself driving more than I expected, and feeling rather sick of condos, which I had lived in my entire adult life. And so, when a house opened up near an old friend, I snapped it up, remodeled it to my liking and then, this past year, decided to live there full time.

    It’s fashionable to hate on the suburbs, particularly for millenials just a bit younger than I am; I was born at the tail-end of Generation X, and my experience in small town Wisconsin was one of leaving the house in the morning on my bike and not coming home until dinner, hopefully in one piece. It was, all things considered, pretty idyllic, but I can imagine that the clampdown on youth freedom that happened over the last few decades, along with the rise of indoor activities like video games and smartphones, made the suburbs feel increasingly alienating and isolating. What a relief to move to the big city, particularly in the 2010’s when Uber came along.

    Uber Resolution

    There were, in the 2010s, few companies more contentious than Uber, and not just because of the scandals and willingness to operate in the gray area of the law. There was a massive debate over whether or not the company was even a viable business. Hubert Horan, in his seemingly never-ending series insisting that Uber would never be profitable, twice attacked me personally (and dishonestly) for believing that Uber would scale into profitability:

    One does not have to immediately accept all of those conclusions to see that Thompson’s various claims suggesting that Uber might someday have a viable welfare enhancing business are not backed by any hard evidence about efficiency advantages or sustainable profitability. All of Uber’s growth has required massive investor subsidies — $2 billion in 2015 and $3 billion in 2016. All of these subsidies have been destroying competitors who are more efficient but can’t withstand years of subsidies from Silicon Valley billionaires. Thompson argues that Uber has grown total market demand and offered greater service options at night. True, but all due to unsustainable predatory subsidies. Thompson says that Uber’s app gives it the great competitive advantage of controlling its customers. False — people don’t like Uber because the app has a neat user interface, people like Uber because the app shows more cabs at lower prices than competitors can offer. All of those cabs and low prices are due to unsustainable, predatory subsidies. Thompson insists “the fact remains that both Uber riders and drivers continue to vote with their feet” justifies his belief that Uber’s approach to regulation is right, but again ignores that they are not voting for the more efficient producer, but for massive service subsidies. Thompson is falsely claiming that Uber’s growth reflects the free choice of consumers in a competitive market. Uber’s predatory subsidies are designed to undermine the processes by which competitive markets help allocate resources, and then to eliminate competition altogether.

    If these benefits were created by legitimate efficiencies, as Thompson imagines, there would be evidence showing how they made Uber more cost competitive, or how they similarly transformed competition in other markets. To refute the points here about Uber’s predatory, market-distorting subsidies, Thompson would need evidence that Uber has scale economies powerful enough to quickly convert $3 billion operating losses into sustainable profits, and evidence that Uber has competitive advantages overwhelming enough to explain driving everyone else out of the industry. Since Thompson does not have any of this evidence, he can’t claim Uber has produced benefits for anyone but itself.

    Well, here we are in 2025, and over the last 12 months Uber has made $4.5 billion in operating profit, and that number is trending upwards (and doesn’t include the significant profits Uber makes from its non-controlling interests in other mobility companies that it gained thanks to its aggressive expansion); no, I didn’t have evidence of that profit in 2017, but I did understand how scale works to transform money-losing software-based Aggregators into profitable behemoths in the long-run.

    Another classic of the Uber bear genre was this 2014 post by NYU finance professor Aswath Damodaran attempting to determine Uber’s true value; the startup had just raised $1.2 billion at a $17 billion valuation, and according to Damodaran’s calculations, “it is difficult to justify a price greater than $10 billion” (his actual valuation was $5.9 billion). Investor Bill Gurley — before his dramatic powerplay that led to the ouster of founder Travis Kalanick — explained what Damodaran got wrong in How to Miss By a Mile: An Alternative Look at Uber’s Potential Market Size:

    The funny thing about “hard numbers” is that they can give a false sense of security. Young math students are warned about the critical difference between precision and accuracy. Financial models, especially valuation models, are interesting in that they can be particularly precise. A discounted cash flow model can lead to a result with two numbers right of the decimal for price-per-share. But what is the true accuracy of most of these financial models? While it may seem like a tough question to answer, I would argue that most practitioners of valuation analysis would state “not very high.” It is simply not an accurate science (the way physics is), and seemingly innocuous assumptions can have a major impact on the output. As a result, most models are used as a rough guide to see if you are “in the ball park,” or to see if a particular stock is either wildly under-valued or over-valued…

    Damodaran uses two primary assumptions that drive the core of his analysis. The first is TAM, and the second is Uber’s market share within that market. For the market size, he states, “For my base case valuation, I’m going to assume that the primary market Uber is targeting is the global taxi and car-service market.” He then goes on to calculate a global estimate for the historical taxi and limousine market. The number he uses for this TAM estimate is $100 billion. He then guesses at a market share limit for Uber – basically a maximum in terms of market share the company could potentially achieve. For this he settles on 10%. The rest of his model is rather straightforward and typical. In my view, there is a critical error in both of these two core assumptions.

    Gurley argued — correctly in retrospect, given that Uber’s gross bookings over the last 12 months were $93 billion in rides and $86 billion in deliveries — that Damodaran failed to consider how a radically better experience could dramatically expand the addressable market, and completely missed the potential for network effects leading to an outsized share of that expanded market.

    I do feel Uber’s effects even out here in the suburbs: when I lived in Madison decades ago, there only seemed to be about five taxis in the whole city, and they were only ever at the airport; now a ride is six minutes away, and I’m sure it would be even shorter if I were more centrally located. That’s particularly appreciated in a place like Wisconsin, not only because of the cold, but also the culture of drinking; the reduction in drunk driving alone has long placed Uber solidly on the “societal good” side of the ledger, at least in my book.

    Full Self-Driving (Supervised)

    Of course I rarely take Ubers: if you’re in the suburbs you drive, and fortunately, I like driving. That’s not the case for everyone, however: while my wife has driven in Taiwan for years, she’s always been nervous about doing the same in America, with its higher speeds, longer distances, and more uncertain directions. That’s why I got her a Tesla: instead of her driving the car, her car drives her.

    I’ve actually dawdled in writing this Article because I wanted to try out v14 of Full Self-Driving (Supervised) first, but it’s been over a month since its release and I still don’t have the Update, so my experience is based on v13. That’s ok, though, because Full Self-Driving (Supervised) is actually pretty amazing. It really does go from origin to destination without intervention pretty much all-of-the-time (v14 reportedly addresses the actually leaving the driveway and parking part of things), although I take over more than my wife does.

    My issue with Full Self-Driving (Supervised) is two-fold: the first is that it is the absolute best worst driver in the world. What I mean is that Full Self-Driving (Supervised) always handles the situation in front of it with aplomb, including tricky merges, construction, etc. I’m particularly impressed at how it stays with traffic, including speeding when appropriate. That’s the best part. The worst part is that Full Self-Driving (Supervised) seems to have zero planning: it will change lanes even though a turn or an exit is half a mile away, which is particularly galling when an exit lane is backed up; if you don’t take over that leads to an embarrassing attempt to merge back in a quarter mile down the road. In other words, Full Self-Driving (Supervised) gets in more messes than it should because of a lack of foresight, but it handles those messes perfectly. As someone who thinks well ahead of my route in an endless pursuit of efficiency this drives me crazy, but honestly, I would take best worst driver over the vast majority of drivers I encounter on the road.

    My second issue is related to why I keep writing out the whole name: the “Supervised” part drives me absolutely batty. Yes, yes, I shouldn’t look at my phone, but is it better to be forced to exit a perfectly competent — more than competent — driving mode to manually steer while sending a text? More galling is when I am looking ahead at a turn — which necessitates turning my head — and get yelled at by my own car to pay attention. I am paying attention, by actually trying to plan more than two steps ahead!

    Regardless, I absolutely do believe that Full Self-Driving (Supervised) is good enough to be Unsupervised, at least in good weather; it’s a bummer to realize that that still may not happen for a long time, and even when it does, the price may be things like actually flowing with traffic, even if it’s a few miles over the speed limit. Even then, however, what exists today — and make no mistake, Full Self-Driving (Supervised), with its ability to follow a route, is a step-change from lane-following adaptive cruise control — is enough to make a meaningful difference to someone like my wife. It’s a lot easier to enjoy the big house and yard when you have the capability to go somewhere else.

    The Convenience Delta

    One challenge I didn’t anticipate was that while trash pickup comes once a week, recycling pick-up is only every other week; that’s a problem given the number of cardboard boxes we go through, mostly from Amazon.

    In all seriousness, Amazon has transformed suburban living. It was always the case that the idea of dashing off to the nearby store was more theory than reality, even when I lived downtown; at a minimum I usually still drove. Next day delivery, however, completely changes the mental calculus: the likelihood I will run out of time to go to the store tips the balance towards just ordering what you need the moment you need it; the next day — and sometimes sooner — it’s on your porch (Walmart deserves a callout here: their delivery is usually even faster if you order something in store).

    Of course it’s nice to not have to worry about your delivery disappearing, or have to cart it up the stairs or in the elevator; you also have the suburban advantage of having places to store supplies, so you don’t run out in the first place. That was always true though — it’s why big box retailers were very much a product of the suburbs — but marrying that advantage to maximum convenience is a big win.

    Food delivery definitely isn’t as good, particularly for the Asian food I sometimes crave; our family has always been one to cook our own food, however, which is of course easier with a big kitchen (and three different types of grills). The better restaurant options are also all downtown, so that’s a minus, but hey, you can always Uber. The broader takeaway is that while there are still certain conveniences that come from a central location, the convenience delta — thanks first and foremost to Amazon — has been dramatically reduced.

    Uber’s Robotaxi Risk

    There is a point to this diary, and it comes back to Uber. Not only was I a bull during Uber’s rise, I’ve also been fairly optimistic about the company’s fortunes when it comes to robotaxis. From an Update late last year:

    Robotaxis are a technology, not a market — a means, not an end, if you will. Markets are defined by demand, and the demand to be tapped is transportation. And, in this market, the dominant player is Uber; no they don’t have their own robotaxis, but from a consumer perspective, they might as well: the rider doesn’t own the vehicle they ride in, they summon it from an app, and they just walk away when the ride is done. The experience — if not the novelty — is the same with a human driver or a robotaxi.

    Moreover, the human drivers come with some big advantages from Uber’s perspective: they bear their own depreciation costs, and can make individual decisions about the marginal rate necessary to provide supply, which is another way of saying that Uber can more easily scale up and down to meet demand by using price as a signal. It is an open question as to whether robotaxis can ever economically scale to meet demand: having enough capacity for peak demand means a lot of robotaxis sitting idle a lot of the time, while maximizing utilization means insufficient supply during peak periods.

    This last point is why my assumption is that Uber will very much be relevant in the robotaxi era: their supply network will be essential for scaling up-and-down within cities, and serving all of the areas that the centralized fleets do not. What is less clear is their long-term profitability, which may be somewhat out of their control.

    That last sentence was about Uber’s diminished bargaining vis-à-vis a centralized robotaxi operator versus individual drivers, and it’s an important one in terms of Uber’s long-term valuation. However, as robotaxis continue to expand — Waymo is now in five cities (three via their own service, two via Uber), Tesla (with human supervisors in the car) in two, and Amazon’s Zoox in one — I do wonder if I am making a similar mistake to Horan and Damodaran.

    First, like Horan, am I too caught up in the current economics of robotaxis? As an apostle of zero marginal costs I am intrinsically allergic to the depreciation inherent in the cars themselves, along with the significant marginal costs in terms of energy and insurance; Uber side-stepped this by offloading those costs to the drivers. Can scale solve this? At some point — Cybercab already points to this future — vehicles will be purpose-built at scale to be robotaxis, and my experience with Full Self-Driving (Supervised) has me convinced that insurance costs will be manageable, not just because of scale, but because there will be fewer accidents.

    Second, like Damodaran, am I limiting my thinking by focusing on the current market — even if that market is already massively larger than the taxi & limo market ever was? The experience of a Waymo is certainly magical; it’s also peaceful, and by removing the human from the equation, provides a sense of safety and security that Uber has always struggled with. This last point could address a major suburban point point, which is kids: the lockdown in kids’ freedom corresponded with a dramatic rise in organized activities, the sheer volume of which leaves lots of parents feeling like unpaid Uber drivers themselves. Some may rely on Uber to solve this problem; it seems likely to me far more would be willing to entrust their children to a Waymo.

    That does still leave the peak demand question: even if kids become a major market, what do all of these rapidly depreciating cars do during the day? And thus we arrive at why Amazon acquire Zoox: the obvious answer is delivery. The only thing better than next day delivery is same day delivery; the only thing better than same day delivery is same hour delivery. The best way to make that happen in a cost-effective way is to have a huge number of robotaxis on the road that actually aren’t making the decision that prices aren’t high enough, at least as long as those prices cover the marginal cost of a trip, which, in the case of a robotaxi, includes energy but not a human.

    Of course you still have to get the package to the doorstep, which is where robots come in; Tesla is explicitly going in this directions. From The Information:

    Optimus is Tesla’s biggest long-term bet. Musk has said there will eventually be more humanoid robots than cars in the world, and that Optimus will one day be responsible for about 80% of Tesla’s market capitalization. Inside Tesla, he’s pushed the Optimus team to find ways to use the robot in tandem with another big, nearer-term bet: the Cybercab, according to a person with direct knowledge.

    That includes Musk’s desire to have the Optimus robot sit in the Cybercab so it can deliver packages. That should be possible: newer versions of the Optimus robot are capable of consistently lifting and moving around with roughly 25-pound objects for three to four hours on a 30 minute charge, another person with direct knowledge said.

    But the connection between the robot’s torso and legs isn’t flexible enough to allow it to seamlessly get in and out of a Cybercab, according to the first person. Tesla would need to redesign the robot to change that or use a different vehicle for deliveries more tailored for Optimus’ shape, that person said.

    This is obviously all still a ways out, but it all feels a lot more possible today than it did even a year ago; relatedly, it feels a lot more uncertain that Uber will have a long-term role to play — and the company may agree! I thought this announcement from Nvidia at GTC Washington D.C. was a bearish indicator for the company:

    Nvidia today announced it is partnering with Uber to scale the world’s largest level 4-ready mobility network, using the company’s next-generation robotaxi and autonomous delivery fleets, the new Nvidia Drive AGX Hyperion 10 autonomous vehicle (AV) development platform and Nvidia Drive AV software purpose-built for L4 autonomy.

    By enabling faster growth across the level 4 ecosystem, Nvidia can support Uber in scaling its global autonomous fleet to 100,000 vehicles over time, starting in 2027. These vehicles will be developed in collaboration with Nvidia and other Uber ecosystem partners, using Nvidia Drive. Nvidia and Uber are also working together to develop a data factory accelerated by the Nvidia Cosmos world foundation model development platform to curate and process data needed for autonomous vehicle development.

    Nvidia Drive AGX Hyperion 10 is a reference production computer and sensor set architecture that makes any vehicle L4-ready. It enables automakers to build cars, trucks and vans equipped with validated hardware and sensors that can host any compatible autonomous-driving software, providing a unified foundation for safe, scalable and AI-defined mobility.

    Uber is bringing together human drivers and autonomous vehicles into a single operating network — a unified ride-hailing service including both human and robot drivers. This network, powered by Nvidia Drive AGX Hyperion-ready vehicles and the surrounding AI ecosystem, enables Uber to seamlessly bridge today’s human-driven mobility with the autonomous fleets of tomorrow.

    The thing about Uber the first time around is that it wasn’t simply providing a fancy app for the taxi & limo market; it was providing an entirely new experience for both drivers and riders that was orthogonal to that market, which let it create a far larger one. This deal with Nvidia envisions a different sort of evolution, where Uber’s existing market slowly becomes autonomous; that’s possible, even if it means significantly higher capital costs for Uber (and cars that cost more, since they are retrofitted instead of purpose-built).

    What is also possible, however, is that Uber gets Uber-ed: a completely new experience for both drivers (as in they don’t exist) and riders (including kids and packages delivered at the marginal cost of energy) ends up being orthogonal to the Uber market, and far larger. Moreover, this market will, for specific qualitative reasons around safety and security, be inaccessible to Uber’s core business, meaning the entire vision of “bringing together human drivers and autonomous vehicles into a single operating network” ends up being a liability instead of an asset.

    The End of Urbanism?

    There are larger sociological and political questions around things like urban versus suburban living, just as there were when suburbs were built out in the first place. I do believe that the suburbs are very much back, and not just because I’m back in the suburbs; what will be a fascinating question for historians is the chicken-and-egg one between technology driving this shift, versus benefiting from it.

    What is worth considering, however, is if the last wave of urbanism, which started in the 1990s and peaked in the 2010s, might be the last, at least in the United States (Asia and its massive metropolises are another story). The potential physical transformation in transportation and delivery I am talking about is simply completing the story that started with entertainment and television in the first wave of suburbia, and then information and interactivity via the Internet, particularly since COVID. There are real benefits to being in person, just like there are to living in the city, but the relative delta to working remote or living in the suburbs has decreased dramatically; meanwhile, offices and urban living can never match the advantages inherent to working from a big home with a big yard.

    Whether or not this is good thing is a separate discussion; I will say it has been good for me, and it’s poised to get even better.


    Get notified about new Articles


  • The Benefits of Bubbles

    Listen to this post:

    It’s funny to remember that a decade ago there were enough people convinced we were in a bubble that I felt compelled to write an Article entitled It’s Not 1999; that was right then, and it’s obviously right now, when we have a clear counter-example: this is a bubble.

    How else to describe a single company — OpenAI — making $1.4 trillion worth of deals (and counting!) with an extremely impressive but commensurately tiny $13 billion of reported revenue? Sure, the actual number may be higher, but that is still two orders of magnitude less than the amount of infrastructure OpenAI has publicly committed to buy over the coming years, and they are not the only big spenders. Over the past week every big tech company (except Apple) has significantly expanded their capital expenditure plans, and there is no sign of anyone slowing down.

    This does, understandably, have people wringing their hands. What goes up must come down, which is to say bubbles that inflate eventually pop, with the end result being a recession and lots of bankrupt companies. And, not to spoil the story, that will almost certainly happen to the AI bubble as well. What is important to keep in mind, however, is that that is not the end of the story, at least in the best case. Bubbles have real benefits.

    Financial Speculation and Physical Capacity

    The definitive book on bubbles has long been Carlota Perez’s Technological Revolutions and Financial Capital.1 Bubbles were — are — thought to be something negative and to be avoided, particularly at the time Perez published her book. The year was 2002 and much of the world was in a recession coming off the puncturing of the dot-com bubble.

    Perez didn’t deny the pain: in fact, she noted that similar crashes marked previous revolutions, including the Industrial Revolution, railways, electricity, and the automobile. In each case the bubbles were not regrettable, but necessary: the speculative mania enabled what Perez called the “Installation Phase”, where necessary but not necessarily financially wise investments laid the groundwork for the “Deployment Period”. What marked the shift to the deployment period was the popping of the bubble; what enabled the deployment period were the money-losing investments.

    In the case of the dotcom bubble, the money-losing investments that mattered were not actually the dotcom companies that mark that era in Silicon Valley lore: yes, a lot of people lost money on insane IPOs, but the loss was mostly equity, not debt. Where debt was a problem was in telecom, where a host of companies went bankrupt after a frenzied period of building far more fiber than could ever be justified by current usage, fast though it may have been growing. That fiber, however, became the background of today’s Internet; the fact that it basically existed for free — because the companies who built it went bankrupt — enabled the effectively free nature of the Internet today.

    The Conditions for Cognitive Capacity

    Late last year Byrne Hobart and Tobias Huber made a new contribution to our understanding of bubbles with their book Boom: Bubbles and the End of Stagnation. While Perez focused on the benefits that came from financial speculation leading to long-term infrastructure, Hobart and Huber identified another important feature of what they called “Inflection Bubbles” — the good kind of bubbles, as opposed to the much more damaging “Mean-reversion Bubbles” like the 2000’s subprime mortgage bubble. First, here is Hobart and Huber’s definition of an inflection bubble:

    Inflection-driven bubbles have fewer harmful side effects and more beneficial long-term effects. In an inflection-driven bubble, investors decide that the future will be meaningfully different from the past and trade accordingly. Amazon was not a better Barnes & Noble; it was a store with unlimited shelf space and the data necessary to make personalized recommendations to every reader. Yahoo wasn’t a bigger library; it was a directory and search engine that made online information accessible to anyone. Priceline didn’t want to be a travel agent; it aspired to change the way people bought everything, starting with plane tickets.

    If a mean-reversion bubble is about the numbers after the decimal point, an inflection bubble is about orders of magnitude. A website, a PC, a car, a smartphone — these aren’t five percent better than the nearest alternative. On some dimensions, they’re incomparably better. A smartphone is a slightly more convenient tool than a PC for taking a photo and quickly uploading it to the internet, but it’s infinitely better at navigation. A car is not just slightly faster and more reliable than a horse (although in the early days of the automobile industry, it was apparently common for pedestrians to yell “Get a horse!” at passing motorists); cars transformed American cities. Modern-day Los Angeles is inconceivable on horseback. The manure problem alone beggars the imagination.

    This is what makes inflection bubbles valuable:

    The fundamental utility of inflection bubbles comes from their role as coordinating mechanisms. When one group makes investments predicated on a particular vision of the future, it reduces the risk for others seeking to build parts of that vision. For instance, the existence of internet service providers and search engines made e-commerce sites a better idea; e-commerce sites then encouraged more ad-dependent business models that could profit from directing consumers. Ad-dependent businesses then created more free content, which gave the ISPs a better product to sell. Each sector grew as part of a virtuous circle.

    What I love about this formulation from a tech perspective is that it captures the other side of the dotcom era: no, Silicon Valley didn’t produce any lasting infrastructure (unless you count a surplus of Aeron chairs), but what the mania did produce were a huge number of innovations, invented in parallel, that unlocked the following two decades of growth.

    First, the dotcom era brought nearly the entire U.S. population online, thanks to that virtuous cycle that Hobart and Huber described in the above excerpt. This not only provided the market for the consumer Internet giants that followed, but also prepared an entire generation of future workers to work on the web, unlocking the SaaS enterprise market.

    Second, the intense competition of the dotcom era led to one of my favorite inventions of all time, both because of its impact and because of its provenance.

    Microsoft famously saw Netscape, the OpenAI of the dotcom era, as a massive threat; the company responded with Internet Explorer, and a host of legally questionable tactics to spur its adoption. What is forgotten, however, is that Microsoft was at that time actually quite innovative in terms of pushing browser technology forward, driven by the need to beat Netscape, and one of those innovations was XMLHttpRequest. XMLHttpRequest, introduced with Internet Explorer 5 in 1999, allowed Javascript to make asynchronous HTTP requests without reloading the page; previously to change anything on a webpage meant reloading the entire thing. Now, however, you could interact with a page and have it update in place, without a reload.

    What makes this invention ironic is that this was the key capability that transformed the browser from a media consumption app to a productive one, and it was the productivity capabilities that began the long breakdown of Microsoft’s application moat. Once work could be done in a browser, it would be done everywhere, not just on Windows; this, in the long run, created the conditions for the smartphone revolution and the end of Windows’ dominance. This was, to be clear, but one of a multitude of new protocols and innovations that made today’s tech stack possible; what is important is how many of them were invented at once thanks to the bubble.

    Third, the cost and complexity of serving all of these new use cases drove tremendous innovation on the backend. The Nvidia of the dotcom era was arguably not Cisco, but rather Sun: a huge percentage of venture capital went to buying Sun SPARC/Solaris servers to run these new-fangled companies. Solaris was the most advanced operating system in terms of running large websites, with the most mature TCP/IP stack, multithreading, symmetric multiprocessing, etc. Moreover, the dominance of Solaris meant that it had the largest pool of developers, which meant it was easier to hire if you ran Solaris.

    The problem, however, is that SPARC servers were extremely expensive, to the point of being nearly financially impractical for the largest-scale web applications like Hotmail or Yahoo. That’s why the former (in its startups days) ran its front-end on free software (FreeBSD) on commodity x86 hardware from the beginning, and why the latter made the same shift as it exploded in popularity. Both, however, had custom-built back-ends; it was Google, founded in 1998, that built the entire stack on commodity x86 hardware and Linux, unlocking the scalability that was critical to the huge growth in the Internet that followed.

    This entire stack was the product of a massive amount of uncoordinated coordination: people came online for better applications that ran on hardware powered by software built by a massive array of companies and individuals; that all of this innovation and invention happened at the same time was because of the bubble.

    Oh, and to return to Perez: all of this ran over fiber laid by bankrupt companies. What Perez got right is that bubbles install physical capacity; what Hobart and Huber added is that they also create cognitive capacity, thanks to everyone pulling in the same direction at the exact same time, based not on fiat, but on a shared belief that this time is different.

    Is AI Different?

    This question — or statement — is usually made optimistically. In this case, the optimistic take would be that AI is already delivering tangible benefits, that those benefits are leading to real demand from companies and consumers, and that all of the money being spent on AI will not be wasted but put to productive use. That may still be the case today — all of the hyperscalers claim that demand for their offerings exceeds supply — but if history is any indication we will eventually overshoot.

    There is, however, a pessimistic way to ask that question: will the AI bubble be beneficial like the positive bubbles chronicled by Perez and Hobart and Huber, or is it different? There have been reasons to be worried about both the physical buildout and the cognitive one.

    Start with the physical: a huge amount of the money being spent on AI has gone to GPUs, particularly Nvidia, rocketing the fabless design company to a nearly $5 trillion valuation and the title of most valuable company in the world. The problem from a Perez perspective is that all of this spending on chips is, relative to the sort of infrastructure she wrote about — railroads, factories, fiber, etc. — short-lived. Chips break down and get superseded by better ones; most hyperscalers depreciate them over five years, and that may be generous. Whatever the correct number is, chips don’t live on as fully-depreciated assets that can be used cheaply for years, which means that to the extent speculative spending goes towards GPUs is the extent to which this bubble might turn out to be a disappointing one.

    Fortunately, however, there are two big areas of investment that promise to have much more long-term utility, even if the bubble pops.

    The first is fabs — the places where the chips are made. I’ve been fretting about declining U.S. capacity in this area, and the attendant dependence on Taiwan, the most fraught geopolitical location in the world, for years, and for much of that time it wasn’t clear that anything would be done about it. Fast forward to today, and not only are foundries like TSMC and Samsung building fabs in the U.S., but the U.S. government is now a shareholder in Intel. There is still a long path to foundry independence for the U.S., particularly once you consider the trailing edge as well, but there is no question that the rise of AI has had a tremendous effect in focusing minds and directing investment towards solving a problem that might never have been solved otherwise.

    The second is power. Microsoft CFO Amy Hood said on the company’s earnings call:

    As you know, we’ve spent the past few years not actually being short GPUs and CPUs per se, we were short the space or the power, is the language we use, to put them in. We spent a lot of time building out that infrastructure. Now, we’re continuing to do that, also using leases. Those are very long-lived assets, as we’ve talked about, 15 to 20 years. And over that period of time, do I have confidence that we’ll need to use all of that? It is very high.

    Amazon CEO Andy Jassy made a similar comment on his company’s earnings call:

    On the capacity side, we brought in quite a bit of capacity, as I mentioned in my opening comments, 3.8 gigawatts of capacity in the last year with another gigawatt plus coming in the fourth quarter and we expect to double our overall capacity by the end of 2027. So we’re bringing in quite a bit of capacity today, overall in the industry, maybe the bottleneck is power. I think at some point, it may move to chips, but we’re bringing in quite a bit of capacity. And as fast as we’re bringing in right now, we are monetizing it.

    As I noted yesterday, this actually surprised me: I assumed that chips were in short supply, and the power shortage was looming, but actually power is already the limiter. This is both disappointing and unsurprising, given how power generation capacity growth has stagnated over the last two decades:

    U.S. electricity generation by major energy source

    At the same time, this is also encouraging: the fastest way to restart growth — and hopefully at an even higher rate than the fifty years that preceded this stagnation — is to have massive economic incentives to build, combined with massive government incentives to eliminate red tape. AI provides both, and my hope is that the fact we are already hitting the power wall means that growth gets started that much sooner.

    It’s hard to think of a more useful and productive example of a Perez-style infrastructure buildout than power. It’s sobering to think about how many things have never been invented because power has never been considered a negligible input from a cost perspective; if AI does nothing more than spur the creation of massive amounts of new power generation it will have done tremendous good for humanity. Indeed, if you really want to push on the bubble benefit point, wiping away the cost of building new power via bankruptcy of speculative investors — particularly if a lot of that power has low marginal fuel costs, like solar or nuclear — could be transformative in terms of what might be invented in the future.

    To that end, I’m more optimistic today than I was even a week ago about the AI bubble achieving Perez-style benefits: power generation is exactly the sort of long-term payoff that might only be achievable through the mania and eventual pain of a bubble, and the sooner we start feeling the financial pressure — and the excitement of the opportunity — to build more power, the better.


    I’ve been less worried about the cognitive capacity payoff of the AI bubble for a while: while there might have been concern about OpenAI having an insurmountable lead, or before that Google being impregnable, nearly everyone in Silicon Valley is now working on AI, and so is China. Innovations don’t stay secret for long, and the time leading edge models stay in the lead is often measured in weeks, not years. Meanwhile, consumer uptake of AI is faster than any other tech product by far.

    What is exciting about the last few weeks, however, is that there is attention being paid to other parts of the stack, beyond LLMs. For example, last week I interviewed Substrate founder James Proud about his attempt to build a new kind of lithography machine as the center of a new American foundry. I don’t know if Proud will succeed, but the likelihood of anyone even trying — and of getting funding — is dramatically higher in the middle of this bubble than it would have been a decade ago.

    It was also last week that Extropic announced a completely new kind of chip, one based not on binary 1s and 0s, but on probabilistic entropy measurements, that could completely transform diffusion models. Again, I don’t know if it will succeed, but I love that the effort exists, and is getting funding. And meanwhile, there are massive investments by every hyperscaler and a host of startups to make new chips for AI that promise to be cheaper, faster, more efficient, etc. All of these efforts are getting funding in a way they wouldn’t if we weren’t in a bubble.

    Hobart and Huber write in Boom:

    Not all bubbles destroy wealth and value. Some can be understood as important catalysts for techno-scientific progress. Most novel technology doesn’t just appear ex nihilo, entering the world fully formed and all at once. Rather, it builds on previous false starts, failures, iterations, and historical path dependencies. Bubbles create opportunities to deploy the capital necessary to fund and speed up such large-scale experimentation — which includes lots of trial and error done in parallel — thereby accelerating the rate of potentially disruptive technologies and breakthroughs.

    By generating positive feedback cycles of enthusiasm and investment, bubbles can be net beneficial. Optimism can be a self-fulfilling prophecy. Speculation provides the massive financing needed to fund highly risky and exploratory projects; what appears in the short term to be excessive enthusiasm or just bad investing turns out to be essential for bootstrapping social and technological innovations…A bubble can be a collective delusion, but it can also be an expression of collective vision. That vision becomes a site of coordination for people and capital and for the parallelization of innovation. Instead of happening over time, bursts of progress happen simultaneously across different domains. And with mounting enthusiasm…comes increased risk tolerance and strong network effects. The fear of missing out, or FOMO, attracts even more participants, entrepreneurs, and speculators, further reinforcing this positive feedback loop. Like bubbles, FOMO tends to have a bad reputation, but it’s sometimes a healthy instinct. After all, none of us wants to miss out on a once-in-a-lifetime chance to build the future.

    This is why I’m excited to talk about new technologies, the prospect for which I don’t know. The more I don’t know projects there are, the more likely there is to be one that succeeds. And, if you want an investment that pays off not for a few years, and not even for a few decades, but literally forever, then your greatest hope should be invention and innovation.

    Stagnation: The Bubble Alternative

    Hobart and Huber actually begin their book not by talking about ancient history, but about this century, and stagnation.

    The symptoms of technological, economic, and cultural stagnation can be detected everywhere. Some of the evidence is hard to quantify, but it can perhaps best be summarized by a simple thought experiment: Will children born today experience as much change as children born a century ago—a time when cars, electrical appliances, synthetic materials, and telephones were still in their infancy? Futurists and science-fiction authors once prophesied an era of abundant energy due to nuclear fission, the arrival of full automation, the colonization of the solar system, the end of poverty, and the attainment of immortality. In contrast, futurists today ask questions about how soon and how catastrophically civilization will collapse.

    There is a science-fiction innovation that has been hovering around the edges of the tech industry for the last decade: virtual and augmented reality. It hasn’t gotten far. Meta has, since it started breaking out Reality Labs financials in Q4 2020, recognized $10.8 billion in revenue against $83.2 billion in costs; the total losses are far higher when you consider that the company bought Oculus VR for $2 billion six years before that breakout. Apple, meanwhile, announced the Vision Pro in 2023, launched it in 2024, and has barely talked about it since — and certainly not on earnings calls.

    Both companies would argue that the technology just isn’t there yet, and to the extent AR and VR are compelling, it’s because of the money and time they have spent developing it. I wonder, however, about a counter-factual where AR and VR were developed by a constellation of startups, not big companies: how much more innovation might there have been? Or, perhaps the bigger problem is that there was not — and, given that all of the investment is a line item in large company budgets, could not be — a bubble around AR and VR.

    More generally, tech simply wasn’t much fun by the time 2020 rolled around. You had your big five tech companies who had each carved out their share of the market, unassailable in their respective domains, and the startup industry was basically itself another big tech company: Silicon Valley Inc., churning out cookie-cutter SaaS companies with a proven formula and low risk. In fact, it’s the absence of risk that Hobart and Huber identify as the hallmark of stagnation:

    Of course, the causes of stagnation are complex. But what these symptoms of stagnation and decline have in common is that they result from a societal aversion to risk, which has been on the rise for the past few decades. Societal risk intolerance expresses itself almost everywhere — in finance, culture, politics, education, science, and technology. Broadly, there seems to be a collective desire to suppress and control all risks and conserve what is at the expense of breaking the terminal horizon of the present and accelerating toward what could be.

    This is why Hobart told me in a Stratechery Interview that Boom was ultimately an exhortation:

    What I took away is your book was much more of a sociological exposé, like spiritual almost, and I shouldn’t say almost because you were actually quite explicit about it. It’s like you were seeking — the goal of this book, it feels like — is to call forth the spirit of the bubble as opposed to have some sort of technocratic overview. You give us useful history, but there’s not really charts or microeconomics, it’s an exhortation. Is that what you were going for?

    Byrne Hobart: Yes, it is an exhortation. We do want people to pick up a copy and quit their job halfway through reading it or drop out of school and start something crazy. I don’t want to be legally liable if you do something sufficiently crazy, and I think that the spiritual element is something that we did want to talk about in the book, because I think if you — you can apply this totally secular framework to it, and it’s perfectly valid. Of course, if it is a mostly materialist framing of things, then it has a lot more real world data because it’s all reliant on that real world data, but if you have the belief or at least suspicion that all of us are unique and special, and there is something that we are, if not put on this Earth to do, at least there are things that we are able to do that other people wouldn’t do as well, that part of our job is to find those things and do them really well. Bubbles play into that in an interesting way because they tell you it’s time, it’s like you wanted to do this kind of thing.

    What is fascinating about the AI bubble is that there is at its core a quasi-spiritual element. There are people working at these labs that believe they are building God; that is how they justify the massive investment in leading edge models that never have the chance to earn back their costs before they are superceded by someone else. That’s why they push for policies that I think are bad for innovation and bad for national security. I don’t like these side effects, to be clear, but I appreciate the importance of the belief and the motivation.

    And, I must say, it certainly is fun and compelling in a way that tech was not a few years ago. Bubbles may end badly, but history does not end: there are benefits from bubbles that pay out for decades, and the best we can do now is pray that the mania results in infrastructure and innovation that make this bubble worth it.


    1. I highly recommend this overview if you are not familiar. 


    Get notified about new Articles


  • Resiliency and Scale

    Listen to this post:

    There seems, at first glance, to be little in common between the two big stories of the last two weeks. On October 9, China announced expansive export controls on rare earths, which are critical to nearly all tech products; then, on October 20, US-East-1, the oldest and largest region of Amazon Web Services, suffered a DNS issue that impacted cloud services that people didn’t even know they used, until they were no longer available.

    There is, however, a commonality, one that cuts to the heart of accepted wisdom about both the Internet and international trade, and serves as a reminder that what actually happens in reality matters more than what should happen in theory.

    US-East-1 and the End of Resiliency

    The Internet story is easier to tell. While the initial motivation for ARPANET, the progenitor of the Internet, was to share remote computing resources, the more famous motivation of surviving a nuclear attack did undergird a critical Internet technology: packet switching. Knocking out one node of the Internet should not break the whole thing, and, technically, it doesn’t. And yet we have what happened this week: US-East-1 is but one node on the Internet, but it is so critical to so many applications that it effectively felt like the Internet was broken.

    The reasoning is straightforward: scale and inertia. Start with the latter: Northern Virginia was a place that, in the 1990s, had relatively cheap and reliable power, land, and a fairly benign natural-disaster profile; it also had one of the first major Internet exchange points, thanks to its proximity to Washington D.C., and was centrally located between the west coast and Europe. That drew AOL, the largest Internet Service Provider of the 1990s, which established the region as data center central, leading to an even larger buildout of critical infrastructure, and making it the obvious location to place AWS’s first data center in 2006.

    That data center became what is known as US-East-1, and from the beginning it has been the location with the most capacity, the widest variety of instance types, and the first region to get AWS’s newest features. It’s so critical that AWS itself has repeatedly been shown to have dependencies on US-East-1; it’s also the default location in tutorials and templates used by developers around the world. You might make the case that “no one got fired for using US-East-1”, at least until now.

    Amazon, meanwhile, has invested billions of dollars into AWS over the last two decades, making the case that enterprises ought not waste their time and money building out and maintaining their own servers: even if the costs penciled out similarly, the flexibility of being able to scale up and scale down instantly was worth shifting capital costs to operational ones.

    The fact that this was not only a winning argument but an immensely profitable one became clear with The AWS IPO, which is how I described the Amazon earnings where they first broke out AWS’s financials. For the first decade of AWS’ existence the conventional wisdom was that only Amazon, with its famous appetite for tiny margins, would be able to stomach similarly narrow margins in the cloud; in fact it turned out that AWS was extremely profitable, and that profitability increased with scale.

    And, nestled within AWS’s scale, was US-East-1: that was the place with the cheapest instances, because it had the most, and that is where both startups and established businesses started as they moved to the cloud. Sure, best practices meant you had redundancy, but best practices are not always followed practices, and when it comes to networking, things can break in weird ways, particularly if DNS is involved.

    The larger lesson, however, is that while the Internet provided resiliency in theory, it also dramatically reduced the costs of putting your data and applications anywhere; then, once you could put your data and applications anywhere, everyone put their data and applications in the place that was both the easiest and the cheapest. That, by extension, only increased the scale of the place where everyone put their data and applications, making it even cheaper and easier. The end result is that, as we saw this week, the Internet in practice is less resilient than it was 20 years ago. Back then data centers went down all of the time, but if that data center served a single customer in an office park it didn’t really matter; now one data center in Northern Virgina is a failure point that affects nearly everyone.

    Rare Earths and China Dependency

    Rare earths are very different from packets that move with the speed of light. You have to build massive mines, separate trace minerals from mounds of dirt, then process and refine them to get something useful. It’s a similar story for physical goods generally: you have to get the raw materials, refine and process them, manufacture components, do final assembly, and then ship them to stores and warehouses until they reach their final destinations in workplaces and homes.

    This process was so onerous that, midway through the last century, only a portion of the world’s countries had ever managed to industrialize, and those that did trod similar paths and developed similar capabilities. Geography mattered tremendously, which is why, to take some classic examples, every country had its own car companies, its own chemical companies, etc. Yes, countries did search and colonize the planet in pursuit of raw materials, but the industrial base was firmly established in the homeland.

    Technology of another sort changed this equation; from 2016’s The Brexit Possibility:

    In the years leading up to the 1970s, three technological advances completely transformed the meaning of globalization:

    • In 1963, Boeing produced the 707-320B, the first jet airliner capable of non-stop service from the continental United States to Asia; in 1970 the 747 made this routine.
    • In 1964, the first transpacific telephone cable between the United States and Japan was completed; over the next several years it would be extended throughout Asia.
    • In 1968, ISO 668 standardized shipping containers, dramatically increasing the efficiency with which goods could be shipped over the ocean in particular.

    These three factors in combination, for the first time, enabled a new kind of trade. Instead of manufacturing products in the United States (or Europe or Japan or anywhere else) and trading them to other countries, multinational corporations could invert themselves: design products in their home markets, then communicate those designs to factories in other countries, and ship finished products back to their domestic market. And, thanks to the dramatically lower wages in Asia (supercharged by China’s opening in 1978), it was immensely profitable to do just that.

    What followed over the last several decades was the same establishment of scale and inertia that led to a dependency on US-East-1 on the Internet, only in this case the center of gravity was China. Once the cost of communication and transportation plummeted it suddenly became viable to shift industry across the globe in pursuit of lower labor costs, looser environmental laws, and governments eager to support factory build-outs. Then, over time, scale and inertia took over: if everyone else was building a factory in China, it was easier to build your factory there; if all of the factories for your components were there, it was easier to do final assembly there.

    This pattern applied to rare earths just as much as anything else. China identified rare earths as a strategic priority even as the United States made it increasingly untenable to maintain, much less expand, rare earth mining and processing here; over time nearly every part of the rare earth production chain, from separation to processing to refining to actual usage in final products became centered in China, and any attempts to build out an alternative saw their markets flooded by Chinese supply, driving down prices and dooming projects. Not that end users cared: they could just buy from China, just like everyone everywhere increasingly bought everything from China.

    One of the critiques I’ve previously leveled at classical free trade arguments is that they ignore the importance of learning curves; from A Chance to Build:

    The story to me seems straightforward: the big loser in the post World War 2 reconfiguration I described above was the American worker; yes, we have all of those service jobs, but what we have much less of are traditional manufacturing jobs. What happened to chips in the 1960s happened to manufacturing of all kinds over the ensuing decades. Countries like China started with labor cost advantages, and, over time, moved up learning curves that the U.S. dismantled; that is how you end up with this from Walter Isaacson in his Steve Jobs biography about a dinner with then-President Obama:

    When Jobs’s turn came, he stressed the need for more trained engineers and suggested that any foreign students who earned an engineering degree in the United States should be given a visa to stay in the country. Obama said that could be done only in the context of the “Dream Act,” which would allow illegal aliens who arrived as minors and finished high school to become legal residents — something that the Republicans had blocked. Jobs found this an annoying example of how politics can lead to paralysis. “The president is very smart, but he kept explaining to us reasons why things can’t get done,” he recalled. “It infuriates me.”

    Jobs went on to urge that a way be found to train more American engineers. Apple had 700,000 factory workers employed in China, he said, and that was because it needed 30,000 engineers on-site to support those workers. “You can’t find that many in America to hire,” he said. These factory engineers did not have to be PhDs or geniuses; they simply needed to have basic engineering skills for manufacturing. Tech schools, community colleges, or trade schools could train them. “If you could educate these engineers,” he said, “we could move more manufacturing plants here.” The argument made a strong impression on the president. Two or three times over the next month he told his aides, “We’ve got to find ways to train those 30,000 manufacturing engineers that Jobs told us about.”

    I think that Jobs had cause-and-effect backwards: there are not 30,000 manufacturing engineers in the U.S. because there are not 30,000 manufacturing engineering jobs to be filled. That is because the structure of the world economy — choices made starting with Bretton Woods in particular, and cemented by the removal of tariffs over time — made them nonviable. Say what you will about the viability or wisdom of Trump’s tariffs, the motivation — to undo eighty years of structural changes — is pretty straightforward!

    The other thing about Jobs’ answer is how ultimately self-serving it was. This is not to say it was wrong: Apple could not only not manufacture an iPhone in the U.S. because of cost, it also can’t do so because of capability; that capability is downstream of an ecosystem that has developed in Asia and a long learning curve that China has traveled and that the U.S. has abandoned. Ultimately, though, the benefit to Apple has been profound: the company has the best supply chain in the world, centered in China, that gives it the capability to build computers on an unimaginable scale with maximum quality for not that much money at all.

    The Apple-China story is so compelling because it is so representative of how the U.S. has become dependent on China. What is notable, however, is that this dependency points to another flaw in classic free trade formulations: while in theory free trade and globalization make supply chains more resilient because you can source from anywhere, in practice free trade has destroyed resiliency. Apple CEO Tim Cook famously said in what became known as The Tim Cook Doctrine:

    We believe that we need to own and control the primary technologies behind the products we make, and participate only in markets where we can make a significant contribution.

    The fact of the matter, however, is that Apple’s most important technology — the one architected by Cook himself — is its unmatched capability to make the most sophisticated and profitable devices at astronomical scale, and Apple ultimately does not own and control it: China does.

    So it goes for nearly everything else in the industrial supply chain, including rare earths. Rare earths are not, in fact, rare, but China’s scale and the inertia of the last forty years has led to total dependence on a country that is a geopolitical foe of the United States. And so, once again, removing or reducing the costs of transportation and communication — this time for atoms — did not increase resiliency but rather, thanks to the pursuit of lower costs enabled by scale, destroyed it.

    COVID and Information Resiliency

    There is a happier story to be told about overcoming resiliency collapse, but I should warn you up front, this might be a controversial take. It has to do with the current state of information, the earliest and most popular Internet content.

    Back in March 2020 I wrote an Article entitled Zero Trust Information that made the case that the Internet was under-appreciated as a medium for conveying information that cut against the prevailing wisdom; my go-to example was the Seattle Flu Study, which heroically traced the spread of COVID in the United States at the beginning of 2020, making the (correct) case that the virus was far more widespread in the U.S. than the CDC in particular was willing to admit.

    In truth, however, my optimism was misplaced, or at least early. What followed in the weeks and months and even years afterwards was one of the greatest failures in information discovery and propagation maybe ever. I actually wrote an Update on March 2, 2020 that included a huge amount of relevant COVID information — including the fact that it was both going to infect everyone, and that it was much less fatal than initially assumed — that only became widely accepted years later (and still isn’t accepted by a big chunk of the population).

    It’s hard not to think about how much differently we might have handled the ensuing months and years if just those two facts were widely accepted, much less other banal observations like the fact that of course natural immunity is a real thing, or that airborne viruses are all-but-inescapable indoors, but much less of an issue outdoors. Unfortunately what happened is that by 2020 information distribution was highly centralized on Facebook, Twitter, and YouTube, and all three companies went to extraordinary lengths to limit the aperture of acceptable discourse on topics that contained a great deal of unknowns; indeed, it’s possible that that March 2 Update, had it been posted on one of those platforms, would have at one point earned me a ban. In short, our resiliency in terms of information propagation was by 2020 completely destroyed, and we all suffered the consequences.

    Then Elon Musk bought Twitter.

    What is fascinating about what has happened in the years since Musk’s purchase is not that Twitter has become a fountain of truth, even if it did in some respects become considerably freer. More importantly, Musk’s purchase and ensuing political advocacy provided the impetus for a number of Twitter alternatives, including Threads, Mastodon, and BlueSky.

    Each of these networks has its own focus and mores and overall culture. What is critical about their existence, however, is not that any one of them has a monopoly on the truth: rather, given that such a monopoly is impossible, it’s heartening that there is more than one forum. To that end, should a COVID-like episode arise today, there may be an easily distinguishable and widely-held-on-the-platform X truth, and Threads truth, and Mastodon truth, and BlueSky truth; the fact that none of those truths will be completely right — and in come cases completely at odds — is not a bug but a feature: that’s actual resiliency, because it increases the likelihood that we collectively arrive at the right answer sooner than we did in the COVID era.

    The Costs of Resiliency

    What is worth noting is that the only way we arrived at this point is through a fair bit of value destruction: Musk overpaid for Twitter, and losing a monopoly on short-form text communication diminished the value further. I think, however, that the collective outcome was positive.

    Unwinding US-East-1 dependencies will also take a similar sort of pain: businesses will need to spend money to truly understand their stack, and build actual resiliency into their systems, such that one region on one cloud provider going down doesn’t screw up their business; it can be done, it just needs the budget.

    And, in the end, we can do something similar with China. There, though, the difference between atoms and bits is very profound, and exceptionally costly. Overcoming the advantages of scale and decades-long learning curves will be very painful and very expensive; the only solution to the inevitable destruction of resiliency that comes from decreased transportation and communications costs is to increase costs elsewhere, even if those costs are artificial and lead to deadweight loss.

    I am, needless to say, much more optimistic about our willingness to accept the costs of moving some bits around than I am the willingness to accept the drastically larger and longer costs of moving atoms. If we don’t, however, then we need to be clear that the true price being paid for global efficiency is national resiliency. Pursuing the former led to the destruction of the latter; there’s no way back other than destroying some value along the way.



    Get notified about new Articles


  • OpenAI’s Windows Play

    Listen to this post:

    OpenAI’s flood of announcements is getting hard to keep up with. A selection — not exhaustive! — from just the last month:

    The last two announcements just dropped yesterday, and actually bring clarity and coherence to the entire list. In short, OpenAI is making a play to be the Windows of AI.

    For nearly two decades smartphones, and in particular iOS, have been the touchstones in terms of discussing platforms. It’s important to note, however, that while Apple’s strategy of integrating hardware and software was immensely profitable, it entailed leaving the door open for a competing platform to emerge. The challenge of being a hardware company is that by virtue of needing to actually create devices you can’t serve everyone; Apple in particular didn’t have the capacity or desire to go downmarket, which created the opportunity for Android to not only establish a competing platform but to actually significantly exceed iOS in market share.

    That means that if we want a historical analogy for total platform dominance — which increasingly appears to be OpenAI’s goal — we have to go back further to the PC era and Windows.

    Platform Establishment

    Before there was Windows there was DOS; before DOS, however, there was a fast-talking deal-making entrepreneur named Bill Gates. From The Truth About Windows Versus the Mac:

    In the late 1970s and very early 1980s, a new breed of personal computers were appearing on the scene, including the Commodore, MITS Altair, Apple II, and more. Some employees were bringing them into the workplace, which major corporations found unacceptable, so IT departments asked IBM for something similar. After all, “No one ever got fired for buying IBM.”

    IBM spun up a separate team in Florida to put together something they could sell IT departments. Pressed for time, the Florida team put together a minicomputer using mostly off-the shelf components; IBM’s RISC processors and the OS they had under development were technically superior, but Intel had a CISC processor for sale immediately, and a new company called Microsoft said their OS — DOS, which they acquired from another company — could be ready in six months. For the sake of expediency, IBM decided to go with Intel and Microsoft.

    The rest, as they say, is history. The demand from corporations for IBM PCs was overwhelming, and DOS — and applications written for it — became entrenched. By the time the Mac appeared in 1984, the die had long since been cast. Ultimately, it would take Microsoft a decade to approach the Mac’s ease-of-use, but Windows’ DOS underpinnings and associated application library meant the Microsoft position was secure regardless.

    There is nothing like IBM and its dominant position in enterprise today; rather, the route to becoming a platform is to first be a massively popular product. Acquiring developers and users is not a chicken-and-egg problem: it is clear that you must get users first, which attracts developers, enhancing your platform in a virtuous cycle; to put it another way, first a product must Aggregate users and then it gets developers for free.

    ChatGPT is exactly that sort of product, and at yesterday’s DevDay 2025 keynote CEO Sam Altman and team demonstrated exactly that sort of pull; from The Verge:

    OpenAI is introducing a way to work with apps right inside ChatGPT. The idea is that, from within a conversation with the chatbot, you can essentially tag in apps to help you complete a task while ChatGPT offers context and advice. The company showed off a few different ways this can work. In a live demo, an OpenAI employee launched ChatGPT and then asked Canva to create a poster of a name for a dog-walking business; after a bit of waiting, Canva came back with a few different examples, and the presenter followed up by asking for a generated pitch deck based on the poster. The employee also asked Zillow via ChatGPT to show homes for sale in Pittsburgh, and it created an interactive Zillow map — which the employee then asked follow-up questions about.

    Apps available inside ChatGPT starting today will include Booking.com, Canva, Coursera, Expedia, Figma, Spotify, and Zillow. In the “weeks ahead,” OpenAI will add more apps, such as DoorDash, OpenTable, Target, and Uber. OpenAI recently started allowing ChatGPT users to make purchases on Etsy through the chatbot, part of its overall push to integrate it with the rest of the web.

    It’s fair to wonder if these app experiences will measure up to these company’s self-built apps or websites, just as there are questions about just how well the company’s Instant Checkout will convert; what is notable, however, is that I disagree that this represents a “push to integrate…with the rest of the web”.

    This is the opposite: this is a push to make ChatGPT the operating system of the future. Apps won’t be on your phone or in a browser; they’ll be in ChatGPT, and if they aren’t, they simply will not exist for ChatGPT users. That, by extension, means the burden of making these integrations work — and those conversions performant — will be on third party developers, not OpenAI. This is the power that comes from owning users, and OpenAI is flexing that power in a major way.

    Second Sourcing

    There is a second aspect to the IBM PC strategy, and that is the role of AMD. From a 2024 Update:

    While IBM chose Intel to provide the PC’s processor, they were wary of being reliant on a single supplier (it’s notable that IBM didn’t demand the same of the operating system, which was probably a combination of not fully appreciating operating systems as a point of integration and lock-in for 3rd-party software, which barely existed at that point, and a recognition that software is just bits and not a physical good that has to be manufactured). To that end IBM demanded that Intel license its processor to another chip firm, and AMD was the obvious choice: the firm was founded by Jerry Sanders, a Fairchild Semiconductor alum who had worked with Intel’s founders, and specialized in manufacturing licensed chips.

    The relationship between Intel and AMD ended up being incredibly fraught and largely documented by endless lawsuits (you can read a brief history in that Update); the key point to understand, however, is that (1) IBM wanted to have dual suppliers to avoid being captive to an essential component provider and (2) IBM had the power to make that happen because they had the customers who were going to provide Intel so much volume.

    The true beneficiary of IBM’s foresight, of course, was Microsoft, which controlled the operating system; IBM’s mandate is why it is appropriate that “Windows” comes first in the “Wintel” characterization of the PC era. Intel reaped tremendous profits from its position in the PC value chain, but more value accrued to Microsoft than anyone else.

    This question of who will capture the most profit from the AI value chain remains an open one. There’s no question that the early winner is Nvidia: the company has become the most valuable in the world by virtue of its combination of best-in-class GPUs, superior networking, and CUDA software layer that locks people into Nvidia’s own platform. And, as long as power is the limiting factor, Nvidia is well-placed to maintain its position.

    What Nvidia is not shy about is capturing its share of value, and that is a powerful incentive for other companies in the value chain to look for alternatives. Google is the furthest along in this regard thanks to its decade-old investment in TPUs, while Amazon is seeking to mimic their strategy with Trainium; Microsoft and Meta are both working to design and build their own chips, and Apple is upscaling Apple Silicon for use in the data center.

    Once again, however, the most obvious and most immediately available alternative to Nvidia is AMD, and I think the parallels between yesterday’s announcement of an OpenAI-AMD deal and IBM’s strong-arming of Intel are very clear; from the Wall Street Journal:

    OpenAI and chip-designer Advanced Micro Devices announced a multibillion-dollar partnership to collaborate on AI data centers that will run on AMD processors, one of the most direct challenges yet to industry leader Nvidia. Under the terms of the deal, OpenAI committed to purchasing 6 gigawatts worth of AMD’s chips, starting with the MI450 chip next year. The ChatGPT maker will buy the chips either directly or through its cloud computing partners.

    AMD chief Lisa Su said in an interview Sunday that the deal would result in tens of billions of dollars in new revenue for the chip company over the next half-decade. The two companies didn’t disclose the plan’s expected overall cost, but AMD said it costs tens of billions of dollars per gigawatt of computing capacity. OpenAI will receive warrants for up to 160 million AMD shares, roughly 10% of the chip company, at 1 cent per share, awarded in phases, if OpenAI hits certain milestones for deployment. AMD’s stock price also has to increase for the warrants to be exercised.

    If OpenAI is the software layer that matters to the ecosystem, then Nvidia’s long-term pricing power will be diminished; the company, like Intel, may still take the lion’s share of chip profits through sheer performance and low-level lock-in, but I believe the most important reason OpenAI is making this deal is to lock in its own dominant position in the stack. It is pretty notable that this announcement comes only weeks after Nvidia’s investment in OpenAI; that, though, is another affirmation that the company who has the users has the ultimate power.

    There is one other part of the stack to keep an eye on: TSMC. Both Nvidia and AMD make their chips with the Taiwanese giant, and while TSMC is famously reticent to take price, they are positioned to do so in the long run. Altman surely knows this as well, which means that I wouldn’t be surprised if there is an Intel announcement sooner rather than later; maybe there is fire behind that recent smoke about AMD talking with Intel?

    The AI Linchpin

    When I started writing Stratechery, Windows was a platform in decline, superceded by mobile and, surprisingly enough, increasingly challenged by its all-but-vanquished ancient foe, the Mac. To that end, one of my first pieces about Microsoft was about then-CEO Steve Ballmer’s misguided attempt to focus on devices instead of services. I wrote a few years later in Microsoft’s Monopoly Hangover:

    The truth is that both [IBM and Microsoft] were victims of their own monopolistic success: Windows, like the System/360 before it, was a platform that enabled Microsoft to make money in all directions. Both companies made money on the device itself and by selling many of the most important apps (and in the case of Microsoft, back-room services) that ran on it. There was no need to distinguish between a vertical strategy, in which apps and services served to differentiate the device, or a horizontal one, in which the device served to provide access to apps and services. When you are a monopoly, the answer to strategic choices can always be “Yes.”

    Microsoft at that point in time no longer had that luxury: the company needed to make a choice — the days of doing everything were over — and that choice should be services (which is exactly what Satya Nadella did).

    Ever since the emergence of ChatGPT made OpenAI The Accidental Consumer Tech Company I have been making similar arguments about OpenAI: they need to focus on the consumer opportunity and leave the enterprise API market to Microsoft. Not only would focus help the company capture the consumer opportunity, there was the opportunity cost of GPUs used for the API that couldn’t be used to deliver consumers a better experience across every tier.

    I now have much more appreciation for OpenAI’s insistence on doing it all, for two reasons. First, this is a company in pure growth mode, not in decline. Tradeoffs are in the long run inevitable, but why make them before you need to? It would have been a mistake for Microsoft to restrict Windows to only the enterprise in the 1980s, even if the company had to low-key retreat from the consumer market over the last fifteen years; there was a lot of money to make before that retreat needed to happen! OpenAI, meanwhile, is the hottest brand in AI, so why not make a play to own it all, from consumer touchpoint to API to everything in-between?

    Second, we’ve obviously crossed the line into bubble territory, which always was inevitable. The question now is whether or not this is a productive bubble: what durable infrastructure will be built by eventually bankrupt companies that we benefit from for years to come?

    GPUs are not that durable infrastructure; data centers are more long-lasting, but not worth the financial pain of a bubble burst. The real payoff would be a massive build-out in power generation, which would be a benefit for the next half century. Another potential payoff would be the renewed viability of Intel, and as I noted above, OpenAI may be uniquely positioned and motivated to make that happen.

    More broadly, this play to be the Windows of AI effectively positions OpenAI as the linchpin of the entire AI buildout. Just look at what the mere announcement of partnerships with OpenAI has done for the stocks of Oracle and AMD. OpenAI is creating the conditions such that it is the primary manifestation of the AI bubble, which ensures the company is the primary beneficiary of all of the speculative capital flooding into the space. Were the company more focused, as I have previously advised, they may not have the leverage to get enough funding to meet those more modest (but still incredible) goals; now it’s hard to see them not getting whatever money they want, at least until the bubble bursts.


    What’s amazing about this overview is that I only scratched the surface of what OpenAI announced both yesterday and over the last month — and I haven’t even mentioned Sora (although I covered that topic yesterday). What the company is seeking to achieve is incredibly audacious, but also logical, and something we’ve seen before:

    And, interestingly enough, there is an Apple to OpenAI’s Microsoft: it’s Google, with their fully integrated stack, from chips to data centers to models to end user distribution channels. Instead of taking on a menagerie of competitors, however, Google is facing an increasingly unified ecosystem, organized, whether they wish to be or not, around OpenAI. Such is the power of aggregating demand and the phenomenon that is ChatGPT.



    Get notified about new Articles


  • Sora, AI Bicycles, and Meta Disruption

    Listen to this post:

    The App Store charts tell the story, at least for the first week of AI-generated video apps:

    This doesn’t, somewhat embarrassingly, match my initial impressions: I liked the Vibes addition to the MetaAI app and was somewhat cool on Sora. I spent much of last week’s episode of Sharp Tech exploring why my initial impressions were so off base, and I think M.G. Siegler — who was sucked into Sora immediately — captures a few of them in Sora’s Slop Hits Different:

    Anyway, what’s different, and what I underestimated about Sora, is that the AI content here is not just randomly generated things. It’s content that’s either loaded with “cameos” from your connections or it’s “real” world content that’s, well, hilarious. Not all of it, of course. But a lot of it! In this regard, it’s really not too dissimilar from TikTok — and back in the day, Vine! This is a lot more like those social networks but with the main difference being that it’s a lot easier to create such content thanks to AI.

    I think that’s the real revelation here. It’s less about consumption and more about creation. I previously wrote about how I was an early investor in Vine in part because it felt like it could be analogous to Instagram. Thanks in large part to filters, that app made it easy for anyone to think they were good enough to be a photographer. It didn’t matter if they were or not, they thought they were — I was one of them — so everyone posted their photos. Vine felt like it could have been that for video thanks to its clever tap-to-record mechanism. But actually, it became a network for a lot of really talented amateurs to figure out a new format for funny videos on the internet. When Twitter acquired the company and dropped the ball, TikTok took that idea and scaled it (thanks to ByteDance paying um, Meta billions of dollars for distribution, and their own very smart algorithms).

    In a way, Sora feels like enabling everyone to be a TikTok creator.

    I feel blessed for a whole host of reasons, many of them related to the fact I’ve been able to carve out a career as a creator. Sure, I call myself an analyst, and I write about primarily big tech companies, but one thing I realized over the years is that the success of Stratechery is tied to it being a creative endeavor; there have been a lot of analysts over the years who have launched similar sites, but what was often missing was the narrative element. The best Articles on Stratechery tell a story, with a beginning, middle, and end, and the analysis is along for the ride; analysis alone doesn’t move the needle.

    That I tell stories is itself a function of the way I think: I have a larger meta story in my head about how the world works, and I’m always adding and augmenting that story; that’s why, in various interviews, I’ve noted that being wrong is often the most inspiring (albeit painful) place to be. That means my story is incomplete, and I need to deepen my understanding of the world I’m seeking to chronicle. I certainly have that opportunity right now.

    My Creativity Blindspot

    This is what I wrote in my Update about Sora:

    Indeed, it feels like each company has an entirely different target audience: YouTube is making tools for creators, Meta is building the ultimate lean back dream-like experience, and OpenAI is making an app that is, in my estimation, the easiest for normal people to use.

    In this new competition, I prefer the Meta experience, by a significant margin, and the reason why goes back to one of the oldest axioms in technology: the 90/9/1 rule.

    • 90% of users consume
    • 9% of users edit/distribute
    • 1% of users create

    If you were to categorize the target market of these three AI video entrants, you might say that YouTube is focused on the 1% of creators; OpenAI is focused on the 9% of editors/distributors; Meta is focused on the 90% of users who consume. Speaking as someone who is, at least for now, more interested in consuming AI content than in distributing or creating it, I find Meta’s Vibes app genuinely compelling; the Sora app feels like a parlor trick, if I’m being honest, and I tired of my feed pretty quickly. I’m going to refrain on passing judgment on YouTube, given that my current primary YouTube use case is watching vocal coaches break down songs from KPop Demon Hunters.

    I honestly have no idea if my evaluation of these apps is broadly applicable; as I’ve noted repeatedly, I’m hesitant to make any pronouncements about what resonates with society broadly given that I am the weirdo in the room. Still, I do think it’s striking how this target market evaluation tracks with the companies themselves: YouTube has always prioritized creators, while OpenAI’s business model is predicated on people actively using AI; it’s Meta that has stayed focused on the silent majority that simply consumes, and as a silent consumer, I still like Vibes!

    As I noted at the beginning, the verdict is in, and my evaluation of these apps is not broadly applicable. Way more people like Sora than Vibes, and OpenAI has another viral hit. What I hear from people who love the app, however, is very much in line with what Siegler wrote: yes, they are browsing the feed, but the real lure is losing surprisingly large amounts of time making content — Sora lets them be a content creator.

    This was a blind spot for me because I don’t have that itch! I’m creating content constantly — three Articles/Updates, an Interview, and three podcast episodes a week is enough for me, thank you very much. When I am vegging out on my phone, I want to passively consume, and I personally found the Vibes mix of fantastical environments and beautiful visages calming and inspiring; almost everyone else feels different:

    I had to laugh at this because I’ve spent way too much time watching Apple’s Aerial Video screensavers; apparently my tastes are consistent! Beyond that, however, is a second blind spot: how much of the 90/9/1 rule is a law of the universe, versus a manifestation of barriers when it comes to creation? At the risk of sounding like a snob, have I become the sort of 1%-er who is totally out of touch?

    The AI Bicycle

    Back in 2022, when AI image generation was just starting to get good, I wrote about The AI Unbundling and the idea propagation chain:

    The five parts of the idea propagation value chain: creation, substantiation, duplication, distribution, consumption

    The evolution of human communication has been about removing whatever bottleneck is in this value chain. Before humans could write, information could only be conveyed orally; that meant that the creation, vocalization, delivery, and consumption of an idea were all one-and-the-same. Writing, though, unbundled consumption, increasing the number of people who could consume an idea.

    Writing unbundled consumption from the rest of the value chain

    Now the new bottleneck was duplication: to reach more people whatever was written had to be painstakingly duplicated by hand, which dramatically limited what ideas were recorded and preserved. The printing press removed this bottleneck, dramatically increasing the number of ideas that could be economically distributed:

    The printing press unbundled duplication and distribution from creation

    The new bottleneck was distribution, which is to say this was the new place to make money; thus the aforementioned profitability of newspapers. That bottleneck, though, was removed by the Internet, which made distribution free and available to anyone.

    The Internet unbundled distribution from duplication

    What remains is one final bundle: the creation and substantiation of an idea. To use myself as an example, I have plenty of ideas, and thanks to the Internet, the ability to distribute them around the globe; however, I still need to write them down, just as an artist needs to create an image, or a musician needs to write a song. What is becoming increasingly clear, though, is that this too is a bottleneck that is on the verge of being removed.

    This is what was unlocked by Sora: all sorts of people without the time or inclination or skills or equipment to make videos could suddenly do just that — and they absolutely loved it. And why wouldn’t they? To be creative is to be truly human — to actually think of something yourself, instead of simply passively consuming — and AI makes creativity as accessible as a simple prompt.

    I think this is pretty remarkable, so much so that I’ve done a complete 180 on Sora: this new app from OpenAI may be the single most exciting manifestation of AI yet, and the most encouraging in terms of AI’s impact on humans. Everyone — including lots of people in my Sora feed — are leaning into the concept of AI slop, which I get: we are looking at a world of infinite machine-generated content, and a lot of it is going to be terrible.

    At the same time, how incredible is it to give everyone with an iPhone a creative outlet? It reminds me of one of my favorite Steve Jobs moments, just before he died, at the introduction of the iPad 2; I wrote about it in 2024’s The Great Flattening:

    My favorite moment in that keynote — one of my favorite Steve Jobs’ keynote moments ever, in fact — was the introduction of GarageBand. You can watch the entire introduction and demo, but the part that stands out in my memory is Jobs — clearly sick, in retrospect — moved by what the company had just produced:

    I’m blown away with this stuff. Playing your own instruments, or using the smart instruments, anyone can make music now, in something that’s this thick and weighs 1.3 pounds. It’s unbelievable. GarageBand for iPad. Great set of features — again, this is no toy. This is something you can really use for real work. This is something that, I cannot tell you, how many hours teenagers are going to spend making music with this, and teaching themselves about music with this.

    Jobs wasn’t wrong: global hits have originated on GarageBand, and undoubtedly many more hours of (mostly terrible, if my personal experience is any indication) amateur experimentation. Why I think this demo was so personally meaningful for Jobs, though, is that not only was GarageBand about music, one of his deepest passions, but it was also a manifestation of his life’s work: creating a bicycle for the mind.

    I remember reading an Article when I was about 12 years old, I think it might have been in Scientific American, where they measured the efficiency of locomotion for all these species on planet earth. How many kilocalories did they expend to get from point A to point B, and the condor won: it came in at the top of the list, surpassed everything else. And humans came in about a third of the way down the list, which was not such a great showing for the crown of creation.

    But somebody there had the imagination to test the efficiency of a human riding a bicycle. Human riding a bicycle blew away the condor, all the way off the top of the list. And it made a really big impression on me that we humans are tool builders, and that we can fashion tools that amplify these inherent abilities that we have to spectacular magnitudes, and so for me a computer has always been a bicycle of the mind, something that takes us far beyond our inherent abilities.

    I think we’re just at the early stages of this tool, very early stages, and we’ve come only a very short distance, and it’s still in its formation, but already we’ve seen enormous changes, but I think that’s nothing compared to what’s coming in the next 100 years.

    In Jobs’ view of the world, teenagers the world over are potential musicians, who might not be able to afford a piano or guitar or trumpet; if, though, they can get an iPad — now even thinner and lighter! — they can have access to everything they need. In this view “There’s an app for that” is profoundly empowering.

    Well, now there’s an AI for that, and it’s accessible to everyone. And yes, I get the objections. I slave over these posts, thinking carefully about the structure and every word choice; it seems cheap to ask an LLM to generate the same. I’m certain that artists feel the same about AI images, or musicians about AI music, or YouTube and TikTok creators about Sora videos; what about the craft?

    That, though, is an easy concern to have when you already have a creative outlet; it’s also easy to make the case that more content means more compelling content to consume, even if the percentage of what is great is very small.

    What I didn’t fully appreciate, however, is what falls in the middle: the fact that so many more people get to be creators, and what a blessing that is. How many people have had ideas in their head, yet were incapable of substantiating them, and now can? I myself benefited greatly from the last unbundling — the ability for anyone to distribute content; why should I begrudge the latest unbundling, and the many more people who will benefit from AI substantiation of their creative impulses? Bicycles for all!

    Instagram’s Social Umbrella

    Siegler in his post discussed how he once thought Vine could be like Instagram, which made it easy to feel like a good photographer with its filters, but that was only step one; Chris Dixon described Instagram’s evolution as Come for the Tool, Stay for the Network:

    A popular strategy for bootstrapping networks is what I like to call “come for the tool, stay for the network.” The idea is to initially attract users with a single-player tool and then, over time, get them to participate in a network. The tool helps get to initial critical mass. The network creates the long term value for users, and defensibility for the company.

    Here are two historical examples: 1) Delicious. The single-player tool was a cloud service for your bookmarks. The multiplayer network was a tagging system for discovering and sharing links. 2) Instagram. Instagram’s initial hook was the innovative photo filters. At the time some other apps like Hipstamatic had filters but you had to pay for them. Instagram also made it easy to share your photos on other networks like Facebook and Twitter. But you could also share on Instagram’s network, which of course became the preferred way to use Instagram over time.

    Dixon wrote that post in 2015, and Instagram has since gone much further than that, as I documented in 2021’s Instagram’s Evolution:

    • There was the tool to network evolution that Dixon talked about.
    • The second evolution was the addition of video.
    • The third evolution was the introduction of the algorithmic feed.
    • The fourth evolution was Stories, driven by competition with Snapchat.
    • The fifth evolution was what I was writing about in that Article: the commitment to short-form video, driven by competition with TikTok.

    That last evolution is fully baked in at this point; late last month Instagram announced that it was changing Instagram’s navigation to focus on private messaging and Reels; I didn’t explicitly cover the 2013 addition of Instagram Direct, but it certainly is the case that messaging is where social networking happens today. What is public is pure entertainment, where the content you see is pulled from across the network and tailored for you specifically.

    I think this evolution was both necessary and inevitable; I first wrote that Facebook needed to move in this direction in 2015’s Facebook and the Feed:

    Consider Facebook’s smartest acquisition, Instagram. The photo-sharing service is valuable because it is a network, but it initially got traction because of filters. Sometimes what gets you started is only a lever to what makes you valuable. What, though, lies beyond the network? That was Facebook’s starting point, and I think the answer to what lies beyond is clear: the entire online experience of over a billion people. Will Facebook seek to protect its network — and Zuckerberg’s vision — or make a play to be the television of mobile?

    It wasn’t until TikTok peeled off a huge amount of attention that Facebook finally realized that viewing itself as a social network was actually limiting its potential. If the goal was to monopolize user attention — the only scarce resource on the Internet — then artificially limiting what people saw to their social network was to fight with one hand tied behind your back; TikTok was taking share not just because of its format, but also because it wasn’t really a social network at all.

    This is all interesting context for how OpenAI characterized Sora in their introductory post: it’s a social app.

    Today, we’re launching a new social iOS app just called “Sora,” powered by Sora 2. Inside the app, you can create, remix each other’s generations, discover new videos in a customizable Sora feed, and bring yourself or your friends in via cameos. With cameos, you can drop yourself straight into any Sora scene with remarkable fidelity after a short one-time video-and-audio recording in the app to verify your identity and capture your likeness…

    This app is made to be used with your friends. Overwhelming feedback from testers is that cameos are what make this feel different and fun to use — you have to try it to really get it, but it is a new and unique way to communicate with people. We’re rolling this out as an invite-based app to make sure you come in with your friends. At a time when all major platforms are moving away from the social graph, we think cameos will reinforce community.

    First, just because Meta needed to move beyond the social network doesn’t mean social networking isn’t still valuable, or appealing. As an analogy, consider the concept of a pricing umbrella: when something becomes more expensive, it opens up the market for a lower-priced competitor. In this case Instagram’s evolution has created a social umbrella: sure, Instagram content may be “better” by virtue of being pulled from anywhere, but that means there is now a space for a content app that is organized around friends.

    Second, remember the creativity point above: one of the challenges of restricting Instagram content to just what your social network posted is that your social network may not post very many interesting things. That gap was initially filled by following influencers, but now Instagram simply goes out and finds what you are interested in without having to do anything. In Sora, however, your network is uniquely empowered to be creative, increasing the amount of interesting content in a network-mediated context (and, of course, Sora is also pulling from elsewhere as well to populate your feed).

    What you’re seeing, if you squint, is disruption: Instagram has gone “up-market” in terms of content, leaving space for a new entrant; that new entrant, meanwhile, is not simply cheaper/smaller. Rather, it’s enabled by a new technological paradigm that lets it compete orthogonally with the incumbent. Granted, that new paradigm is very expensive, particularly compared to the content that Instagram gets for free, but the extent it restores value to your social network is notable.

    Meta Concerns

    I am on the record as being very bullish about the impact of AI on Meta’s business:

    • It’s good for their ad business in the short, medium, and long-term (and YouTube’s as well).
    • More content benefits the company with the most popular distribution channels.
    • AI will be the key to unlocking both AR and VR.

    The key to everything, however, is maintaining the hold Meta has on user attention, and the release of both Vibes and Sora has me seriously questioning point number two.

    What I appreciate about both of these apps is the fact they are explicitly AI-content; I said in my Update about Vibes:

    One of the reasons why AI slop is so annoying is — paradoxically — the fact that a lot of it has gotten quite good. That means that when consuming content you have to continually be ascertaining if what you see is real or AI-generated; to put it in the terms of the Article I just quoted, you might want to lean back, but if you don’t want to be taken in or make a fool of yourself then you have to constantly be leaning forward to figure out what is or isn’t AI.

    What this means for Vibes is the fact it is unapologetically and explicitly all AI is quite profound: it’s a true lean-back experience, where the fact none of it is real is a point of interest and — if Holz is right — inspiration and imagination. I find it quite relaxing to consume, in a way I don’t find almost any other feed on my phone.

    The reason this is problematic for Meta (and YouTube) is that I’m not sure the company can counter Sora — or any other AI-generated content app that appears — in the same way they countered Snapchat and TikTok. Both challengers introduced new formats — Stories in the case of Instagram, and short-form video in the case of TikTok — but the content was still produced by humans; that made it much more palatable to stuff those formats into Instagram.

    AI might be different: Meta certainly has data on this question, but I could imagine a scenario where users are actually annoyed and turned off by mixing AI-generated content with human content — and because Instagram isn’t really a social network anymore, the fact that that content might be made by or include your friends might not be enough. Implicit in this observation is the fact that I don’t think that human content is going anywhere; there just might be a smaller percentage of time devoted to it, and that’s a problem for a company predicated on marshaling attention.

    The second issue for Meta is that their AI capabilities simply don’t match OpenAI, or Google’s for that matter. It’s clear that Meta knows this is the case — look no further than this summer’s hiring spree and total overhaul of their AI approach — but creating something like Sora is a lot more difficult than copying Stories or short-form video. I imagine this shortcoming will be rectified, but Sora is in the market now.

    I also think that it is fair to raise some questions about point three. I have been a vocal proponent of AI being the key to the Metaverse, but my tastes in content may not be very broadly applicable! I loved Vibes because to me it felt like virtual reality, but if it was virtual reality, and no one liked it, maybe the concept actually isn’t that appealing? Time will tell, but I do keep coming back to the social aspects of Sora: people like the real world, and they like people they know, and virtual reality in particular just might not be that broadly popular.

    And, while I’m here, I continue to think that Meta’s recent financial success is not entirely organic:

    It turns out I was right last quarter that Meta had a lot of room to increase Reels monetization, but not just because they could target ads better (that was a part of it, as I noted above): rather, it turns out that short-form video is so addictive that Meta can simply drive more engagement — and thus more ad inventory — by pushing more of it. That’s impression driver number one — and the most important one. The second one is even more explicit: Meta simply started showing more ads to people (i.e. “ad load optimization”).

    All of this ties back to where I started, about how Meta learned that you have to give investors short term results to get permission for long term investments. I don’t think it’s a coincidence that, in the same quarter where Meta decided to very publicly up its investment in the speculative “Superintelligence”, users got pushed more Reels and Facebook users in particular got shown more ads. The positive spin on this is that Meta has dials to turn; by the same token, investors who have flipped from intrinsically doubting Meta to intrinsically trusting them should realize that it was the pre-2022 Meta, the one that regularly voiced the importance of not pushing too many ads in order to preserve the user experience, that actually deserved the benefit of the doubt for growth that was purely organic. This last quarter is, to my mind, a bit more pre-determined.

    CEO Mark Zuckerberg framed the company’s new Personal Superintelligence like this:

    As profound as the abundance produced by AI may one day be, an even more meaningful impact on our lives will likely come from everyone having a personal superintelligence that helps you achieve your goals, create what you want to see in the world, experience any adventure, be a better friend to those you care about, and grow to become the person you aspire to be.

    Meta’s vision is to bring personal superintelligence to everyone. We believe in putting this power in people’s hands to direct it towards what they value in their own lives.

    This is distinct from others in the industry who believe superintelligence should be directed centrally towards automating all valuable work, and then humanity will live on a dole of its output. At Meta, we believe that people pursuing their individual aspirations is how we have always made progress expanding prosperity, science, health, and culture. This will be increasingly important in the future as well.

    I agree with the sentiment, but it’s worth being honest about today’s reality: Meta’s financial fortunes, at least for now, are in fact tied up in a centralized content engine that gives users “a dole of its output”; it’s nice from an investor perspective that Meta can turn the dials and get people to spend that much more time in Instagram. I for one can’t say that I feel particularly great when I’m done watching Reels for longer than I planned, and it’s certainly not a creative endeavor on my part — that’s for the content creators.

    OpenAI, meanwhile, with both ChatGPT and Sora, is in fact placing easily accessible tools in people’s hands today, first with text and now with video. And, as I noted above, I actually find it exciting precisely because of the possibility that many more people are on the verge of discovering a creativity streak they didn’t even know they had, now that AI is available to substantiate it. So much Meta optimism is, paradoxically, pessimistic about the human condition; it may be the case that, to the extent that AI makes humans better, is the extent that Meta faces disruption.



    Get notified about new Articles


  • The YouTube Tip of the Google Spear

    Listen to this post:

    Action is happening up-and-down the LLM stack: Nvidia is making deals with Intel, OpenAI is making deals with Oracle, and Nvidia and OpenAI are making deals with each other. Nine years after Nvidia CEO Jensen Huang hand-delivered the first Nvidia DGX-1 AI computer to OpenAI, the chip giant is investing up to $100 billion in the AI lab, which OpenAI will, of course, spend on Nvidia AI systems.

    This ouroboros of a deal certainly does feel a bit frothy, but there is a certain logic to it: Nvidia is uniquely dominant in AI thanks to the company’s multi-year investment in not just superior chips but also an entire ecosystem from networking to software, and has the cash flow and stock price befitting its position in the AI value chain. Doing a deal like this at this point in time not only secures the company’s largest customer — and rumored ASIC maker — but also gives Nvidia equity upside beyond the number of chips it can manufacture. More broadly, lots of public investors would like the chance to invest in OpenAI; I don’t think Nvidia’s public market investors are bothered to have now acquired that stake indirectly.

    The interconnectedness of these investments reflects the interconnectedness of the OpenAI and Nvidia stories in particular: Huang may have delivered OpenAI their first AI computer, but it was OpenAI that delivered Nvidia the catalyst for becoming the most valuable company in the world, with the November 2022 launch of ChatGPT. Ever since, the assumption of many in tech has been that the consumer market in particular has been OpenAI’s to lose, or perhaps more accurately, monetize; no company has ever grown faster in terms of users and revenue, and that’s before they had an advertising model!

    And beyond the numbers, have you used ChatGPT? It’s so useful. You can look up information, or format text, and best of all you can code! Of course there are other models like Anthropic’s Claude, which has excelled at coding in particular, but surely the sheer usefulness makes ultimate success inevitable!

    A Brief History of Social Media

    If a lot of those takes sound familiar, it’s because I’ve made some version of most of them; I also, perhaps relatedly, took to Twitter like a fish to water. Just imagine, an app that was the nearly perfect mixture of content I was interested in and people I wanted to hear from, and interact with. Best of all it was text: the efficiency of information acquisition was unmatched, and it was just as easy to say my piece.

    It took me much longer to warm up to Facebook, and, frankly, I never was much of a user; I’ve never been one to image dump episodes of my life, nor have I had much inclination to wade through others’. I wasn’t interested in party photos; I lusted after ideas and arguments, and Twitter — a view shared by much of both tech and media — was much more up my alley.

    Despite that personal predilection, however, and perhaps because of my background in small town Wisconsin and subsequently living abroad, I retained a strong sense of the importance of Facebook. Sure, the people who I was most interested in hearing from and interacting with may have been the types to leave their friends and family for the big city, but for most people, friends and family were the entire point of life generally, and by extension, social media specifically.

    To that end, I was convinced from the beginning that Facebook was going to be a huge deal, and argued so multiple times on Stratechery; social media was ultimately a matter of network effects and scale, and Facebook was clearly on the path to domination, even as much of the Twitterati were convinced the company was the next MySpace. I was similarly bullish about Instagram: no, I wasn’t one to post a lot of personal pictures, but while I personally loved text, most people liked photos.

    What people really liked most of all, however — and not even Facebook saw this coming — was video. TikTok grew into a behemoth with the insight that social media was only ever a stepping stone to personal entertainment, of which video was the pinnacle. There were no network effects of the sort that everyone — including regulators — assumed would lead to eternal Facebook dominance; rather, TikTok realized that Paul Krugman’s infamous dismissal of the Internet actually was somewhat right: most people actually don’t have anything to say that is particularly compelling, which means that limiting the content you see to your social network dramatically decreases the possibility you’ll be entertained every time you open your social networking app. TikTok dispensed with this artificial limitation, simply showing you compelling videos period, no matter where they came from.

    The Giant in Plain Sight

    Of course TikTok wasn’t the first company to figure this out: YouTube was the first video platform, and from the beginning focused on building an algorithm that focused more on giving you videos you were interested in than in showing you what you claimed to want to see.

    YouTube, however, was and probably always has been my biggest blind spot: I’m just not a big video watcher in general, and YouTube seemed like more work than short-form video, which married the most compelling medium with the most addictive delivery method — the feed. Sure, YouTube was a great acquisition for Google — certainly in line with the charge to “organize the world’s information and make it universally accessible and useful” — but I — and Google’s moneymaker, Search — was much more interested in text, and pictures if I must.

    The truth, however, is that YouTube has long been the giant hiding in plain sight: the service is the number one streaming service in the living room — bigger than Netflix — and that’s the company’s 3rd screen after mobile and the PC, where it has no peer. More than that, YouTube is not just the center of culture, but the nurturer of it: the company just announced that it has paid out more than $100 billion to creators over the last four years; given that many creators earn more from brand deals than they do from YouTube ads, that actually understates the size of the YouTube economy. Yes, TikTok is a big deal, but TikTok stars hope to make it on YouTube, where they can actually make a living.

    And yet, YouTube sometimes seems like an afterthought, at least to people like me and others immersed in the text-based Internet. Last week I was in New York for YouTube’s annual “Made on YouTube” event, but the night before I couldn’t remember the name; I turned to Google, natch, and couldn’t figure it out. The reason is that talk about YouTube mostly happens on YouTube; I, and Google itself, still live in a text-based world.

    That is the world that was rocked by ChatGPT, especially Google. The company’s February 2023 introduction of Bard in Paris remains one of the most surreal keynotes I’ve ever watched: most of the content was rehashed, the presenters talked as if they were seeing their slides for the first time, and one of the demos of a phone-based feature neglected to remember to have a phone on hand. This was a company facing a frontal assault on their most obvious and profitable area of dominance — text-based information retrieval — and they were completely flat-footed.

    Google has, in the intervening years, made tremendous strides to come back, including dumping the Bard name in favor of Gemini, itself based on vastly improved underlying models. I’m also impressed by how the company has incorporated AI into search; not only are AI Overviews generally useful, they’re also incredibly fast, and as a bonus have the links I sometimes prefer already at hand. Ironically, however, you could make the case that the biggest impact LLMs have had on Search is giving a federal judge an excuse to let Google continue paying its biggest would-be competitors (like Apple) to simply offer their customers Google instead. The biggest reason to be skeptical of the company’s fortunes in AI is that they had the most to lose; the company is doing an excellent job of minimizing the losses.

    What I would submit, however, is that Google’s most important and most compelling AI announcements actually don’t have anything to do with Search, at least not yet. These announcements start, as you might expect, with Google’s Deep Mind Research Lab; where they hit the real world, however, is on YouTube — and that, like the user-generated streaming service, is a really big deal.

    The DeepMind-to-YouTube Pipeline

    A perfect example of the DeepMind-to-YouTube pipeline was last week’s announcement of Veo 3-based features for making YouTube Shorts. From the company’s blog post:

    We’ve partnered with Google DeepMind to bring a custom version of their most powerful video generation model, Veo 3, to YouTube. Veo 3 Fast is designed to work seamlessly in YouTube Shorts for millions of creators and users, for free. It generates outputs with lower latency at 480p so you can easily create video clips – and for the first time, with sound – from any idea, all from your phone.

    This initial launch will allow you to not only generate videos, but also use one video to animate another (or a photo), stylize your video with a single touch, and add objects. You can also create an entire video — complete with voiceover — from a collection of clips, or convert speech to song. All of these features are a bit silly, but, well, that’s often where genius — or at least virality — comes from.

    Critics, of course, will label this an AI slop machine, and they’ll be right! The vast majority of content created by these tools will be boring and unwatched. That, however, is already the case with YouTube: the service sees 500 hours of content uploaded every minute, and most of that content isn’t interesting to anyone; the magic of YouTube, however, is the algorithm that finds out what is actually compelling and spreads it to an audience that wants exactly that.

    To put it another way, for YouTube AI slop is a strategy credit: given that the service has already mastered organizing overwhelming amounts of content and only surfacing what is good, it, more than anyone else, can handle exponentially more content which, through the sheer force of numbers, will result in an absolute increase of content that is actually compelling.

    That’s not the only strategy credit YouTube has; while the cost of producing AI-generated video will likely be lower than the cost of producing human-generated video, at least in the long run, the latter’s costs are not borne by TikTok or Meta (Facebook and Instagram are basically video platforms at this point). Rather, the brilliance of the user-generated content model is that creators post their content for free! This, however, means that AI-generated video is actually more expensive, at least if it’s made on TikTok or Meta’s servers. YouTube, however, pays its creators, which means that for the service AI-generated video actually has the potential to lower costs in the long run, increasing the incentive to leverage DeepMind’s industry-leading models.

    In short, while everyone immediately saw how AI could be disruptive to Search, AI is very much a sustaining innovation for YouTube: it increases the amount of compelling content in absolute terms, and it does so with better margins, at least in the long run.

    Here’s the million billion trillion dollar question: what is going to matter more in the long run, text or video? Sure, Google would like to dominate everything, but if it had to choose, is it better to dominate video or dominate text? The history of social networking that I documented above suggests that video is, in the long run, much more compelling to many more people.

    To put it another way, the things that people in tech and media are interested in has not historically been aligned with what actually makes for the largest service or makes the most money: people like me, or those reading me, care about text and ideas; the services that matter specialize in videos and entertainment, and to the extent that AI matters for the latter YouTube is primed to be the biggest winner, even as the same people who couldn’t understand why Twitter didn’t measure up to Facebook go ga-ga over text generation and coding capabilities.

    AI Monetization

    The potential impact of AI on YouTube’s fortunes isn’t just about AI-created videos; rather, the most important announcement of last week’s event was the first indicator that AI can massively increase the monetization potential of every video on the streaming service. You might have missed the announcement, because YouTube underplayed it; from their event blog post:

    We’re adding updates to brand deals and Shopping to make brand collaborations easier than ever. We’re accelerating these deals through a new initiative and new product features to make sure those partnerships succeed – like the ability to add a link to a brand’s site in Shorts. And YouTube Shopping is expanding to more markets and merchants and getting help from AI to make tagging easier.

    It’s just half a sentence — “getting help from AI to make tagging easier” — but the implications of those eight words are profound; here’s how YouTube explained the feature:

    We know tagging products can be time-consuming, so to make the experience better for creators, we’re leaning on an AI-powered system to identify the optimal moment a product is mentioned and automatically display the product tag at that time, capturing viewer interest when it’s highest. We’ll also begin testing the ability to automatically identify and tag all eligible products mentioned in your video later this year.

    The creator who demonstrated the feature — that right there is a great example of how YouTube is a different world than the one I and other people in the media inhabit — was very enthusiastic about the reduction in hassle and time-savings that would come from using AI to do a menial task like tagging sponsored products; that sounds like AI at its best, freeing up creative people to do what they do best.

    There’s no reason, however, why auto-tagging can’t become something much greater; in fact, I already explained the implications of this exact technology in explaining why AI made me bullish on Meta:

    This leads to a third medium-term AI-derived benefit that Meta will enjoy: at some point ads will be indistinguishable from content. You can already see the outlines of that given I’ve discussed both generative ads and generative content; they’re the same thing! That image that is personalized to you just might happen to include a sweater or a belt that Meta knows you probably want; simply click-to-buy.

    It’s not just generative content, though: AI can figure out what is in other content, including authentic photos and videos. Suddenly every item in that influencer photo can be labeled and linked — provided the supplier bought into the black box, of course — making not just every piece of generative AI a potential ad, but every piece of content period.

    The market implications of this are profound. One of the oddities of analyzing digital ad platforms is that some of the most important indicators are counterintuitive; I wrote this spring:

    The most optimistic time for Meta’s advertising business is, counter-intuitively, when the price-per-ad is dropping, because that means that impressions are increasing. This means that Meta is creating new long-term revenue opportunities, even as its ads become cost competitive with more of its competitors; it’s also notable that this is the point when previous investor freak-outs have happened.

    When I wrote that I was, as I noted in the introduction, feeling more cautious about Meta’s business, given that Reels is built out and the inventory opportunities of Meta AI were not immediately obvious. I realize now, though, that I was distracted by Meta AI: the real impact of AI is to make everything inventory, which is to say that the price-per-ad on Meta will approach $0 for basically forever. Would-be competitors are finding it difficult enough to compete with Meta’s userbase and resources in a probabilisitic world; to do so with basically zero price umbrella seems all-but-impossible.

    This analysis was spot-on; I just pointed it at the wrong company. This opportunity to leverage AI to make basically every pixel monetizable absolutely exists for Meta; Meta, however, has to actually develop the models and infrastructure to do it at scale. Google is already there; it was the company universally decried for being slow-moving that announced the first version of this feature last week.

    I can’t overstate what a massive opportunity this is: every item in every YouTube video is well on its way to being a monetizable surface. Yes, that may sound dystopian when I put it so baldly, but if you think about it you can see the benefits; I’ve been watching a lot of home improvement videos lately, and it sure would be useful to be able to not just identify but helpfully have a link to buy a lot of the equipment I see, much of which is basically in the background because it’s not the point of the video. It won’t be long until YouTube has that inventory, which it could surface with an affiliate fee link, or make biddable for companies who want to reach primed customers.

    More generally, you can actually envision Google pulling this off: the company may have gotten off to a horrible start in the chatbot era, but the company has pulled itself together and is increasingly bringing its model and infrastructure leadership to bear, even as Meta has had to completely overhaul their AI approach after hitting a wall. I’m sure CEO Mark Zuckerberg will figure it out, but Google — surprise! — is the company actually shipping.

    A Bull’s Journey

    Or, rather, YouTube is. Close readers of Stratechery have been observing — and probably, deservedly, smirking — at this most unexpected evolution:

    That quote is from Paradigm Shifts and the Winner’s Curse, an Article that was mostly about my concerns about Apple and Amazon, and reads:

    And, by the same token, I’m much more appreciative of Google’s amorphous nature and seeming lack of strategy. That makes them hard to analyze — again, I’ve been honest for years about the challenges I find in understanding Mountain View — but the company successfully navigated one paradigm shift, and is doing much better than I originally expected with this one. Larry Page and Sergey Brin famously weren’t particularly interested in business or in running a company; they just wanted to do cool things with computers in a college-like environment like they had at Stanford. That the company, nearly thirty years later, is still doing cool things with computers in a college-like environment may be maddening to analysts like me who want clarity and efficiency; it also may be the key to not just surviving but winning across multiple paradigms.

    Appreciating the benefits of Google being an amorphous blob where no one knows what is going on, least of all leadership, is a big part of my evolution; this Article is the second part: that blob ultimately needs a way to manifest the technology it manages to come up with, and if you were to distill my worries about Google in the age of AI it would be to wonder how the company could become an answer machine — which Page and Brin always wanted — when it risked losing the massive economic benefits that came from empowering users to choose the winners of auctions Google conducted for advertisers.

    That, however, is ultimately the text-based world, and there’s a case to be made that, in the long run, it simply won’t matter as much as the world of video. Again, the company is doing better with Search than I expected, and I’ve always been bullish about the impact of AI on the company’s cloud business; the piece I’ve missed, however, is that Google already has the tip of the spear for its AI excellence to actually go supernova: YouTube, the hidden giant in plain sight, a business that is simultaneously unfathomably large, and also just getting started.



    Get notified about new Articles


  • iPhones 17 and the Sugar Water Trap

    Listen to this post:

    I think the new iPhones are pretty great.

    The base iPhone 17 finally gets some key features from the Pro line, including the 120Hz Promotion display (the lack of which stopped me from buying the most beautiful iPhone ever). The iPhone Air, meanwhile, is a marvel of engineering: transforming the necessary but regretful camera bump into an entire module that houses all of the phone’s compute is Apple at its best, and reminiscent of how the company elevated the necessity of a front camera assembly into the digital island, a genuinely useful user interface component.

    The existence of the iPhone Air, meanwhile, seems to have given the company permission to fully lean into the “Pro” part of the iPhone Pro. I think the return to aluminum is a welcome one (and if the unibody construction is as transformative as it was for MacBooks, the feel of the phone should be a big step up), the “vapor chamber” should alleviate one of the biggest problems with previous Pro models and provide a meaningful boost in performance, and despite focusing on cameras every year for years, the latest module seems like a big step up (and the square sensor in the selfie camera is a brilliant if overdue idea). Oh, and the Air’s price point — $999, the former starting price for the Pro — finally gave Apple the opening to increase the Pro’s price by $100.

    And, I must add, it’s nice to have a retort to everyone complaining about size and weight: if that is what is important to you, get an Air! I’ll take my (truly) all-day battery life and giant screen, thank you very much, and did I mention that the flagship color is Stratechery orange?

    What was weird to me yesterday, however, is that my enthusiasm over Apple’s announcement didn’t seem to be broadly shared. There was lots of moaning and groaning about weight and size (get an iPhone Air!), gripes about the lack of changes year-over-year, and general boredom with the pre-recorded presentation (OK, that’s a fair one, but at least Apple ventured to some other cities instead of endlessly filming in and around San Francisco). This post on Threads captured the sentiment:

    This is honestly very confusing to me: the content of the post is totally contradicted by the image! Just look at the features listed:

    • There is a completely new body material and design
    • There is a new faster chip, with GPUs actually designed for AI workloads (a reminder that Apple’s neural engine was designed for much more basic machine learning algorithms, not LLMs)
    • There is a 50% increase in RAM
    • The front camera sensor has 2x the pixels, and is square
    • The telephoto lense has 4x the pixels, allowing for 8x hardware zoom
    • There is a much larger battery, thanks to the Pro borrowing the Air’s trick of bundling all of the electronics in a larger yet more aesthetically pleasing plateau
    • There is much better cooling, allowing for better sustained performance
    • There is faster charging

    This is a lot more than a 10% difference over last year’s phone! Basically every aspect of the iPhone Pro got better, and did I mention the Stratechery orange?

    I could stop there, playing the part of the analytics nerd, smugly asserting my list of numbers and features to tell everyone that they’re wrong, and, when it comes to the core question of the year-over-year improvement in iPhone hardware, I would be right! I think, however, that the widespread insistence that this was a blah update — even when the reality is otherwise — exposes another kind of truth: people are calling this update boring not because the iPhones 17 aren’t great, but because Apple no longer captures the imagination.

    Apple in the Background

    One of the advantages of living abroad is how you gain a new perspective on your home country; one of the surprises of moving back is running head-on into accumulated gradual changes that most people may not have noticed as they happened, but that you experience all at once.

    To that end, I have, for the last several years, noted how, from a Stratechery perspective, iPhone launches just aren’t nearly as big of a deal as they were when I first started. Back then I would spend weeks before the event predicting what Apple would announce, and would spend weeks afterwards breaking down the implications; now I usually dash off an Update that, in recent years, has been dominated by discussions about price and elasticity and Apple’s transition to being a services company.

    What was shocking to me, however, was actually watching the event in real time: my group chats and X feed acknowledged that the event was happening, but I had the distinct impression that almost no one was paying much attention, which was not at all the case a decade ago. And, particularly when it comes to tech discussion, you can understand why: by far the biggest thing in tech — and on Stratechery — is AI, and Apple simply isn’t a meaningful player.

    Indeed, the most important news they have made has been their announcement that they were significantly delaying major features that they promised (and advertised!) as a part of Apple Intelligence, followed by a string of news and rumors about reorganizations and talent losses, and questions about whether or not they should partner with or acquire AI companies to do what Apple seems incapable of doing themselves. Until those questions are rectified, why should anyone who cares about AI — which is to say basically everyone else in the industry — care about square camera sensors or vapor chambers?

    Apple’s Enviable Position

    I can, if I put my business analyst hat on, make the case that Apple is doing better than ever, and not just in terms of making money. One underdiscussed takeaway from this year’s announcements is that the company, which originally had the iPhone on a two-year refresh cycle in terms of industrial design, before slipping to a three-year cycle (X/XS/11 and 12/13/14) over the last decade, is back to two years: the iPhones 15 and 16 were the same, but the iPhone 17 Pro in particular is completely new, and there is a completely new model in the Air as well. That suggests a company that is gaining vigor, not losing it.

    Meanwhile, there is the aforementioned Services business, which is growing inexorably, thanks both to the continually growing installed base, and the fact that people continue to spend more time on their phones, not less. Yes, a lot of that Services growth comes from Google traffic acquisition cost payments and App Store fees, but those aren’t necessarily a bad thing: the former speaks to Apple’s dominant position in the attention value chain, and the former not only to the company’s hold on payments, but also the massive growth that has happened in new business models like subscriptions.

    Moreover, you can further make the case that the fundamentals that drive those businesses mean that Apple is poised to be a winner in AI, even if Apple Intelligence is delayed: Apple is positioned to be kingmaker — or gatekeeper — to AI companies who need a massive userbase to justify their astronomical investments, and to the extent that subscriptions are a core piece of the AI monetization puzzle is the extent to which the App Store is positioned for even more recurring revenue growth.

    And besides, isn’t it a good thing that Apple is unique amongst its Big Tech peers in having dramatically lower capital expenditures, even as they are making just as much money as ever? Since when did it become a crime to not just maintain but actually grow profit margins, as Apple has for the last several years?

    The Cost of Pure Profit

    Back when I started Stratechery — and back when iPhone launches were the most important days in all of tech — Apple was locked into tooth-and-nail competition with Google for the smartphone space. And, in the midst of that battle, Google made a critical error: for several years in the early 2010s the company forgot that the point of Android was to ensure access to Google services, and started using Google services to differentiate Android in its fight with the iPhone.

    The most famous example was Google Maps, a version of which launched with the iPhone. When it came time to re-up the deal Google wanted too much data and the ability to insert too many ads for Apple’s liking, so the latter launched its own product — which sucked, particularly at the beginning. Over time, however, Apple Maps has become a very competent product, and critically, it’s the default on iPhones. The implication of that is not that Apple won, but rather that Google lost: maps are a critical source of useful data for an advertising company, and Google lost a huge amount of useful signal from its most valuable users.

    The most important outcome of the early smartphone wars, however, particularly the Maps fiasco, was the extent to which both companies determined to not make the same mistake again: Google would ensure that iPhones were a first-class client for its services, and would pay ever more money for the right to be the default for Search in particular. Apple, meanwhile, seemed to get the even better end of the deal: the company would simply not compete with Google, and add those payments directly to its bottom line.

    This, of course, is why Judge Amit Mehta’s decision last week about remedies in the Google Search default placement antitrust case — specifically, the fact that he allowed those payments to continue — was hailed as a victory not just for Google but also Apple, which would see the $20+ billion of pure profit it got from Mountain View continue to flow.

    What I think is under-appreciated, however, is that the old cliché is true: nothing is free. Apple paid a price for those payments, but it’s not one that has shown up on the bottom line, at least not yet. I wrote about Maps last year in Friendly Google and Enemy Remedies and concluded:

    The lesson Google learned was that Apple’s distribution advantages mattered a lot, which by extension meant it was better to be Apple’s friend than its enemy…It has certainly been profitable for Apple, which has seen its high-margin services revenue skyrocket, thanks in part to the ~$20 billion per year of pure profit it gets from Google without needing to make any level of commensurate investment.

    That right there is the cost I’m referring to: the investment Apple might have made in a search engine to compete with Google are not costs that, once spent, are gone forever, like jewels in an iPhone game; rather, the reason it’s called an “investment” is that it pays off in the long run.

    The most immediate potential payoff would have been search ad revenues that Apple might have earned in an alternate timeline where they competed with Google instead of getting paid off by them. This, to be sure, would likely have been less on both the top and especially bottom lines, so skepticism about the attractiveness of this approach is fair.

    There is, however, another sort of payoff that comes from this kind of investment, and that’s the accumulation of knowledge and capabilities inherent in building products. In this case, Apple completely forewent any sort of knowledge or capability accumulation in terms of gathering, reasoning across, and serving large amounts of data; when you put it that way, is it any surprise that the company suddenly finds itself on the back foot when it comes to AI? Apple is suddenly trying to flex muscles that were, by-and-large, unimportant for the company’s core iPhone business because Google took care of it; had the company been competing in search for the last decade — even if they weren’t as good as Google — they would likely at a minimum have a functional Siri!

    This gets at the most amazing paradox of Mehta’s reasoning for not banning Google payments. Mehta wrote:

    If adopted, the remedy would pose a substantial risk of harm to OEMs, carriers, and browser developers…Distributors would be put to an untenable choice: either (1) continue to place Google in default and preferred positions without receiving any revenue or (2) enter distribution agreements with lesser-quality GSEs to ensure that some payments continue. Both options entail serious risk of harm.

    This is certainly true when it comes to small-scale outfits like Mozilla; Mehta, however, was worried about Apple as well. This was the second in Mehta’s list of “significant downstream effects…possibly dire” that might result from banning Google payments:

    Fewer products and less product innovation from Apple. Rem. Tr. at 3831:7-10 (Cue) (stating that the loss of revenue share would “impact [Apple’s] ability at creating new products and new capabilities into the [operating system] itself”). The loss of revenue share “just lets [Apple] do less.” Id. at 3831:19 (Cue).

    This is obviously not true in an absolute sense: Apple made just shy of $100 billion in profit over the last 12 months; losing ~20% of that would hurt, but the company would still have money to spend. Of course, you might make the case that it is true in practice, since investors might not tolerate the loss of margins that ending the Google deal would entail, which might compel management to decrease what it spends on innovation. I tend to think that investors would actually punish Apple more for innovating less, but that’s not the point I’m focused on.

    Rather, what Judge Mehta seems oblivious to is the extent to which his downstream fears already manifested. Apple has had fewer products and less innovation precisely because they have been paid off by Google, and worse, that lack of investment is compounding with the rise of AI.

    The Sugar Water Trap

    Apple took the liberty of opening yesterday’s presentation with a classic Steve Jobs quote:

    "Design is not just what it looks like and feels like. Design is how it works." — Steve Jobs

    Setting aside the wisdom of using that quote when the company is about to launch a controversial new user interface design that critics complain sacrifices legibility for beauty (although, to be honest, I don’t think it looks great either), that wasn’t the Steve Jobs quote this presentation and Apple’s general state of affairs made me think of. What I was thinking of was the question Jobs posed to then PepsiCo President John Sculley when he was recruiting him to be the CEO of Apple in the early 1980s:

    Do you want to sell sugar water for the rest of your life or come with me and change the world?

    iPhones are a great business — one of the best businesses ever, in fact — because Apple managed to marry the malleability of software with the tangibility and monetization potential of hardware. Indeed, the fact that we will always need hardware to access software — including AI — speaks to not just the profitability but also the durability of Apple’s model.

    The problem, however, is that simply staying in their lane, content to be a hardware provider for the delivery of others’ innovation, feels a lot more like Sculley than Jobs. Jobs told Walter Isaacson for his biography:

    My passion has been to build an enduring company where people were motivated to make great products. Everything else was secondary. Sure, it was great to make a profit, because that was what allowed you to make great products. But the products, not the profits, were the motivation. Sculley flipped these priorities to where the goal was to make money. It’s a subtle difference, but it ends up meaning everything: the people you hire, who gets promoted, what you discuss in meetings.

    Apple, to be fair, isn’t selling the same sugar water year-after-year in a zero sum war with other sugar water companies. Their sugar water is getting better, and I think this year’s seasonal concoction is particularly tasty. What is inescapable, however, is that while the company does still make new products — I definitely plan on getting new AirPods Pro 3s! — the company has, in pursuit of easy profits, constrained the space in which it innovates.

    That didn’t matter for a long time: smartphones were the center of innovation, and Apple was consequently the center of the tech universe. Now, however, Apple is increasingly on the periphery, and I think that, more than anything, is what bums people out: no, Apple may not be a sugar water purveyor, but they are farther than they have been in years from changing the world.



    Get notified about new Articles


  • U.S. Intel

    Listen to this post:

    Now that everyone is using ChatGPT, the lazy columnist’s trick of quoting Wikipedia to open an Article is less cliché than it is charming (at least that’s my excuse). Anyhow, here is Wikipedia’s definition of “steelmanning”:

    A steel man argument (or steelmanning) is the opposite of a straw man argument. Steelmanning is the practice of applying the rhetorical principle of charity through addressing the strongest form of the other person’s argument, even if it is not the one they explicitly presented. Creating the strongest form of the opponent’s argument may involve removing flawed assumptions that could be easily refuted or developing the strongest points which counter one’s own position. Developing counters to steel man arguments may produce a stronger argument for one’s own position.

    The beauty of being in the rather lonely position of supporting the U.S. government taking an equity stake in Intel is that I don’t have to steelman the case about it being a bad idea. Scott Lincicome, for example, had a good Twitter thread and Washington Post column explaining why this is a terrible idea; this is the opening of the latter:

    President Donald Trump’s announcement on Friday that the U.S. government will take a 10 percent stake in long-struggling Intel marks a dangerous turn in American industrial policy. Decades of market-oriented principles have been abandoned in favor of unprecedented government ownership of private enterprise. Sold as a pragmatic and fiscally responsible way to shore up national security, the $8.9 billion equity investment marks a troubling departure from the economic policies that made America prosperous and the world’s undisputed technological leader.

    Lincicome lists a number of problems with this transaction, including (but not limited to!):

    • Intel making decisions for political rather than commercial considerations
    • Intel’s board prioritizing government interests over their fiduciary duties
    • Other companies being pressured to purchase Intel products, weakening their long-term position.
    • Disadvantaging the competitive position of other companies
    • Incentivizing the misallocation of private capital

    Lincicome and all of the other critics of this deal are absolutely correct about all of the downsides. The problem with their argument, however, is the lack of steelmanning, in two respects: first, Lincicome’s Twitter thread doesn’t mention “China” or “Taiwan” once (the Washington Post column mentions China, but not in a national security context). Second, Lincicome et al refuse to grapple with the possibility that chips generally, and foundries specifically, really are a unique case.

    The Geopolitical Case

    There is a reason I’ve written so much about chips, and for many years before the AI wave brought the industry to prominence; start with 2020’s Chips and Geopolitics:

    The international status of Taiwan is, as they say, complicated. So, for that matter, are U.S.-China relations. These two things can and do overlap to make entirely new, even more complicated complications.

    Geography is much more straightforward:

    A map of the Pacific

    Taiwan, you will note, is just off the coast of China. South Korea, home to Samsung, which also makes the highest end chips, although mostly for its own use, is just as close. The United States, meanwhile, is on the other side of the Pacific Ocean. There are advanced foundries in Oregon, New Mexico, and Arizona, but they are operated by Intel, and Intel makes chips for its own integrated use cases only.

    The reason this matters is because chips matter for many use cases outside of PCs and servers — Intel’s focus — which is to say that TSMC matters. Nearly every piece of equipment these days, military or otherwise, has a processor inside. Some of these don’t require particularly high performance, and can be manufactured by fabs built years ago all over the U.S. and across the world; others, though, require the most advanced processes, which means they must be manufactured in Taiwan by TSMC.

    This is a big problem if you are a U.S. military planner. Your job is not to figure out if there will ever be a war between the U.S. and China, but to plan for an eventuality you hope never occurs. And in that planning the fact that TSMC’s foundries — and Samsung’s — are within easy reach of Chinese missiles is a major issue.

    The rise of AI makes these realities — and related issues like chip controls — even more pressing. I made the argument earlier this year in AI Promise and Chip Precariousness that the U.S. should be seeking to make China more dependent on both U.S. chip companies and TSMC manufacturing, even as it was doing the opposite. The motivation was to preserve dominance in AI, but this ignored the reality I just laid out: AI depends on chips, and those chips are made next door to China; that means that stopping China could be worse than China succeeding:

    It’s also worth noting that success in stopping China’s AI efforts has its own risks: another reason why China has held off from moving against Taiwan is the knowledge that every year they wait increases their relative advantages in all the real world realities I listed above; that makes it more prudent to wait. The prospect of the U.S. developing the sort of AI that matters in a military context, however, even as China is cut off, changes that calculus: now the prudent course is to move sooner rather than later, particularly if the U.S. is dependent on Taiwan for the chips that make that AI possible.

    Beyond the human calamity that would result from a Chinese attack on Taiwan, there is the economic calamity downstream of not just losing AI chips, but chips of all sorts, including the basic semiconductors that power not just computers but basically everything in the world. And, to that end, it’s worth pointing out that an Intel that succeeds doesn’t fully address our chip dependency on Taiwan. It is, however, a pre-requisite, and any argument about the U.S. government’s involvement with Intel must grapple with this reality.

    Decisions Over Decades

    There was one line in Lincicome’s Article that definitely made me raise my eyebrows (emphasis mine):

    The semiconductor industry, more than most, requires nimble responses to rapidly changing technology and market conditions. Intel already faces significant operational and competitive challenges; it has been a technological laggard for more than a decade as Nvidia, AMD, TSMC and other competitors have raced ahead. Adding a layer of political oversight to Intel’s already-complex turnaround effort is far more likely to hinder than help.

    I get Lincicome’s point, which certainly applies to the technology industry broadly; just look at all of the upheaval that has happened in the two-and-a-half years since ChatGPT launched. I would argue, however, that chips are different: Intel is a technological laggard because of choices made decades ago; it just takes a really long time for the consequences of mistakes to show up.

    Starting a new blog is a bit like a band publishing their debut album: you’re full of takes that you’ve held for years and have been waiting to unleash. In my case, I had been worried about Intel ever since they missed out on mobile, which meant they missed out on the associated volume that came from making chips for every smartphone in the world. Volume is critical when it comes to managing the ever-expanding cost of staying on the leading edge: as the cost of fabs has surged from hundreds of millions to tens of billions of dollars, the only way to fab chips profitably is to have enough volume over which to spread those massive capital expenditures.

    And so, within a month of launching Stratechery, I wrote that Intel needed to become a foundry — i.e. make chips for other companies — if they wanted to remain viable in the long run. And, to be honest, I had been saving up that take for so long that I thought I was too late; after all, I started Stratechery in 2013, six years into the mobile era, and given the massive changes Intel would have to undergo to become a customer service organization, I thought they needed to make that change at least three years earlier.

    And then, for the next eight years, Intel’s stock went up and up, as the company rode the cloud boom that was the yin to the smartphone’s yang. If anyone had read my 2013 Article and sold their Intel shares, or worse, shorted them, they would have lost their shirt!

    In the end, however, my take was correct, even if it was un-investable. First Intel fell behind TSMC, who was powered by massive orders from Apple in particular, and then, on the company’s last earnings call, CEO Lip-Bu Tan admitted the reality of what could have been forecasted when Steve Jobs walked onto that 2007 MacWorld stage:

    Up to and through Intel 18A, we could generate a reasonable return on our investments with only Intel Products. The increase in capital cost at Intel 14A, make it clear that we need both Intel products, and a meaningful external customer to drive acceptable returns on our deployed capital, and I will only invest when I’m confident those returns exist.

    This is the rotten tree sprung from the seed of Intel’s mobile failure: the company could afford to miss out on a massive market for nearly two decades, but when it comes to 14A, the company simply can’t sell enough chips on its own to justify the investment.

    What is worse is the tree that wasn’t planted: the real payoff from Intel building a foundry business in 2010, or 2013 as I recommended, is that they would have been ready for the AI boom. Every hyperscaler is still complaining that demand exceeds supply for AI chips, even as Intel can’t win customers for its newest process that is actually best-suited for AI chips. The company simply has too many other holes in its offering, including the sort of reliability and throughput that is essential to earning customer trust.

    In short, contra Lincicome, Intel’s problem is not short-term decision-making, because Intel is in the business of making chips, and making chips is a decades-long endeavor of building expertise, gaining volume, moving down the learning curve, and doing it all again and again to the tune of tens of billions of dollars a year in capex.

    That, by extension, is why the stakes today are so high. The problem facing the U.S. is not simply the short-term: the real problems will arise in the 2030s and beyond. Semiconductor manufacturing decision-making does not require nimbleness; it requires gravity and the knowledge that abandoning the leading edge entails never regaining it.

    Competing with TSMC

    This also puts to rest one of the traditional objections to government intervention in support of an incumbent: in almost every case that investment crowds out new companies, companies that are, yes, more nimble and more capable of meeting the moment. The reality of semiconductor manufacturing, however, is that the path is far too long and arduous to ever fill the vacuum that Intel’s exit would leave. Actually, though, that last line is not quite right: Intel’s biggest problem is that its market challenges are closer to that mythical startup that will never exist.

    Suppose our mythical startup somehow received hundreds of billions of dollars worth of funding, and somehow moved down the decades-long learning curve that undergirds modern silicon manufacturing: to make the business work our mythical startup would actually need to find customers.

    Our mythical startup, however, doesn’t exist in a vacuum: it exists in the same world as TSMC, the company who has defined the modern pure play foundry. TSMC has put in the years, and they’ve put in the money; TSMC has the unparalleled customer service approach that created the entire fabless chip industry; and, critically, TSMC, just as they did in the mobile era, is aggressively investing to meet the AI moment. If you’re an Nvidia, or an Apple in smartphones, or an AMD or a Qualcomm, why would you take the chance of fabricating your chips anywhere else? Sure, TSMC is raising prices in the face of massive demand, but the overall cost of a chip in a system is still quite small; is it worth risking your entire business to save a few dollars for worse performance with a worse customer experience that costs you time to market and potentially catastrophic product failures?

    We know our mythical startup would face these challenges because they are the exact challenges Intel faces. Intel may need “a meaningful external customer to drive acceptable returns on [its] deployed capital”, but Intel’s needs do not drive the decision-making of those external customers, despite the fact that Intel, while not fully caught up to TSMC, is at least in the ballpark, something no startup could hope to achieve for decades.

    Intel’s Credibility Problem

    These realities are why I argued a year ago that the U.S. government needed to prop up demand for Intel manufacturing, a point I reiterated earlier this year. And, to steelman the argument of those opposed to this deal, there are ways to do that without acquiring part of the company.

    The problem, however, comes back to what Tan said on that earnings call: beyond all of the challenges above, what company is going to go through the trouble of getting their chip working on Intel’s process if it’s possible that the company is going to abandon manufacturing on the next process? It’s a catch-22: Intel needs an external customer to make its foundry viable, but no external customer will go with Intel if there is a possibility that Intel Foundry will not be viable. In other words, the stakes have changed from even earlier this year: Intel doesn’t just need demand, it needs to be able to credibly guarantee would-be customers that it is in manufacturing for the long haul.

    A standalone Intel cannot credibly make this promise. The path of least resistance for Intel has always been to simply give up manufacturing and become another TSMC customer; they already fab some number of their chips with the Taiwanese giant. Such a decision would — after some very difficult write-offs and wind-down operations — change the company into a much higher margin business; yes, the company’s chip designs have fallen behind as well, but at least they would be on the most competitive process, with a lot of their legacy customer base still on their side.

    The problem for the U.S. is that that then means pinning all of the country’s long-term chip fabrication hopes on TSMC and Samsung not just building fabs in the United States, but also building up a credible organization in the U.S. that could withstand the loss of their headquarters and engineering knowhow in their home countries. There have been some important steps in this regard, but at the end of the day it seems reckless for the U.S. to place both its national security and its entire economy in the hands of foreign countries next door to China, allies or not.

    Given all of this, acquiring 10% of Intel, terrible though it may be for all of the reasons Lincicome articulates — and I haven’t even touched on the legality of this move — is I think the least bad option. In fact, you can even make the case that a lot of what Lincicome views as a problem has silver linings:

    • Intel deciding to stay in manufacturing is arguably making a political decision, not a commercial one; however, it is important for the U.S. that Intel stay in manufacturing.
    • Intel prioritizing government interests — which are inherently focused on national security and the long-term viability of U.S. semiconductor manufacturing — over their fiduciary duties could just as easily be framed as valuing the long-term over the short term; had Intel done just that over the last two decades they wouldn’t be in this position.
    • Other companies being pressured to purchase Intel products is exactly what Intel needs to not just be viable in manufacturing, but also to actually get better.
    • If TSMC and Samsung are disadvantaged by not making chips in the U.S., that’s a good thing from the U.S. perspective. Both companies are investing here; the U.S. wants more.
    • Private capital prioritizing U.S. manufacturing is a good thing!

    The single most important reason for the U.S. to own part of Intel, however, is the implicit promise that Intel Foundry is not going anywhere. There simply isn’t a credible way to make that promise without having skin in the game, and that is now the case.

    Steelmanning

    I’ll be honest: there is a very good chance this won’t work. Intel really is a mess: they are actively hostile to customers, no one in the industry trusts them, they prioritize the wrong things even today (i.e. technical innovation with backside power over yields for chips which don’t necessarily have interference issues), and that’s even without getting into the many problems with their business. Moreover, I led with Lincicome’s argument because I agree! Government involvement in private business almost always ends badly.

    At the same time, the China concerns are real, Intel Foundry needs a guarantee of existence to even court customers, and there really is no coming back from an exit. There won’t be a startup to fill Intel’s place. The U.S. will be completely dependent on foreign companies for the most important products on earth, and while everything may seem fine for the next five, ten, or even fifteen years, the seeds of that failure will eventually sprout, just like those 2007 seeds sprouted for Intel over the last couple of years. The only difference is that the repercussions of this failure will be catastrophic not for the U.S.’s leading semiconductor company, but for the U.S. itself.



    Get notified about new Articles


  • Facebook is Dead; Long Live Meta

    Listen to this post:

    I don’t do memes frequently, but I’ve used this one twice:

    The GTA meme about "Here we go again" as applied to Facebook

    Alas, I think I need to retire it. From the Wall Street Journal:

    Meta Platforms posted 22% revenue growth in the second quarter, showing its core ad business remains strong at a time when the company is investing billions of dollars into artificial intelligence. For the first time this year, the company held its top capital-spending projections steady, a move that is sure to ease investor concerns over Chief Executive Mark Zuckerberg’s AI spending spree. Shares rose by more than 11% in after-hours trading.

    The results show the extent to which Meta’s advertising business will continue to pay for its outsize AI ambitions, as well as how the AI tools are contributing to its strength. Sales came in at $47.5 billion, ahead of analyst expectations. Net income for the April-to-June period was $18.3 billion, also ahead of market expectations. Meta also said it expects to post between 17% and 24% revenue growth year-over-year for the current quarter. “The investments it’s making in AI are already paying off in its ads business,” said Jasmine Enberg, principal analyst at research firm eMarketer.

    I’m obviously a bit tardy getting to these earnings, which is something that wouldn’t have happened if Meta had had very bad results the same quarter in which they started very publicly spending hundreds of millions of dollars on talent to overhaul their AI strategy. This stands in marked contrast to the 2021-2022 period, when Meta renamed itself from Facebook in conjunction with unveiling just how much money it was spending/losing on the Metaverse, even as its core business slowed significantly, driving the stock to below $100 and compelling me to write Meta Myths.

    Lessons Learned

    This wasn’t an accident, but rather evidence of a lesson learned; CFO Susan Li explained what that lesson was on an episode of John Collison’s Cheeky Pint podcast:

    Susan Li: It’s pretty standard after earnings calls, where you touch base with some number of your largest investors. Sadly, it is not one-on-one, it’s one of you and many, many people from their teams. And most of the time, they just ask you to clarify things. Obviously, everything is Reg FD compliant, but it mostly takes the form of questions. And in October 2022, for the first time, there were sometimes no questions. I mean, there was a call where basically one of the portfolio managers said, “We actually don’t have any questions for you today. We just want you to hear feedback from us.”

    John Collison: Wow. More of a comment than a question.

    SL: Yes. It was actually very memorable.

    JC: And it was blunt feedback, I presume.

    SL: Yes. And one of the things that really stuck with me from one of those conversations is someone said, “Look, I get that you’re building the future of computing and the next mobile platform and all that, and that is great, and I am glad someone wants to do it and I am rooting for you, but why should I invest in your stock today? Why don’t I just wait for your phone equivalent, your scaled consumer product to come out and invest in you then, and you tell me that that’s going to be years away?”

    And the way that question was framed actually really stuck with me, and is the way that, frankly, now Mark and I think about this. Which is like, great, we’ve got a lot of these bets, and the bets are technologically exciting. People can get excited about them and the vision of the world. But as investors, they’re like, “Cool, why don’t I just wait for your bets to be ready to succeed before I come?” We need people to invest with us along the way. When we think about the financial outlook of the company, a large part of it is not just, okay, cool, you’re building the next massive platform out here in some decades, it’s, why would you hold our shares until then? What do we need to keep delivering in terms of consolidated results?

    The very first words out of Zuckerberg’s mouth on the earnings call stated Meta’s new thinking explicitly:

    We had another strong quarter with more than 3.4 billion people using at least one of our apps each day, and strong engagement across the board. Our business continues to perform very well, which enables us to invest heavily in our AI efforts.

    There have been historical examples of public companies having permission from investors to invest far ahead of profits, most notably Amazon. Matthew Yglesias famously wrote on Slate in 2013:

    Amazon kept up its streak of being awesome this afternoon by announcing a 45 percent year-on-year decline in profits measuring Q4 2012 against Q4 2011. Not because sales went down, mind you. They’re up. Revenue is up. The company’s razor-thin profit margins just got even thinner, and in total the company lost $39 million in 2012.

    The company’s shares are down a bit today, but the company’s stock is taking a much less catastrophic plunge in already-meager profits than Apple, whose stock plunged simply because its Q4 profits increased at an unexpectedly slow rate. That’s because Amazon, as best I can tell, is a charitable organization being run by elements of the investment community for the benefit of consumers. The shareholders put up the equity, and instead of owning a claim on a steady stream of fat profits, they get a claim on a mighty engine of consumer surplus. Amazon sells things to people at prices that seem impossible because it actually is impossible to make money that way. And the competitive pressure of needing to square off against Amazon cuts profit margins at other companies, thus benefiting people who don’t even buy anything from Amazon.

    It’s a truly remarkable American success story. But if you own a competing firm, you should be terrified. Competition is always scary, but competition against a juggernaut that seems to have permission from its shareholders to not turn any profits is really frightening.

    This was obviously incorrect: Amazon was “losing money” (while still having fabulous free cash flow, mind you) because they were investing heavily in the infrastructure that supported The Amazon Tax; for several years now the company’s biggest “problem” has arguably been that they are throwing off more money than they can manage to reinvest, even with their massive buildout of AWS and their logistics network. Critically, however, it is easy to draw a line from those two buildouts in particular to future cash flows, and investors were willing to give Jeff Bezos the benefit of the doubt.

    One of the weird things about Meta, on the other hand, is how the company never seemed to get the benefit of the doubt, at least up until the last couple of years; I wrote in Meta Myths:

    Myspace is, believe it or not, still around; however, it has been irrelevant for so long that I needed to look it up to remember if the name used camel case or not (it doesn’t). It does, though, still seem to loom large in the mind of Meta skeptics certain that the mid-2000’s social network’s fate was predictive for the company that supplanted it.

    There is something about social media that has always made investors intrinsically suspicious of the long-term; given that, Meta’s simultaneous slowdown in growth (mostly driven by ATT), combined with the public shift in focus to the Metaverse, fueled concern that the company was desperately trying to pivot away from a failing business — and losing billions of dollars to do so, with very few tangible results.

    Things are very different now, however, in two regards. First, there is the aforementioned commitment by management to deliver results now, not just promises about the future; second, investor sentiment about Meta really has done a 180: if anything, my impression is that Meta is not just getting the benefit of the doubt, but actually more credit than they deserve.

    AI That Matters

    This is a very good quarter to deliver the updated version of my favorite Meta chart:

    Last quarter I wrote about Meta’s Deteriorating Ad Metrics:

    The first thing to note is the leveling out of impression growth even as Daily Active Person’s growth increased slightly; what this means is that the Reels inventory growth tailwind may be over. That’s always a bearish signal, as I noted a year ago in Meta and Reasonable Doubt

    Secondly, and more concerningly, is that while price-per-ad growth outpaced impression growth, it also decreased quarter-over-quarter. One of the reasons why 2023 was such a good time to own Meta is that a combination of figuring out ATT and expanding Reels inventory meant that both impressions and price-per-ad increased simultaneously; in contrast, it is a lot less attractive when both are decreasing simultaneously…

    The two previous quarters where the growth rates for impressions and price-per-ad simultaneously decreased were 4Q 2016 and 1Q 2019; the positive takeaway is that I’m not actually totally certain what happened in 4Q 2016, and I wasn’t concerned at the time. 1Q 2019 is more interesting, in that I chalked up the issue to increased time spent on Stories cannibalizing Feed at a time when Stories had fewer impressions and worse monetization; to that end, the positive spin on this point is that Meta still has a lot of room to increase Reels monetization.

    This quarter looks much better: impressions increased growth significantly; when that happens you should expect price-per-ad to decline, but notably, the decline was relatively attenuated compared to last quarter. This is a great result that drove the blowout numbers that made investors willing to give Zuckerberg a big fat permission slip for his AI talent splurge.

    What, though, were the drivers of these significantly improved ad fundamentals? Zuckerberg was quick to credit AI:

    On advertising, the strong performance this quarter is largely thanks to AI unlocking greater efficiency and gains across our ads system. This quarter, we expanded our new AI-powered recommendation model for ads to new surfaces and improved its performance by using more signals and a longer context. It’s driven roughly 5% more ad conversions on Instagram and 3% on Facebook.

    I think this credit is fair, particularly in terms of the relative lack of decline in price-per-ad (given the impressions growth). However, what is important to note is that the “AI” Zuckerberg is referring to is not LLMs; Li, to her credit, spent an extensive amount of time in her prepared remarks detailing the AI systems — Andromeda, GEM, and Lattice — that are actually driving these improvements. These are all impressive, to be sure, but the reason I suggested that investor sentiment may have swung too much away from suspicion to gullibility was the way in which I saw a lot of folks conflate Meta’s new SuperIntelligence investments with their great results, as if the former caused the latter; in fact, Zuckerberg clearly stated that they were totally different initiatives:

    I think the trajectory on this stuff is very optimistic. And I think it’s one of the interesting challenges in running a business like this now is there’s just a very high chance, it seems, like the world is going to look pretty different in a few years from now. And on the one hand, there are all these things that we can do, there are improvements to our core products that exist.

    It’s the “improvements to our core products that exist” that drove these results; Zuckerberg now knows — but seemingly still chafes, just a tad — that there is an opportunity cost that must be paid in terms of driving those improvements to get permission to invest in superintelligence. To put it another way, (1) Zuckerberg really wants to focus on the latter, and (2) now knows that the way to get investor permission to do so is to deliver short-term results. That, by extension, makes me think the AI story isn’t the whole story when it comes to Meta’s ads.

    Reels and Dials

    To go back to the chart above, previous drivers of increased impressions growth were meaningful product additions that vastly increased ad inventory; from Meta and Reasonable Doubt:

    There are two-and-a-half inversions of impression and price-per-ad growth rates on this chart:

    • In 2017 Meta saturated the Instagram feed with ads; this led impressions growth to drop and the price-per-ad to increase; then, in 2018, Instagram started to monetize Stories, leading to increased growth in impressions and corresponding decreases in price-per-ad growth.
    • In 2020 Meta saturated Instagram Stories; this once again led impressions growth to drop and the price-per-ad to increase; then, while COVID provided a boost in 2021, 2022 saw a significant increase in Reels monetization, leading to increased growth in impressions and a decrease in price-per-ad growth (which, as noted above, was made more extreme by ATT).
    • Since the middle of last year Meta Impressions growth is once again dropping as Reels becomes saturated; this is leading to an increase in price-per-ad growth (although the lines have not yet crossed).

    The most optimistic time for Meta’s advertising business is, counter-intuitively, when the price-per-ad is dropping, because that means that impressions are increasing. This means that Meta is creating new long-term revenue opportunities, even as its ads become cost competitive with more of its competitors; it’s also notable that this is the point when previous investor freak-outs have happened.

    The problem I saw in that Article is that Meta didn’t have any obvious product improvements on the horizon that would meaningfully increase inventory. Yes, Threads and WhatsApp ads are coming, but they’re still not big revenue drivers. And yet, we got impression growth all the same! Here was the closest thing we got to an explanation from Li in her prepared remarks:

    Impression growth accelerated across all regions due primarily to engagement tailwinds on both Facebook and Instagram and, to a lesser extent, ad load optimizations on Facebook.

    Li also said on the results follow-up call:

    So on your first question about impression growth, the worldwide impression growth acceleration that we saw in Q2 was driven primarily by incremental engagement on video and Feed surfaces, which benefited from a bunch of the ranking optimizations that we made to our content recommendations on Facebook and Instagram, and then to a lesser extent some of the ad load optimizations.

    It turns out I was right last quarter that Meta had a lot of room to increase Reels monetization, but not just because they could target ads better (that was a part of it, as I noted above): rather, it turns out that short-form video is so addictive that Meta can simply drive more engagement — and thus more ad inventory — by pushing more of it. That’s impression driver number one — and the most important one. The second one is even more explicit: Meta simply started showing more ads to people (i.e. “ad load optimization”).

    All of this ties back to where I started, about how Meta learned that you have to give investors short term results to get permission for long term investments. I don’t think it’s a coincidence that, in the same quarter where Meta decided to very publicly up its investment in the speculative “Superintelligence”, users got pushed more Reels and Facebook users in particular got shown more ads. The positive spin on this is that Meta has dials to turn; by the same token, investors who have flipped from intrinsically doubting Meta to intrinsically trusting them should realize that it was the pre-2022 Meta, the one that regularly voiced the importance of not pushing too many ads in order to preserve the user experience, that actually deserved the benefit of the doubt for growth that was purely organic. This last quarter is, to my mind, a bit more pre-determined.

    Social Network R.I.P.

    The link in the previous paragraph is to then-Facebook’s Q3 2016 earnings call; that’s as good a place as any to find one of a bajillion quotes from Zuckerberg about connecting people:

    Over the next 10 years, we’re going to continue to invest in the platforms and technologies that will connect more people and more places and allow everyone in the world to have a voice. We focused our long-term innovation roadmap around three areas: connectivity initiatives that bring more people online; artificial intelligence; and virtual and augmented reality.

    The “artificial intelligence” and “virtual and augmented reality” investments are obviously still there in a major way; what is interesting to note, however, is that the word “connect” — at least in the context of what Zuckerberg repeatedly stated was Facebook’s mission — only appeared once on this earnings call:

    Over the last few months we have begun to see glimpses of our AI systems improving themselves. The improvement is slow for now, but undeniable. Developing superintelligence — which we define as AI that surpasses human intelligence in every way — we think is now in sight.

    Meta’s vision is to bring personal superintelligence to everyone — so that people can direct it towards what they value in their own lives. We believe this has the potential to begin an exciting new era of individual empowerment. A lot has been written about the economic and scientific advances that superintelligence can bring. I am extremely optimistic about this. But I think that if history is a guide, then an even more important role will be how superintelligence empowers people to be more creative, develop culture and communities, connect with each other, and lead more fulfilling lives.

    To build this future, we’ve established Meta Superintelligence Labs, which includes our foundations, product, and FAIR teams, as well as a new lab that is focused on developing the next generation of our models. We’re making good progress towards Llama 4.1 and 4.2 — and in parallel, we’re also working on our next generation of models that will push the frontier in the next year or so.

    Long gone are the days when Meta’s self-identification as a social network led them to miss fundamental shifts in the Internet; this commitment to delivering superintelligence, with the expressed goal of helping people “direct it towards what they value in their own lives” — note the explicit callout of individualism, not community — is the end of a shift that Meta was both late to and has also been undergoing for years.

    And, I would note, it’s that shift that provided the dials I wrote about in the previous section. Pushing people ever more short-form video — with ever more advertisements — is intrinsically isolating and contrary to social interaction (making it easier to share Reels with your chat groups notwithstanding). Indeed, I think this connection is inevitable: I suspect that what companies most strenuously push as their mission is often the most favorable interpretation of their downsides. Apple talks about integration, while their critics argue they control too much; Google talks about capturing knowledge and making it useful, while their critics argue they violate privacy; Meta used to talk about connection, while their critics fretted about the negative effects of peer pressure and falling into the wrong communities, and many investors nervously awaited the next viral social network. Now Meta talks about empowering the individual, even as they justify their investment to investors by pushing users ever deeper into individually-tailored short form video feeds.

    At a minimum, it gives the company a solid legal defense; the company recently wrote in a filing defending itself from the FTC’s charge that it monopolized social networking:

    The evidence decisively demonstrated that Meta – once an online “Facebook” for connecting students – has evolved into a diverse global provider of entertaining and informative content that competes with increasingly similar social apps including TikTok, YouTube, iMessage, and others. Times, technologies, and norms of use change, and the trial evidence proved that Meta has adapted due to competitive pressure in this fast-moving industry. Today, only a fraction of time spent on Meta’s services – 7% on Instagram, 17% on Facebook – involves consuming content from online “friends” (“friend sharing”). A majority of time spent on both apps is watching videos, increasingly short-form videos that are “unconnected” – i.e., not from a friend or followed account – and recommended by AI-powered algorithms Meta developed as a direct competitive response to TikTok’s rise, which stalled Meta’s growth. The FTC now concedes this development “brings Meta into competition with TikTok and YouTube.” That concession means that there is no valid PSNS market, which is the sole market the FTC asserts.

    This is obviously correct; indeed, maybe investors were actually right all along: being a social network wasn’t ultimately sustainable, and the fact that Meta is stronger than ever, with entirely new justifications for being, is the real reason why they deserve the benefit of the doubt.



    Get notified about new Articles


  • Paradigm Shifts and the Winner’s Curse

    Listen to this post:

    It’s fun — and often accurate — to think of tech companies in pairs. Apple and Microsoft defined the PC market; Microsoft and Intel won it. Google and Meta dominate digital advertising; Apple and Google won mobile. That, however, is not the defining pair of the smartphone era, which ran from the introduction of the iPhone in 2007 to the launch of ChatGPT in 2022; rather, the two most important companies of the last two decades of tech were Apple and Amazon, specifically AWS.

    The Apple part is easy: the iPhone market created the smartphone paradigm, from its user interface (touch) to its distribution channel (the App Store), and was richly rewarded with a bit under half of the unit marketshare and a bit under all of the total profits. Google did well to control the rest in terms of the Android operating system, and profit from it all thanks to Google Search, but it was Search that remained their north star; the company’s primary error in the era was the few years they let the tail (Android) wave the dog (Google).

    The AWS part is maybe less obvious, but no less critical — and the timing is notable. Amazon created AWS in 2006, just 10 months before the iPhone unveiling, and the paradigm they created was equally critical to the smartphone era. I explained the link in 2020’s The End of the Beginning:

    This last point gets at why the cloud and mobile, which are often thought of as two distinct paradigm shifts, are very much connected: the cloud meant applications and data could be accessed from anywhere; mobile made the I/O layer available everywhere. The combination of the two make computing continuous.

    A drawing of The Evolution of Computing

    What is notable is that the current environment appears to be the logical endpoint of all of these changes: from batch-processing to continuous computing, from a terminal in a different room to a phone in your pocket, from a tape drive to data centers all over the globe. In this view the personal computer/on-premises server era was simply a stepping stone between two ends of a clearly defined range.

    AWS was not the only public cloud provider, of course — Azure and Google Cloud Platform were both launched in 2008 — but by virtue of being first they both defined the paradigm and also were the the first choice of the universe of applications that ran on smartphones or, more accurately, ran everywhere.

    Smartphone Winners and Losers

    If Apple and AWS were the definers — and thus winners — of the smartphone era, then it was Microsoft and Nokia that were the losers. The reasons for their failure were myriad, but there was one common thread: neither could shake off the overhang of having won their previous paradigm; indeed, both failed in part because they deluded themselves into thinking that their previous domination was an advantage.

    For Microsoft that previous paradigm was the PC and the Windows platform, which the company thought they could extend to mobile; from 2014’s Microsoft’s Mobile Muddle:

    Saying “Microsoft missed mobile” is a bit unfair; Windows Mobile came out way back in 2000, and the whole reason Google bought Android was the fear that Microsoft would dominate mobile the way they dominated the PC era. It turned out, though, that mobile devices, with their focus on touch, simplified interfaces, and ARM foundation, were nothing like PCs. Everyone had to start from scratch, and if starting from scratch, by definition Microsoft didn’t have any sort of built-in advantage. They were simply out-executed.

    It took Microsoft years — and a new CEO — to realize their mistake, up and to the point where they put their enterprise productivity dominance at risk; from 2015’s Redmond and Reality:

    There’s reality, and there’s Redmond, and if one thing marked the last few years of Steve Ballmer’s tenure as the CEO of Microsoft, it was the sense that those were two distinct locales. In reality, Android (plus AOSP in China) and iOS were carving up the world phone market; in Redmond Ballmer doubled-down on the losing Window Phone bet by buying Nokia. In reality Office was losing relevance because of its absence on the mobile platforms that mattered; in Redmond Ballmer personally delayed Office on iOS until the Windows Modern née Metro version was finished. And in reality, all kinds of startups were taking aim at the Microsoft enterprise stack; in Redmond, Microsoft was determined to own it all, just as they had in the PC era.

    It’s fitting that Microsoft and Nokia ended up together; perhaps they were able to jointly go to therapy for success-induced obliviousness of market realities. Nokia dominated the phone market for the decade prior to the iPhone, and even once the iPhone was announced, blithely assumed that they could simply lean on their existing advantages to fend off the Silicon Valley usurper. From 2013’s Blackberry — and Nokia’s — Fundamental Failing:

    Nokia dominated all the parts of this stack you don’t see: they had, and in some respects, still have, the best supply chain and distribution network. In addition, they had high quality hardware that served every segment imaginable. Notably absent in these strengths is the OS and Apps. By 2009, BlackBerry OS and Symbian were clearly obsolete, and their app ecosystems, such as they were, were eclipsed by iOS and then Android. The problem, as I alluded to above, is that while the OS was ultimately under the control of BlackBerry and Nokia, respectively, and thus could be fixed, the efficacy of their ecosystem wasn’t, and wouldn’t be…

    And so, by far the smartest strategic thing either could have done would have been to accept their weakness — they didn’t have an adequate OS or ecosystem — and focus on their strengths…Nokia should have adopted Android-stock, and used their unmatched supply chain and distribution to do to their competitors, well, exactly what Nokia had been doing to their competitors for the last decade (if you think Samsung is running roughshod over everyone today, in 2007 they could only manage 41 million phones compared to Nokia’s 110 million).

    Both BlackBerry and Nokia would have gotten a good OS and thriving ecosystem for free and been able to compete and differentiate themselves on the exact same vectors they had previously. To put it another way, RIM and Nokia had never been successful because of their OS or ecosystem, yet both decided their best response to iOS and Android was to build a new OS! In fact, the strategic superiority of the Android option for RIM and Nokia was even then so obvious that I suspect their core failing was not so much strategic as it was all-too-human: pride. Owning an ecosystem seems much more important than owning services or supply chains, even if building said ecosystem completely devalues what you’re actually good at.

    If the first commonality in Microsoft and Nokia’s failure is the assumption that dominance in one paradigm would seamlessly translate into dominance in the next, then the second was in not making the strategically obvious choice — embracing iOS and Android for Windows, and Android for Nokia — for fear of losing control and long-term relevance. What separates the two companies is that Microsoft, under CEO Satya Nadella, rectified their mistake, while Nokia doubled-down with Windows Phone; that is why Microsoft still matters today — more than ever, in fact — while Nokia phones no longer exist.

    The two companies that stood in constrast to Microsoft and Nokia were Google and Samsung; while their dominance of the non-iPhone market seems obvious in retrospect, it wasn’t at all pre-ordained. What is impressive about both companies is that they had the opposite of pride: both were quite shameless, in fact. From 2013’s Shameless Samsung:

    Every pre-iPhone phone maker is irrelevant, if they even exist, except for Samsung, who is thriving. Samsung the copycat was smart enough to realize they needed to change, and quickly, and so they did.

    Or maybe it wasn’t being smart. Maybe it was simply not caring what anyone else thought about them, their strategy, or their inspiration. Most successful companies, including Apple, including Google, seem remarkably capable of ignoring the naysayers and simply doing what is right for their company. In the case of smartphones, why wouldn’t you copy the iPhone? Nokia refused and look where that got them!

    We, especially in the West, have a powerful sense of justice and fairness when it comes to product features and being first. Business, though, is not fair, even if it is more just than we care to admit.

    Just as Samsung blatantly copied Apple hardware, Android blatantly copied the iOS interface:

    Android as a concept existed pre-iPhone; as a product, not so much

    Plenty of people mocked Google for this shift, but not me: Apple figured out what worked; it would have been foolish to not copy them.

    Foolish like Microsoft and Nokia.

    Apple, Amazon, and AI

    There were striking resemblances in last week’s earnings calls from Apple and Amazon, not just to each other, but to this early smartphone era that I have just recounted. Both companies are facing questions about their AI strategies — Apple for its failure to invest in a large language model of its own, or deeply partner with a model builder, and Amazon for prioritizing its own custom architectures and under-deploying leading edge Nvidia solutions — and both had similar responses:

    It’s Early

    Tim Cook (from a post-earnings all-hands meeting):

    Cook struck an optimistic tone, noting that Apple is typically late to promising new technologies. “We’ve rarely been first,” the executive told staffers. “There was a PC before the Mac; there was a smartphone before the iPhone; there were many tablets before the iPad; there was an MP3 player before iPod.” But Apple invented the “modern” versions of those product categories, he said. “This is how I feel about AI.”

    Andy Jassy:

    The first thing I would say is that I think it is so early right now in AI. If you look at what’s really happening in the space, it’s very top heavy. So you have a small number of very large frontier models that are being trained that spend a lot on computing, a couple of which are being trained on top of AWS and others are being trained elsewhere. And then you also have, I would say, a relatively small number of very large-scale generative AI applications.

    We Will Serve Actual Use Cases

    Tim Cook:

    We see AI as one of the most profound technologies of our lifetime. We are embedding it across our devices and platforms and across the company. We are also significantly growing our investments. Apple has always been about taking the most advanced technologies and making them easy to use and accessible for everyone, and that’s at the heart of our AI strategy. With Apple Intelligence, we’re integrating AI features across our platforms in a way that is deeply personal, private, and seamless, right where users need them.

    Andy Jassy:

    We have a very significant number of enterprises and startups who are running applications on top of AWS’ AI services and but, like the amount of usage and the expansiveness of the use cases and how much people are putting them into production and the number of agents that are going to exist, it’s still just earlier stage than it’s going to be, and so then when you think about what’s going to matter in AI, what are customers going to care about when they’re thinking about what infrastructure use, I think you kind of have to look at the different layers of the stack. And I think…if you look at where the real costs are, they’re going to ultimately be an inference today, so much of the cost in training because customers are really training their models and trying to figure out to get the applications into production.

    Our Chips Are Best

    Tim Cook:

    Apple Silicon is at the heart of all of these experiences, enabling powerful Apple Intelligence features to run directly on device. For more advanced tasks, our servers, also powered by Apple Silicon, deliver even greater capabilities while preserving user privacy through our Private Cloud Compute architecture. We believe our platforms offer the best way for users to experience the full potential of generative AI. Thanks to the exceptional performance of our systems, our users are able to run generative AI models right on their Mac, iPad, and iPhone. We’re excited about the work we’re doing in this space, and it’s incredibly rewarding to see the strong momentum building.

    Andy Jassy:

    At scale, 80% to 90% of the cost will be an inference because you only train periodically, but you’re spinning out predictions and inferences all the time, and so what they’re going to care a lot about is they’re going to care about the compute and the hardware they’re using. We have a very deep partnership with Nvidia and will for as long as I can foresee, but we saw this movie in the CPU space with Intel, where customers are anchoring for better price performance. And so we built just like in the CPU space, where we built our own custom silicon and building Graviton which is about 40% more price performance than the other leading x86 processors, we’ve done the same thing on the custom silicon side in AI with Trainium and our second version of Trainium2…it’s about 30% and 40% better price performance than the other GPU providers out there right now, and we’re already working on our third version of Trainium as well. So I think a lot of the compute and the inference is going to ultimately be run on top of Trainium2.

    We Have the Data

    Tim Cook:

    We’re making good progress on a more personalized Siri, and we do expect to release the features next year, as we had said earlier. Our focus from an AI point of view is on putting AI features across the platform that are deeply personal, private, and seamlessly integrated, and, of course, we’ve done that with more than 20 Apple Intelligence features so far, from Visual Intelligence to Clean Up to Writing Tools and all the rest.

    Andy Jassy:

    People aren’t paying as close attention as they will and making sure that those generative AI applications are operating where the rest of their data and infrastructure. Remember, a lot of generative AI inference is just going to be another building block like compute, storage and database. And so people are going to actually want to run those applications close to where the other applications are running, where their data is. There’s just so many more applications and data running in AWS than anywhere else.

    Both Apple and Amazon’s arguments are very plausible! To summarize each:

    Apple: Large language models are useful, but will be a commodity, and easily accessible on your iPhone; what is the most useful to people, however, is AI that has your private data as context, and only we can provide that. We will provide AI with your data as context at scale and at low cost — both in terms of CapEx and OpEx — by primarily running inference on device. People are also concerned about sharing their personal data with AI companies, so when we need more capabilities we will use our own compute infrastructure, which will run on our own chips, not Nvidia chips.

    Amazon: Large language models are useful, but will be a commodity, and widely available on any cloud. What is the most useful to companies, however, is AI that has your enterprise data as context, and more enterprises are on AWS than anywhere else. We will provide AI with a company’s data as context at scale and at low cost — both in terms of CapEx and OpEx — by primarily running inference on our own AI chips, not Nvidia chips.

    What is notable about both arguments — and again, this doesn’t mean they are wrong! — is how conveniently they align with how the companies operated in the previous era. Apple powered apps with Apple Silicon on the edge with an emphasis on privacy, and Amazon powered apps in the cloud with its own custom architecture focused first and foremost on low costs.

    The AI Paradigm

    The risk both companies are taking is the implicit assumption that AI is not a paradigm shift like mobile was. In Apple’s case, they assume that users want an iPhone first, and will ultimately be satisfied with good-enough local AI; in AWS’s case, they assume that AI is just another primitive like compute or storage that enterprises will tack onto their AWS bill. I wrote after last fall’s re:Invent:

    The emphasis on “choice” in the presentation, first in terms of regular AWS, and then later in terms of AI, is another way to say that the options are, in the end, mere commodities. Sure, the cutting edge for both inference and especially training will be Nvidia, and AWS will offer Nvidia instances (to the extent they fit in AWS’ network), but AWS’s bet is that a necessary component of generative AI being productized is that models fade in importance. Note this bit from Garman leading up to his Bedrock discussion:

    We talked about wanting this set of building blocks that builders could use to invent anything that they could imagine. We also talked about how many of the cases we walked through today, that we’ve redefined how people thought about these as applications change. Now people’s expectations are actually changing for applications again with generative AI, and increasingly my view is generative AI inference is going to be a core building block for every single application. In fact, I think generative AI actually has the potential to transform every single industry, every single company out there, every single workflow out there, every single user experience out there…

    This expansive view of generative AI’s importance — notice how Garman put it on the same level as the compute, storage, and database primitives — emphasizes the importance of it becoming a commodity, with commodity-like concerns about price, performance, and flexibility. In other words, exactly what AWS excels at. To put it another way, AWS’s bet is that AI will be important enough that it won’t, in the end, be special at all, which is very much Amazon’s sweet spot.

    Go back to that illustration from The End of the Beginning: Apple and Amazon are betting that AI is just another primitive in continuous computing that happens everywhere.

    A drawing of The Evolution of Computing

    The most optimistic AI scenarios, however, point to something new:

    A new paradigm of agents and augmentation may lie beyond the cloud and smartphones.

    A better word for “Anywhere” is probably autonomous, but I wanted to stick with the “Where” theme; what I’m talking about, however, is agents: AI doing work without any human involvement at all. The potential productivity gains for companies are obvious: there is a massive price umbrella for inference costs if the end result is that you don’t need to employ a human to do the same work. In this world what matters most is performance, not cost, which means that Amazon’s obsession with costs is missing the point; it’s also a world where the company’s lack of a competitive leading edge model makes it harder for them to compete, particularly when there is another company in the ecosystem — Google — that not only has its own custom chip strategy (TPUs), but also is integrating those chips with its competitive leading edge large language model (Gemini).

    Tim Cook, meanwhile, has talked for years now about his excitement about AR glasses, which fit with the idea of augmentation; Mark Gurman reported in Bloomberg earlier this year:

    Still, all of this is a stepping stone toward Cook’s grand vision, which hasn’t changed in a decade. He wants true augmented reality glasses — lightweight spectacles that a customer could wear all day. The AR element will overlay data and images onto real-world views. Cook has made this idea a top priority for the company and is hell-bent on creating an industry-leading product before Meta can. “Tim cares about nothing else,” says someone with knowledge of the matter. “It’s the only thing he’s really spending his time on from a product development standpoint.”

    Still, it will take many years for true AR glasses to be ready. A variety of technologies need to be perfected, including extraordinarily high-resolution displays, a high-performance chip and a tiny battery that could offer hours of power each day. Apple also needs to figure out applications that make such a device as compelling as the iPhone. And all this has to be available in large quantities at a price that won’t turn off consumers.

    What seems likely to me is that for this product to succeed, Apple will need to figure out generative AI as well; I posited last year that generative AI will undergird future user interfaces in The Gen AI Bridge to the Future. From a section recounting my experience with Meta’s Orion AR glasses:

    This, I think, is the future: the exact UI you need — and nothing more — exactly when you need it, and at no time else. This specific example was, of course, programmed deterministically, but you can imagine a future where the glasses are smart enough to generate UI on the fly based on the context of not just your request, but also your broader surroundings and state.

    This is where you start to see the bridge: what I am describing is an application of generative AI, specifically to on-demand UI interfaces. It’s also an application that you can imagine being useful on devices that already exist. A watch application, for example, would be much more usable if, instead of trying to navigate by touch like a small iPhone, it could simply show you the exact choices you need to make at a specific moment in time. Again, we get hints of that today through deterministic programming, but the ultimate application will be on-demand via generative AI.

    This may sound fanciful, but then again, I wrote in early 2022 that generative AI would be the key to making the metaverse viable:

    In the very long run this points to a metaverse vision that is much less deterministic than your typical video game, yet much richer than what is generated on social media. Imagine environments that are not drawn by artists but rather created by AI: this not only increases the possibilities, but crucially, decreases the costs.

    That may have also sounded fanciful at the time, but it’s already reality: just yesterday Google DeepMind announced Genie 3; from their blog post:

    Today we are announcing Genie 3, a general purpose world model that can generate an unprecedented diversity of interactive environments. Given a text prompt, Genie 3 can generate dynamic worlds that you can navigate in real time at 24 frames per second, retaining consistency for a few minutes at a resolution of 720p.

    […] Achieving a high degree of controllability and real-time interactivity in Genie 3 required significant technical breakthroughs. During the auto-regressive generation of each frame, the model has to take into account the previously generated trajectory that grows with time. For example, if the user is revisiting a location after a minute, the model has to refer back to the relevant information from a minute ago. To achieve real-time interactivity, this computation must happen multiple times per second in response to new user inputs as they arrive…

    Genie 3’s consistency is an emergent capability. Other methods such as NeRFs and Gaussian Splatting also allow consistent navigable 3D environments, but depend on the provision of an explicit 3D representation. By contrast, worlds generated by Genie 3 are far more dynamic and rich because they’re created frame by frame based on the world description and actions by the user.

    We are still far from the metaverse, to be clear, or on-demand interfaces in general, but it’s stunning how much closer we are than a mere three years ago; to that end, betting on current paradigms may make logical sense — particularly if you dominate the current paradigm — but things really are changing with stunning speed. Apple and Amazon’s risk may be much larger than either appreciate.

    Google Appreciation

    Genie 3 is, as I noted, from Google, and thinking about these paradigm shifts — first the shift to mobile, and now the ongoing one to AI — has made me much more appreciative and respectful of Google. I recounted above how the company did what was necessary — including overhauling Android to mimic iOS — to capture its share of the mobile paradigm; as we approach the three year anniversary of ChatGPT, it’s hard to not be impressed at how the company has gone all-in on relevancy with AI.

    This wasn’t a guarantee: two months after ChatGPT, in early 2023, I wrote AI and the Big Five, and expressed my concerns about the company’s potential disruption:

    That, though, ought only increase the concern for Google’s management that generative AI may, in the specific context of search, represent a disruptive innovation instead of a sustaining one. Disruptive innovation is, at least in the beginning, not as good as what already exists; that’s why it is easily dismissed by managers who can avoid thinking about the business model challenges by (correctly!) telling themselves that their current product is better. The problem, of course, is that the disruptive product gets better, even as the incumbent’s product becomes ever more bloated and hard to use — and that certainly sounds a lot like Google Search’s current trajectory.

    I’m not calling the top for Google; I did that previously and was hilariously wrong. Being wrong, though, is more often than not a matter of timing: yes, Google has its cloud and YouTube’s dominance only seems to be increasing, but the outline of Search’s peak seems clear even if it throws off cash and profits for years.

    Meanwhile, I wasn’t worried about Apple and Amazon at all: I saw AI as being a complement for Apple, and predicted that the company would invest heavily in local inference; when it came to Amazon I was concerned that they might suffer from not have an integrated approach a la Google, but predicted that AI would slot in cleanly to their existing cloud business. In other words, exactly what Apple and Amazon’s executives are banking on.

    I wonder, however, if there is a version of this analysis that, were it written in 2007, might have looked like this:

    Nokia will be fine; once they make a modern OS, their existing manufacturing and distribution advantages will carry the day. Microsoft, meanwhile, will mimic the iPhone UI just like they once did the Mac, and then leverage their app advantage to dominate the lower end of the market. It’s Google, which depends on people clicking on links on a big desktop screen, that is in danger.

    I don’t, with the benefit of having actually known myself in 2007, think that would have been my take (and, of course, much of the early years of Stratechery were spent arguing with those who held exactly those types of views). I was, however, a Google skeptic, and I’m humble about that. And, meanwhile, I have that 2023 Article, where, in retrospect, I was quite rooted in the existing paradigm — which favors Apple and Amazon — and skeptical of Google’s ability and willingness to adapt.

    Today I feel differently. To go back to the smartphone paradigm, the best way to have analyzed what would happen to the market would have been to assume that the winners of the previous paradigm would be fundamentally handicapped in the new one, not despite their previous success, but because of it. Nokia and Microsoft pursued the wrong strategies because they thought they had advantages that ultimately didn’t matter in the face of a new paradigm.

    If I take that same analytical approach to AI, and assume that the winners of the previous paradigm will be fundamentally handicapped in the new one, not despite their previous success, but because of it, then I ought to have been alarmed about Apple and Amazon’s prospects from the get-go. I’m not, for the record, ready to declare either of them doomed; I am, however, much more alert to the prospect of them making wrong choices for years, the consequences of which won’t be clear until it’s too late.

    And, by the same token, I’m much more appreciative of Google’s amorphous nature and seeming lack of strategy. That makes them hard to analyze — again, I’ve been honest for years about the challenges I find in understanding Mountain View — but the company successfully navigated one paradigm shift, and is doing much better than I originally expected with this one. Larry Page and Sergey Brin famously weren’t particularly interested in business or in running a company; they just wanted to do cool things with computers in a college-like environment like they had at Stanford. That the company, nearly thirty years later, is still doing cool things with computers in a college-like environment may be maddening to analysts like me who want clarity and efficiency; it also may be the key to not just surviving but winning across multiple paradigms.



    Get notified about new Articles