In this issue: - Routers, Apps, AGI—The more the default way to start solving a problem is to converse with an AI chatbot, the more they end up being a frontend to the entire economy. But a frontend is only as useful as the API powering it. Fortunately, OpenAI is building this out—and it's hard to see where that process will stop.
- Peripheral Bets—Of trends, magnitudes, and standard deviations.
- Fragmentation and Industry Maturity—The classic industry growth pattern of starting with vertical integration and moving to a mature, specialized business happens even in businesses that aren't legal.
- Deepfakes as Pull Quotes—Fake videos, real quotes.
- Benchmarks—Are you betting on supply and demand of the Red Metal, or the latest policy shifts from the Orange Man?
- Platforms—Lord, fill my platform with AI slop, but not yet.
Routers, Apps, AGIFor a while in the early 2000s, the story of economic growth was that you could go to a website, type some words into a text bar, and get ten blue links, a few dozen flights, a few hundred products, etc., which were probably roughly what you were looking for if the product, service, or piece of information you were looking for actually existed. In that model, search was a dynamic layer on top of a comparatively static real economy—your search for a flight wouldn’t cause any airlines to adjust their schedule, and asking for a Wikipedia article on some obscure thinker didn’t magic one into existence. (Though, over time, it got harder and harder to find public figures who didn’t have some kind of Wikipedia presence.) The software sector could sustain high growth because, while it was eating the world, there was a whole lot of non-software world to eat, and the hardware, telecom, and venture/growth equity industries supplied plenty of the appropriate enzymes to make that happen. A case where there’s massive fragmentation among customers and suppliers naturally creates demand for centralized platforms that can bring them together, and that’s exactly what the biggest online platforms do. They tended to get their initial wins from scooping up easily-accessible information: - A simple online bookstore is just an HTML frontend to a few wholesalers’ catalogues, though that wasn’t simple to make in 1994;
- Harvard had face books, but no consolidated Facebook, and since most students would be first- or second-degree connections right away, a single campus environment was a great way to figure out how an online social network for people who already knew each other should work;
- Music and video IP was concentrated in a small number of gatekeepers, so streaming operations that could work with a critical mass of them had enough catalog breadth to be the obvious winner. (Though the structure of those two industries, and different consumption patterns, meant that the two didn’t stay aggregated in the same way.)
Ben Thompson formalized this in Aggregation Theory, which became seminal to understanding how search and social, by owning/understanding consumer demand/intent, created (and captured) the plurality of value on the consumer internet. Two weeks ago, OpenAI started incorporating third-party apps directly into ChatGPT, and shipped an SDK to allow other software companies to include their apps. This is partly the next obvious product decision to make—the closer the customer gets to making a purchase before leaving a platform, the more of the contribution profit from that purchase the platform can capture—but it’s more interesting as a potential way to bring Aggregation Theory to B2B software, and beyond. This platform evolution is not just an incremental expansion to where aggregation theory applies, it’s aggregation theory taken to its limits, and eventually applied to the entire economy. It’s at least potentially the last big API, a layer of general-purpose economic middleware, and a means to translate cheap, on-demand intelligence to economic growth. In a business context, suppose you have some commonplace problem, like an annoying rise in customer acquisition costs. You might get routed to an off-the-shelf solution, like a software product; you might eventually get connected to a consultant who can take a look at your paid ad spending and figure out what’s not working, or even to a few people you might want to hire as CMO. If a search product is mostly traversing the static web, it has to make some assumptions about what kind of solution you’re looking for, and pick-push the meta-problem of knowing what to search back to the user or onto whoever wants to get found. (If you search for enough business software-related topics, you’ll see lots of ads from someone who isn’t selling that product, but who’s buying ads that take you to a landing page where they tell you why you need their alternative instead). But in a chatbot environment, those assumptions get relaxed—and if it’s also a chatbot that has memories of some of the data analysis you’ve done, some of the strategic questions you’ve brainstormed, etc., it’s actually in a pretty good position to judge what you need. (Things get trickier from a privacy perspective if at least some of those prospective CMOs are also ChatGPT users, and if it can help evaluate them for you. It’s not clear that there’s an elegant solution here, but the general story with privacy is that users are very tolerant of theirs getting violated if they don’t have to think about it and get benefits from it, too.) Taken to its logical conclusion, this mechanism allows OpenAI to be the intelligent switchboard that not only discovers/understands/captures workflow intent, but also routes it to the best end-to-end solution. This intelligent switchboard turns aggregated workflow intent into a dynamic economic inefficiency index that is programmatically extensible/legible to the growing universe of first party and third party, increasingly AI-powered applications purpose built to solve specific problems end-to-end. Crucially, this system will be maximally conducive to self-improvement via RL: instead of optimizing routing around lagging indicators—who has the largest marketing budget, best Gartner standing, subjective reputation among implementation consultants, executives, etc.—it optimizes around which tool was most effective at turning workflow intent/problems into solutions. Outcome signals drive better routing, which attracts better tools, which improves outcomes in a compounding loop. An interesting further analogy is to think about the router as formalizing a mixture of experts approach to business building where the experts are the universe of AI powered tools that solve specific problems with higher fidelity than a purely general approach (search + code gen/execution) would. In traditional mixture of experts, this allows for huge step changes in computational efficiency during training, and in this case, could lead to the same step changes in efficiency in business building. Progress in information technology can be modeled as a never ending march to solve ever more valuable matching problems. The share of valuable matching problems that have truly been solved by the CPUs, the PC, internet, traditional search, predictive ML (recommendations/ad targeting), and now LLMs/GenAI is extremely small. And that’s despite the fact that these technologies have created matching problems that are entirely native to their infrastructure; the question of which database to use for a project, or which video game to play next, is not necessarily answered entirely by computers. What’s interesting is that the function of an economy is also a never ending march to solve matching problems. Economics as a body of theory is the study of allocating scarce resources; economics as a practice is whatever can be done to improve the efficiency of that allocation. In this sense, the economy can be modeled as a massive, distributed information processing system that tries to most efficiently execute all the intermediate steps involved in these matching problems, and one element of productivity can be modeled as the real time measure of how well the economy is solving these matching problems. In other words, the economy is an emergent superintelligence that somewhat effectively (and for the past 300 years ever more effectively), identifies high value problems and matches those problems to the most-efficient, presently available solution. OpenAI, with app-use/the router has taken a major leap in compressing this existing economic superintelligence into a legible, queryable, steerable system that gets better over time, and eventually one where improvements to the underlying system approximate improvements in actual total factor productivity. They were not the first to attempt this compression. There have been areas where these matching problems have been (programmatically) solved, and in those areas immense amounts of value has been created and captured.The canonical examples are search and direct response ads matching consumer desires with (mostly goods) that fulfill those desires. Matching revealed consumer intent/desire to consumer goods that fulfill those intents/desires is the matching problem Meta, Google and Amazon solved with highest fidelity. Sam Altman, like any good consumer Internet entrepreneur, expressed distaste for ad-based business models early in his career and eventually concluded that they’re too lucrative and incentive-aligning to ignore. In his recent interview with Ben Thompson, he concedes that ads have made his life better: “First of all, on the Instagram ads point, that was actually the thing that made me think, okay, maybe ads don’t always suck. I love Instagram ads, they’ve added value to me, I found stuff I never would’ve found, I bought a bunch of stuff, I actively like Instagram ads.” What he didn’t explicitly say, is that those ads also generate trillions in value for the small, medium, and large businesses who benefit from the world-scale matching (some lossy, some with very high fidelity) of consumer intent to their products/services. And that ads give OpenAI a much more direct incentive to care about when people game the system, by, for example, buying sockpuppet Reddit accounts to post comments singing the praises of their products. For a subscription product, that’s a minor quality problem; for an ad-supported product, it’s direct competition. One issue for Meta is that the only signal it has for product quality is money spent, there is no information about how well the product actually solved someone’s problem, how it performs relative to other products, etc. They’re stuck measuring user churn and tracking reviews as a lossy, laggy indicator. Once you buy a T-shirt from an instagram ad, Meta doesn’t know if you actually wore it, wear it on a regular basis, or if it just sits in your closet. There’s no explicit signal for product performance, satisfaction, or utility. OpenAI’s router enables a new kind of accounting that’s outcome-based rather than transaction-based, where utility of digital products is very measurable. (And this will also give them a data incentive to particularly like to match people to humans rather than software service providers—they get more detailed feedback that way.) What Sam also didn’t say is that, for most of the last 25 years, intent that implied an intelligence- or service-based solution was matched by Google and Facebook with far lower fidelity—usually as a list of instructions for you to execute, or a directory of who might execute them. ChatGPT/chatbots to date have slowly increased the fidelity here, first by using a more complete understanding of intent to provide more comprehensive/high quality information and procedures, then, with the advent of reasoning and tool use, providing the primitives (code gen/interpreter, deep research/search both on public and private data) to actually execute those procedures. We know that solving these matching problems with even relatively lower fidelity creates, and allows companies to capture, massive value. If you don’t believe us, just take a look at the market caps of Google, Meta, Amazon and OpenAI to date or this recent economic impact study from Meta that approximates they’re responsible for $548B in annual economic activity. But this raises the question, how much value can be created and captured if this broader, b2b and beyond, flavor of matching problems are solved with very high fidelity? While there has been a lot of discussion about the app-use/router announcement and what it implies about OpenAI’s general product and platform strategy, no one (including Sam) has publicly or explicitly called it what it is: the meta-discovery that will allow OpenAI to solve, with very high-fidelity, one of the most valuable matching problems in human history. OpenAI will soon be able to match valuable business problems with solutions at much higher fidelity than search/social/current chatbots and at much lower cost than the current non-programmatic meta for business solution discovery and implementation (consultants, software sales cycles, tacit knowledge, human information and relationship networks, etc.) As with any other LLM trend, this is partly predicated on better models. GPT-5 is smart, but probably not smart enough to match the maddening contours of the typical consumer’s mind. But they’ll be collecting data on a large fraction of the purchase process, and there’s a good chance they’ll also see incremental updates as these purchases are consumed. If they ever decide to go deep into AI boyfriends and girlfriends, expect these to be available for free accounts, and expect your beloved to be deeply curious about whether switching toothpaste brands made a big difference in your life. This also shows that OpenAI doesn’t necessarily want to provide all these high-fidelity, end-to-end solutions themselves, or that they believe it will be possible for some super-intelligent model to vibe code them in real time, or that all traditional software is dead. In fact, the App-use and routing capability is a tacit admission that code-gen/interpreter and agentic search are far too elementary of primitives to solve complex business problems with high fidelity and efficiency. And also an admission that no one company or model can develop all the first-party applications to do so, similar to how Google realized that no one company could produce all the useful content on the internet or manually organize it into a directory that was truly comprehensive and maximally useful—and that being the company that organized everyone else’s information was a better business than generating their own. OpenAI is admitting here that their strategy isn’t primarily rooted in getting better and better at gathering the correct context and invoking the correct first-party tools to solve your problems end to end inside ChatGPT. Instead, they’re saying it's about building a machine that deeply understands your problem, and knows how to make that understanding maximally legible and usable by the highest-fidelity (increasingly third-party, increasingly AI-powered) end-to-end solutions. They are acknowledging that for AI systems to become truly bitter lesson-pilled, to generate enough data to improve, they need to take advantage of the inherent composability of software. That means more intent and outcome signals to train on, more counterparties to balance against one another, and more reasons for the user to lean on AI to keep track of everything. Taken together, you can look at LLM chatbots that deeply integrate service providers as the Manhattan Project of economics, a radical experiment in expending huge sums of capital to crush transaction costs and then capture some of the resulting upside. When they say they are building AGI (which Sam defines as AI that can automate most or even all economically valuable tasks), they don’t mean building one super intelligent model that suddenly qualifies; they mean building the coordination machine that systematically discovers and matches the economy’s highest value problems with the best, currently available solutions, ideally while continuing to improve itself. Their view is also that the solutions to which valuable problems get matched will increasingly be AI-native applications/businesses, partially powered by OpenAI’s API, but also by whatever other blends of cheap intelligence (open source, competing general-purpose models, specialized models, etc.) arise. Of course, at first, most of the problems and solutions will be white collar/digital in nature, but as the analog world gets increasingly eaten by low-cost electricity, transistors, and intelligence, this will mean more kinds of workflow intent and more kinds of high-fidelity solutions become legible to the router as a universe of composable, intelligent solutions; the higher the ratio of bits of data created to dollars of GDP, the greater software’s ability to reorganize the economy. This makes problem solving in the physical world as composable as the digital world. If the router can understand what a robot inside a warehouse is looking at, and what it’s trying to do, it can understand which tool that robot needs to fulfill its task. Does it need to call the tool to register the number of widgets in the bin? Does it need to call the tool that helps it move the widget from one section of the warehouse to another in some particularly high-dexterity way? Does it need to call the tool that helps it understand where in the warehouse this widget needs to go? Which tools are currently performing the best in these areas? etc. By taking the first step in solving this matching problem, OpenAI is in pole position to become the meta-layer over the entire AI-powered economy (AGI). It’s been quite clear for anyone paying attention that OpenAI’s internal bear/base/bull-case financial scenarios are: the end of the world, a multi-trillion dollar company, or the Singularity. This analysis is focused on that base case scenario, where OpenAI merely joins the pantheon of the most valuable enterprises in history. Companies of that size and impact have a serious responsibility. In 1927, the US economy went into a brief recession because Ford Motor Company shut down production of the Model T to retool for their next vehicle—one company was big enough to singlehandedly engineer a recession then, and it’s likely that there are individual companies that could do that now (though corporate governance and the scope of regulation have expanded since then, so it’s less likely that they would. They were such a large employer, and cars were such a significant portion of GDP, that this single company's pause caused sequential GDP decline. At that moment, looking at the data, you might have worried: "What if the economy is just becoming the car industry?" But that's not what happened. Today, cars are still a significant chunk of the economy, but a far larger portion of economic activity exists because people can drive to it. The general purpose technology enabled an explosion of complementary industries and activities. Cars were a massive share of the economy during the peak deployment phase, but the more they were used, and the more scale production drove down their price, the more the marginal dollar of capex was going towards taking advantage of how many people owned cars rather than making more cars. And that’s generally true. The more general purpose a technology is, the more its impact shows up in things other than that technology. The written word, the wheel, electrification, the internet—the percentage of GDP we spend directly on these is much smaller than the percentage of GDP that's dependent on them. They're complements to everything, which makes the rest of the economy bigger. Right now, US GDP growth appears to be AI and pretty much nothing else—estimates range up to 92% ($, FT) of US GDP growth coming from the AI buildout. This looks concerning if you extrapolate it. But if AI follows the historical pattern of general purpose technologies, this should be temporary. What we should expect: a period where an unsustainably high fraction of GDP is devoted to the buildout, followed by that fraction declining as AI becomes a low marginal cost way to access a new set of opportunities. Those opportunities—the businesses and activities enabled by cheap, abundant intelligence and efficient coordination—are where the long-term growth will come from. The app-use/router evolution is significant because it might be the maximally general and scalable way to increase economic value created by improvements in the underlying technology: it’s about making the entire economy of solutions more accessible and efficient. The power and compute going into training OpenAI’s models, running OpenAI’s models, etc. is likely to be translated into a substrate that will reduce the friction in every economic transaction going forward and serve as a discovery mechanism for future high value economic transactions. With app use and the router, OpenAI has finally made explicit that AGI will be about coordination and resource allocation. In this framework, we probably shouldn’t be thinking about AGI as something that will take all of our jobs. Instead, we should ask if it can get so good at understanding the economy that it will be able to route human talent, capital, and digital agents to their most productive uses. Sort of like what Elon Musk or other great allocators of capital and labor do at present. That could be a future where more people are empowered to do their best work and where the economy's matching problems get solved with unprecedented fidelity. If we want to understand how much value a company that solves this problem could create, just imagine if Elon were deployed across all areas of the economy. Marc Andreessen an informative riff on this. He concedes that even though Elon is a true anomaly, and that there are likely to be fewer than 10,000 people in the world with his blend of traits/abilities, we could think about progress in entrepreneurship and innovation as the dose of “milli-Elons” all founders can apply to their market. Elon is of course 1000 “milli-Elons”, but it should be the goal of founders’ to operate at whatever dose they are capable of handling, whether that’s 10 “milli-Elons” or 500. What makes Elon so special is that he can hold the real-time state of problems and related engineering details across 6+ companies, in his head, with enough detail that he can get up to speed fast on whichever company is struggling the most. This allows him to parachute into any one of his local economies, quickly identify areas of maximum local economic inefficiency and route the right blend of intelligence (in all its forms) and capital (in all its forms) to remove that bottleneck. In other words, he matches high value problems with high fidelity solutions continuously, and with incredible effectiveness, which is why the productivity growth at his companies is immense. He has an incredible command over the fundamental nature of economy building/business building. A proto-AGI, with its inherent ability to aggregate problems/workflow intent and route to solutions, gives founders and their teams the ability to access 100-500 more milli-Elons than otherwise possible. Depending on your company and role, the scale and nature of your local economy may differ, but the fundamental job of matching the highest value problems with high fidelity solutions remains the same. Therefore, if we have a machine that gives us leverage in this process and have the proclivities and energy to use it, we could have a path towards sustained economic growth that mirrors the conditions one of this year’s Nobel Prize winners, Joel Mokyr, identified as prerequisites for the Industrial Revolution. Mokyr's framework distinguishes between propositional knowledge (understanding why things work) and prescriptive knowledge (knowing how to do things), and argues that sustained growth requires a tight feedback loop between the two. Before the Industrial Revolution, craftsmen might’ve stumbled upon better techniques, but without understanding underlying principles, improvements couldn't be systematically replicated or extended. The Industrial Enlightenment changed this by creating institutions that connected theory and practice. For example, when glassmakers could talk to chemists and mechanics could access thermodynamics, prescriptive improvements fed back into propositional understanding, which enabled systematic improvements in technique. Meanwhile, cheaper and more precise equipment meant that theorists could do better experiments, keeping that feedback loop running. This self-reinforcing loop is what made growth sustained rather than episodic. AGI could be the meta-institution that does for the entire economy what the Industrial Enlightenment did for manufacturing. This leaves us with a good corollary of what the most productive companies of the future may look like: a team of mini-CEOs, a pod shop of expert-generalists, etc. These would be teams of people with exceptional judgement across domains who can most effectively leverage AGI to identify the highest value opportunities/bottlenecks, construct an optimal portfolio of solutions/bets, and capture the value from an ever evolving set of economic inefficiencies. Once that process is systematized, we may have finally reached ASI. Disclosure: long META, GOOGL, AMZN Diff JobsCompanies in the Diff network are actively looking for talent. See a sampling of current open roles below: - A hyper-growth startup that’s turning the fastest growing unicorns’ sales and marketing data into revenue (driven $XXXM incremental customer revenue the last year alone) is looking for a senior/staff-level software engineer with a track record of building large, performant distributed systems and owning customer delivery at high velocity. Experience with AI agents, orchestration frameworks, and contributing to open source AI a plus. (NYC)
- Well funded, Ex-Stripe founders are building the agentic back-office automation platform that turns business processes into self-directed, self-improving workflows which know when to ask humans for input. They are initially focused on making ERP workflows (invoice management, accounting, financial close, etc.) in the enterprise more accurate/complete and are looking for FDEs and Platform Engineers. If you enjoy working with the C-suite at some of the largest enterprises to drive operational efficiency with AI and have 3+ YOE as a SWE, this is for you. (Remote)
- A leading AI transformation & PE investment firm (think private equity meets Palantir) that’s been focused on investing in and transforming businesses with AI long before ChatGPT (100+ successful portfolio company AI transformations since 2019) is hiring Associates, VPs, and Principals to lead AI transformations at portfolio companies starting from investment underwriting through AI deployment. If you’re a generalist with deal/client-facing experience in top-tier consulting, product management, PE, IB, etc. and a technical degree (e.g., CS/EE/Engineering/Math) or comparable experience this is for you. (Remote)
- A transformative company that’s bringing AI-powered, personalized education to a billion+ students is looking for elite, AI-native generalists to build and scale the operational systems that will enable 100 schools next year and a 1000 schools the year after that. If you want to design and deploy AI-first operational systems that eliminate manual effort, compress complexity, and drive scalable execution, please reach out. Experience in product, operational, or commercially-oriented roles in the software industry preferred. (Remote)
- YC-backed founder building the travel-agent for frequent-flyers that actually works is looking for a senior engineer to join as CTO. If you have shipped real, working applications and are passionate about using LLMs to solve for the nuanced, idiosyncratic travel preferences that current search tools can't handle, please reach out. (SF)
Even if you don't see an exact match for your skills and interests right now, we're happy to talk early so we can let you know if a good opportunity comes up. If you’re at a company that's looking for talent, we should talk! Diff Jobs works with companies across fintech, hard tech, consumer software, enterprise software, and other areas—any company where finding unusually effective people is a top priority. Elsewhere
Peripheral Bets
When one part of the economy experiences outsized growth, it can often lead to disproportionate revenue surprises throughout its supply chain. Bloomberg covers former crypto mining plays that are valuable mostly because of their long-term power contracts, and the WSJ highlights money-losing and even pre-revenue bets on AI power demand ($, WSJ). Industries that have big swings in growth in percentage terms tend to trade at low multiples when earnings peak, because investors know that that's temporary. So the purely cyclical players don't benefit that much when a new secular trend boosts their revenue. The real impact is in cases where the growth is high in terms of the standard deviation of demand growth, as with electricity. In a case like that, the long-term fixed-price contracts get revalued, as do the speculative new sources of energy.
Fragmentation and Industry Maturity
The most successful companies in new industries are often vertically-integrated, because there are so many unique inputs and intermediate goods that they need. And in mature industries, there tends to be more specialization, as every organization focuses on its comparative advantage. One place this is apparently happening is in the illegal drug business ($, The Economist), where production and transportation are more fragmented than before, which makes it harder for enforcement to have an impact. Supply elasticity is an important concern for policy, because what's optimal to stop—and where in the supply chain it's best to do that—is a moving target.
Deepfakes as Pull Quotes
Republicans are using an AI-generated video of Chuck Schumer to attack him, and it's working because the video is a deepfake of a quote Schumer gave in print. This is a good example of the self-limiting nature of AI-generated propaganda: it's only going to have an effect if people believe it, but the believable stuff doesn't shift people's opinions all that much, because it already fits in with their worldview. In this case, the deepfake is closer to a formatting change than a fabrication. So in politics, generative AI does mean that whatever a politician does will be translated into whatever medium makes it most damaging, but that was the status quo to begin with.
Benchmarks
When you trade commodity futures, you're trading a somewhat stylized abstract representation of the commodity in question. But those trades can sometimes be settled by the delivery of physical products, so at some point the trade becomes concrete again. This concern sometimes shows up when the metal backing contracts turns out to be fake, but it alos matters when different markets reference different jurisdictions: the London Metals Exchange has been taking share back from the US's Comex because metals prices at the latter are affected by tariffs ($, FT). There's still value in having a market like that; an American copper consumer or producer wants to hedge US prices, because that's what they'll pay or receive. But the US consumes ~8% of global copper production, so for the vast majority of traders, this is just noise. The financial product that best isolates exposure to the exact variable traders want to bet on is the one that gets the liquidity, whether that's in ETFs, in equities that pivot to take advantage of some theme, or in commodities.
Meta has banned third-party AI companies from using WhatsApp as a distribution tool, though it still allows AI-powered chat within the context of a separate product or service. So you can talk to an AI customer service agent through WhatsApp, but if you're going to use WhatsApp to get AI-powered therapy or to write you a quick React app, you're going to use Meta's.
From one perspective, this doesn't align with Meta's incentive: in comparison to all of the other companies building AI products, Meta is best positioned to capture upside from the growth of LLMs and generative AI even if their in-house models don't win. They're still where content will be shared, and if someone else makes the model that Meta uses to generate more variants of an ad in order to find the optimal copy and creative, Meta still gets upside. But the reason Meta is so well-positioned there is that they have a good sense of who's a real user and who's a bot; if more of their users are automated, they risk losing some of that advantage, and it makes sense to block them early.
|