Select Page

Global Advisors | Quantified Strategy Consulting

AI
Quote: Nate B Jones

Quote: Nate B Jones

“Anthropic shipping ‘Co-Work’ as a full product feature. It was built in 10 days with just four people. It was written entirely in Claude Code. And Claude Code, mind you, is an entire product that is less than a year old… The Anthropic team is evolving as they go.” – Nate B Jones – AI News & Strategy Daily

Context of the Quote

On 15 January 2026, Nate B Jones, in his AI News & Strategy Daily update, highlighted Anthropic’s remarkable achievement in shipping ‘Co-Work’ (also styled as Cowork), a groundbreaking AI feature. This quote captures the essence of Anthropic’s rapid execution: developing a production-ready tool in just 10 days using a team of four, with all code generated by their own AI system, Claude Code. Jones emphasises the meta-innovation – Claude Code itself, launched less than a year prior, enabling this feat – signalling how Anthropic is iteratively advancing AI capabilities in real-time.1,5

Who is Nate B Jones?

Nate B Jones is a prominent voice in AI strategy and news aggregation, curating daily insights via his AI News & Strategy Daily platform. His commentary distils complex developments into actionable intelligence for executives, developers, and strategists. Jones focuses on execution speed, product strategy, and the competitive dynamics of AI firms, often drawing from primary sources like announcements, demos, and insider accounts. His analysis in this instance underscores Anthropic’s edge in ‘vibe coding’ – prompt-driven development – positioning it as a model for AI-native organisations.1,7

Backstory of Anthropic’s Cowork

Anthropic unveiled Cowork on 12 January 2026 as a research preview for Claude Max subscribers on macOS. Unlike traditional chatbots, Cowork acts as an autonomous ‘colleague’, accessing designated local folders to read, edit, create, and organise files without constant supervision. Users delegate tasks – such as sorting downloads, extracting expenses from screenshots into spreadsheets, summarising notes, or drafting reports – and approve key actions via prompts. This local-first approach contrasts with cloud-centric AI, restoring agency to personal devices while prioritising user oversight to mitigate risks like unintended deletions or prompt injections.1,2,3,4,6

The tool emerged from user experiments with Claude Code, Anthropic’s AI coding agent popular among developers. Observing non-technical users repurposing it for office tasks, Anthropic abstracted these capabilities into Cowork, inheriting Claude Code’s robust architecture for reliable, agentic behaviour. Built entirely with Claude Code in 10 days by four engineers, it exemplifies ‘AI building AI’, compressing development timelines and widening the gap between AI-leveraging firms and others.1,3,5

Significance in AI Evolution

Cowork marks a shift from conversational AI to agentic systems that act on the world, handling mundane work asynchronously. It challenges enterprise tools like Microsoft’s Copilot by offering proven developer-grade autonomy to non-coders, potentially redefining productivity. Critics note risks of ‘workslop’ – error-prone outputs requiring fixes – but Anthropic counters with transparency, trust-building safeguards, and architecture validated in production coding.2,3,5,6

Leading Theorists and Concepts Behind Agentic AI

  • Boris Cherny: Leader of Claude Code at Anthropic, Cherny coined ‘vibe coding’ – an AI paradigm where high-level prompts guide software creation, minimising manual code. His X announcement confirmed Cowork’s components were fully AI-generated, embodying this hands-off ethos.1
  • Dario Amodei: Anthropic CEO and ex-OpenAI executive, Amodei champions scalable oversight and reliable AI agents. His vision drives Cowork’s supervisor model, ensuring human control amid growing autonomy.3,6
  • Yohei Nakajima: Creator of BabyAGI (2023), an early autonomous agent framework chaining tasks via LLM planning. Cowork echoes this by autonomously strategising and executing multi-step workflows.2
  • Andrew Ng: AI pioneer advocating ‘agentic workflows’ where AI handles routine tasks, freeing humans for oversight. Ng’s predictions align with Cowork’s file manipulation and task queuing, forecasting quieter, faster work rhythms.2,5
  • Lil’ Log (Lilian Weng): OpenAI’s applied AI head, Weng theorises hierarchical agent architectures for complex execution. Cowork’s lineage from Claude Code reflects this, prioritising trust over raw intelligence as the new bottleneck.5

These thinkers converge on agentic AI: systems that plan, act, and adapt with minimal intervention, propelled by models like Claude. Anthropic’s sprint validates their theories, proving AI can ship AI at unprecedented speed.

References

1. https://www.axios.com/2026/01/13/anthropic-claude-code-cowork-vibe-coding

2. https://www.techradar.com/ai-platforms-assistants/claudes-latest-upgrade-is-the-ai-breakthrough-ive-been-waiting-for-5-ways-cowork-could-be-the-biggest-ai-innovation-of-2026

3. https://www.axios.com/2026/01/12/ai-anthropic-claude-jobs

4. https://www.vice.com/en/article/anthropic-introduces-claude-cowork/

5. https://karozieminski.substack.com/p/claude-cowork-anthropic-product-deep-dive

6. https://fortune.com/2026/01/13/anthropic-claude-cowork-ai-agent-file-managing-threaten-startups/

7. https://www.youtube.com/watch?v=SpqqWaDZ3ys

read more
Quote: Demis Hassabis – DeepMind co-founder, CEO

Quote: Demis Hassabis – DeepMind co-founder, CEO

“I think [AI is] going to be like the industrial revolution, but maybe 10 times bigger, 10 times faster. So it’s an incredible amount of transformation, but also disruption that’s going to happen.” – Demis Hassabis – DeepMind co-founder, CEO

Demis Hassabis and the Quote

This striking prediction comes from Demis Hassabis, co-founder and CEO of Google DeepMind. Spoken on The Tech Download (CNBC Original podcast) on 16 January 2026, the quote encapsulates Hassabis’s view of artificial intelligence (AI) as a force dwarfing historical upheavals. He describes AI not merely as an evolution but as a catalyst for radical abundance, potentially leading to prosperity if managed equitably, while acknowledging inevitable job disruptions akin to – yet far exceeding – those of past revolutions.1,2

Backstory of Demis Hassabis

Born in 1976 in London to a Greek Cypriot father and Chinese Singaporean mother, Hassabis displayed prodigious talent early. At age 13, he won a British Tetris championship and published his first computer program in a magazine. By 17, he was the world’s second-highest-ranked chess player for his age group, balancing academics with competitive gaming.1

Hassabis entered the games industry as a teenager, co-designing the 1994 hit Theme Park at Bullfrog Productions and working with Peter Molyneux at Lionhead Studios on titles like Black & White. This foundation in complex simulations honed his skills in modelling human-like behaviours, which later informed his AI pursuits.1

In 2010, aged 34, he co-founded DeepMind with Mustafa Suleyman and Shane Legg, driven by a mission to ‘solve intelligence’ and advance science. Google acquired DeepMind for $400 million in 2014, propelling breakthroughs like AlphaGo (2016), which defeated world Go champion Lee Sedol, and AlphaFold (2020), revolutionising protein structure prediction.1,2

Today, as CEO of Google DeepMind, Hassabis leads efforts towards artificial general intelligence (AGI) – AI matching or surpassing human cognition across domains. He predicts AGI by 2030, describing himself as a ‘cautious optimist’ who believes humanity’s adaptability will navigate the changes.1,3,5

Context of the Quote

Hassabis’s statement reflects ongoing discussions on AI’s societal impact. He envisions AGI ushering in changes ’10 times bigger than the Industrial Revolution, and maybe 10 times faster,’ with productivity gains enabling ‘radical abundance’ – an era where scarcity ends, fostering interstellar exploration if wealth is distributed fairly.1,2

Yet, he concedes risks: job losses mirror the Industrial Revolution’s upheavals, which brought prosperity unevenly. Hassabis urges preparation, recommending STEM studies and experimentation with AI tools to create ‘very valuable jobs’ for the technically savvy. He stresses political solutions for equitable distribution, warning against zero-sum outcomes.1,3,5

Leading Theorists on AI and Transformative Technologies

Hassabis builds on foundational thinkers in AI and technological disruption:

  • Alan Turing (1912-1954): ‘Father of computer science,’ proposed the Turing Test (1950) for machine intelligence, laying theoretical groundwork for AGI.2
  • John McCarthy (1927-2011): Coined ‘artificial intelligence’ in 1956 at the Dartmouth Conference, pioneering AI as a field.2
  • Ray Kurzweil: Futurist predicting the ‘singularity’ – AI surpassing human intelligence by 2045 – influencing DeepMind’s ambitious timelines.1
  • Nick Bostrom: Philosopher warning of superintelligence risks in Superintelligence (2014), echoed in Hassabis’s cautious optimism.1
  • Shane Legg: DeepMind co-founder and chief AGI scientist, formalised AGI mathematically, emphasising safe development.2

These theorists frame AI as humanity’s greatest challenge and opportunity, aligning with Hassabis’s vision of exponential transformation.1,2

 

References

1. https://www.pcgamer.com/software/ai/deepmind-ceo-makes-big-brain-claims-saying-agi-could-be-here-in-the-next-five-to-10-years-and-that-humanity-will-see-a-change-10-times-bigger-than-the-industrial-revolution-and-maybe-10-times-faster/

2. https://www.antoinebuteau.com/lessons-from-demis-hassabis/

3. https://www.businessinsider.com/demis-hassabis-google-deemind-study-future-jobs-ai-2025-6

4. https://www.youtube.com/watch?v=l_vXXgXwoh0

5. https://economictimes.com/tech/artificial-intelligence/ai-will-create-very-valuable-jobs-but-study-stem-googles-demis-hassabis/articleshow/121592354.cms

 

read more
Quote: Nate B Jones – AI News & Strategy Daily

Quote: Nate B Jones – AI News & Strategy Daily

“The one constant right now is chaos. I hear it over and over again from folks: the rate of change, the sheer unpredictability of AI – it’s very difficult to tell what’s up and what’s down.” – Nate B Jones – AI News & Strategy Daily

Context of the Quote

This quote captures the essence of the AI landscape in early 2026, where rapid advancements and unpredictability dominate discussions among professionals. Spoken by Nate B. Jones during his AI News & Strategy Daily segment on 15 January 2026, it reflects feedback from countless individuals grappling with AI’s breakneck pace. Jones highlights how the constant flux – from model breakthroughs to shifting business applications – leaves even experts disoriented, making strategic planning a challenge.1,5

Backstory on Nate B. Jones

Nate B. Jones is a leading voice in practical AI implementation, known for his no-nonsense analysis that cuts through hype. Through his personal site natebjones.com, he delivers weekly deep dives into what truly works in AI, offering actionable frameworks for businesses and individuals. His Substack newsletter, including pieces like ‘2026 Sneak Peek: The First Job-by-Job Guide to AI Evolution’, has become essential reading for those navigating AI-driven disruption.2,3

Jones has personally advised hundreds of professionals on pivoting careers amid AI’s rise. He emphasises execution over mere tooling, stressing accountability, human-AI boundaries, and risk management. In videos such as ‘The AI Moments That Shaped 2025 and Predictions for 2026’, he recaps key events like model wars, Sora’s impact, copyright battles, and surging compute costs, positioning himself as a guide for the ‘frontier’ era of AI.1,4

His content, including AI News & Strategy Daily, focuses on real-world strategy: from compressing research timelines to building secure AI interfaces. Jones warns of a ‘compounding gap’ between the prepared and unprepared, urging a mindset shift for roles in programme management, UX design, QA, and risk assessment.2,5

Leading Theorists on AI Chaos and Unpredictability

The theme of chaos in AI echoes longstanding theories from pioneers who foresaw technology’s disruptive potential.

  • Ray Kurzweil: Futurist and Google director of engineering, Kurzweil popularised the ‘Law of Accelerating Returns’, predicting exponential tech growth leading to singularity by 2045. His books like The Singularity Is Near (2005) describe how AI’s unpredictability stems from recursive self-improvement, mirroring Jones’s observations of model saturation and frontier shifts.
  • Nick Bostrom: Oxford philosopher and author of Superintelligence (2014), Bostrom theorises AI’s ‘intelligence explosion’ – a feedback loop where smarter machines design even smarter ones, creating uncontrollable change. He warns of alignment challenges, akin to the ‘trust deficit’ and human-AI boundaries Jones addresses.2
  • Sam Altman: OpenAI CEO, whom Jones quotes on chatbot saturation. Altman’s views on AI frontiers emphasise moving beyond chat interfaces to agents and capabilities that amplify unpredictability, as seen in 2025’s model evolutions.1
  • Stuart Russell: Co-author of Artificial Intelligence: A Modern Approach, Russell advocates ‘provably beneficial AI’ to tame chaos. His work on value alignment addresses the execution speed and risk areas Jones flags, like bias management and compute explosions.2

These theorists provide the intellectual foundation for understanding AI’s turmoil: exponential progress breeds chaos, demanding strategic adaptation. Jones builds on this by offering tactical insights for 2026, from accountability frameworks to jailbreaking new intelligence surfaces.1,2,3

References

1. https://www.youtube.com/watch?v=YBLUf1yYjGA

2. https://natesnewsletter.substack.com/p/2026-sneak-peek-the-first-job-by-9ac

3. https://www.natebjones.com

4. https://www.youtube.com/watch?v=fbEiYRogYCk

5. https://www.youtube.com/watch?v=pOb0pjXpn6Q

6. https://www.youtube.com/watch?v=ftHsQvdTUww

read more
Quote: Jack Clark – Import AI

Quote: Jack Clark – Import AI

“Since 2020, we have seen a 600 000x increase in the computational scale of decentralized training projects, for an implied growth rate of about 20x/year.” – Jack Clark – Import AI

Jack Clark on Exponential Growth in Decentralized AI Training

The Quote and Its Context

Jack Clark’s statement about the 600,000x increase in computational scale for decentralized training projects over approximately five years (2020-2025) represents a striking observation about the democratization of frontier AI development.1,2,3,4 This 20x annual growth rate reflects one of the most significant shifts in the technological and political economy of artificial intelligence: the transition from centralized, proprietary training architectures controlled by a handful of well-capitalized labs toward distributed, federated approaches that enable loosely coordinated collectives to pool computational resources globally.

Jack Clark: Architect of AI Governance Thinking

Jack Clark is the Head of Policy at Anthropic and one of the most influential voices shaping how we think about AI development, governance, and the distribution of technological power.1 His trajectory uniquely positions him to observe this transformation. Clark co-authored the original GPT-2 paper at OpenAI in 2019, a moment he now reflects on as pivotal—not merely for the model’s capabilities, but for what it revealed about scaling laws: the discovery that larger models trained on more data would exhibit predictably superior performance across diverse tasks, even without task-specific optimization.1

This insight proved prophetic. Clark recognized that GPT-2 was “a sketch of the future”—a partial glimpse of what would emerge through scaling. The paper’s modest performance advances on seven of eight tested benchmarks, achieved without narrow task optimization, suggested something fundamental about how neural networks could be made more generally capable.1 What followed validated his foresight: GPT-3, instruction-tuned variants, ChatGPT, Claude, and the subsequent explosion of large language models all emerged from the scaling principles Clark and colleagues had identified.

However, Clark’s thinking has evolved substantially since those early days. Reflecting in 2024, five years after GPT-2’s release, he acknowledged that while his team had anticipated many malicious uses of advanced language models, they failed to predict the most disruptive actual impact: the generation of low-grade synthetic content driven by economic incentives rather than malicious intent.1 This humility about the limits of foresight informs his current policy positions.

The Political Economy of Decentralized Training

Clark’s observation about the 600,000x scaling in decentralized training projects is not merely a technical metric—it is a statement about power distribution. Currently, the frontier of AI capability depends on the ability to concentrate vast amounts of computational resources in physically centralized clusters. Companies like Anthropic, OpenAI, and hyperscalers like Google and Meta control this concentrated compute, which has enabled governments and policymakers to theoretically monitor and regulate AI development through chokepoints: controlling access to advanced semiconductors, tracking large training clusters, and licensing centralized development entities.3,4

Decentralized training disrupts this assumption entirely. If computational resources can be pooled across hundreds of loosely federated organizations and individuals globally—each contributing smaller clusters of GPUs or other accelerators—then the frontier of AI capability becomes distributed across many actors rather than concentrated in a few.3,4 This changes everything about AI policy, which has largely been built on the premise of controllable centralization.

Recent proof-of-concepts underscore this trajectory:

  • Prime Intellect’s INTELLECT-1 (10 billion parameters) demonstrated that decentralized training at scale was technically feasible, a threshold achievement because it showed loosely coordinated collectives could match capabilities that previously required single-company efforts.3,9

  • INTELLECT-2 (32 billion parameters) followed, designed to compete with modern reasoning models through distributed training, suggesting that decentralized approaches were not merely proof-of-concept but could produce competitive frontier-grade systems.4

  • DiLoCoX, an advancement on DeepMind’s DiLoCo technology, demonstrated a 357x speedup in distributed training while achieving model convergence across decentralized clusters with minimal network bandwidth (1Gbps)—a crucial breakthrough because communication overhead had previously been the limiting factor in distributed training.2

The implied growth rate of 20x annually suggests an acceleration curve where technical barriers to decentralized training are falling faster than regulatory frameworks or policy interventions can adapt.

Leading Theorists and Intellectual Lineages

Scaling Laws and the Foundations

The intellectual foundation for understanding exponential growth in AI capabilities rests on the work of researchers who formalized scaling laws. While Clark and colleagues at OpenAI contributed to this work through GPT-2 and subsequent research, the broader field—including contributions from Jared Kaplan, Dario Amodei, and others at Anthropic—established that model performance scales predictably with increases in parameters, data, and compute.1 These scaling laws create the mathematical logic that enables decentralized systems to be competitive: a 32-billion-parameter model trained via distributed methods can approach the capabilities of centralized training at similar scales.

Political Economy and Technological Governance

Clark’s thinking is situated within broader intellectual traditions examining how technology distributes power. His emphasis on the “political economy” of AI reflects influence from scholars and policymakers concerned with how technological architectures embed power relationships. The notion that decentralized training redistributes who can develop frontier AI systems draws on longstanding traditions in technology policy examining how architectural choices (centralized vs. distributed systems) have political consequences.

His advocacy for polycentric governance—distributing decision-making about AI behavior across multiple scales from individuals to platforms to regulatory bodies—reflects engagement with governance theory emphasizing that monocentric control is often less resilient and responsive than systems with distributed decision-making authority.5

The “Regulatory Markets” Framework

Clark has articulated the need for governments to systematically monitor the societal impact and diffusion of AI technologies, a position he advanced through the concept of “Regulatory Markets”—market-driven mechanisms for monitoring AI systems. This framework acknowledges that traditional command-and-control regulation may be poorly suited to rapidly evolving technological domains and that measurement and transparency might be more foundational than licensing or restriction.1 This connects to broader work in regulatory innovation and adaptive governance.

The Implications of Exponential Decentralization

The 600,000x growth over five years, if sustained or accelerated, implies several transformative consequences:

On AI Policy: Traditional approaches to AI governance that assume centralized training clusters and a small number of frontier labs become obsolete. Export controls on advanced semiconductors, for instance, become less effective if 100 organizations in 50 countries can collectively train competitive models using previous-generation chips.3,4

On Open-Source Development: The growth depends crucially on the availability of open-weight models (like Meta’s LLaMA or DeepSeek) and accessible software stacks (like Prime.cpp) that enable distributed inference and fine-tuning.4 The democratization of capability is inseparable from the proliferation of open-source infrastructure.

On Sovereignty and Concentration: Clark frames this as essential for “sovereign AI”—the ability for nations, organizations, and individuals to develop and deploy capable AI systems without dependence on centralized providers. However, this same decentralization could enable the rapid proliferation of systems with limited safety testing or alignment work.4

On Clark’s Own Policy Evolution: Notably, Clark has found himself increasingly at odds with AI safety and policy positions he previously held or was associated with. He expresses skepticism toward licensing regimes for AI development, restrictions on open-source model deployment, and calls for worldwide development pauses—positions that, he argues, would create concentrated power in the present to prevent speculative future risks.1 Instead, he remains confident in the value of systematic societal impact monitoring and measurement, which he has championed through his work at Anthropic and in policy forums like the Bletchley and Seoul AI safety summits.1

The Unresolved Tension

The exponential growth in decentralized training capacity creates a central tension in AI governance: it democratizes access to frontier capabilities but potentially distributes both beneficial and harmful applications more widely. Clark’s quote and his broader work reflect an intellectual reckoning with this tension—recognizing that attempts to maintain centralized control through policy and export restrictions may be both technically infeasible and politically counterproductive, yet that some form of measurement and transparency remains essential for democratic societies to understand and respond to AI’s societal impacts.

References

1. https://jack-clark.net/2024/06/03/import-ai-375-gpt-2-five-years-later-decentralized-training-new-ways-of-thinking-about-consciousness-and-ai/

2. https://jack-clark.net/2025/06/30/import-ai-418-100b-distributed-training-run-decentralized-robots-ai-myths/

3. https://jack-clark.net/2024/10/14/import-ai-387-overfitting-vs-reasoning-distributed-training-runs-and-facebooks-new-video-models/

4. https://jack-clark.net/2025/04/21/import-ai-409-huawei-trains-a-model-on-8000-ascend-chips-32b-decentralized-training-run-and-the-era-of-experience-and-superintelligence/

5. https://importai.substack.com/p/import-ai-413-40b-distributed-training

6. https://www.youtube.com/watch?v=uRXrP_nfTSI

7. https://importai.substack.com/p/import-ai-375-gpt-2-five-years-later/comments

8. https://jack-clark.net

9. https://jack-clark.net/2024/12/03/import-ai-393-10b-distributed-training-run-china-vs-the-chip-embargo-and-moral-hazards-of-ai-development/

10. https://www.lesswrong.com/posts/iFrefmWAct3wYG7vQ/ai-labs-statements-on-governance

read more
Quote: John Furner – President, CEO Walmart US

Quote: John Furner – President, CEO Walmart US

“The transition from traditional web or app search to agent-led commerce represents the next great evolution in retail. We aren’t just watching the shift, we are driving it.” – John Furner – President, CEO Walmart US

When John Furner speaks about the shift from traditional web or app search to agent-led commerce, he is putting words to a structural change that has been building at the intersection of artificial intelligence, retail strategy and consumer behaviour for more than two decades. His quote does not describe a marginal optimisation of online shopping; it points to a reconfiguration of how demand is discovered, shaped and fulfilled in the digital economy.

John Furner: An operator at the centre of AI-led retail

John Furner built his leadership reputation inside one of the most operationally demanding businesses in the world. Before being named President and CEO of Walmart U.S., and then incoming President and CEO of Walmart Inc., he held a series of roles that grounded him in the realities of store operations, merchandising and labour-intensive retail at scale.1,4 That background matters to the way he talks about AI.

Unlike many technology narratives that begin in the lab, Walmart’s AI story has been forged in distribution centres, supercentres and neighbourhood markets. Under Doug McMillon, and increasingly under Furner, Walmart framed AI not as a side project but as a new backbone for the business.1 Analysts note that as Furner steps into the global CEO role, the board describes the next chapter as one “fueled by innovation and AI”.1 His quote about agent-led commerce sits squarely in that strategic context.

Furner has consistently emphasised pragmatic, measurable outcomes from technology adoption: better inventory accuracy, improved shelf availability, faster fulfilment and fewer customer headaches.1,4 He has also been explicit that every job in the company will change in some way under AI – from collecting trolleys in car parks to technology development and leadership roles.4 In other words, for Furner, agent-led commerce is not simply a new consumer interface; it is a catalyst for rethinking work, operations and value creation across the retail stack.

The specific context of the quote: Walmart, Google and Gemini

The quote originates in the announcement of a partnership between Walmart and Google to bring Walmart and Sam’s Club product discovery directly into Google’s Gemini AI environment.2,3,5 Rather than treating AI search as an external channel to be optimised, the collaboration embeds Walmart’s assortment, pricing and fulfilment options into an intelligent agent that can converse with customers inside Gemini.

In this setting, Furner’s words perform several functions:

  • They frame the shift from keyword-driven search (type an item, browse lists) to goal- or task-based interaction (“help me plan a camping trip”), where an agent orchestrates the entire shopping journey.2,3
  • They signal that Walmart is not content to be a passive catalogue inside someone else’s interface, but intends to shape the emerging standards for “agentic commerce” – an approach where software agents work on behalf of customers to plan, select and purchase.2,3,4
  • They reassure investors and partners that the company sees AI as a core strategic layer, not as an optional experiment or promotional gimmick.1,4,6

The Walmart – Google experience is designed to allow a shopper to ask broad, life-context questions – for example, how to prepare for spring camping – and receive curated product bundles drawn from Walmart and Sam’s Club inventory, updated dynamically as the conversation unfolds.2,3 The system does not simply return search results; it proposes solutions and refines them interactively. The agent becomes a kind of digital retail concierge.

Technically, this is underpinned by the pairing of Gemini’s foundation models with Walmart’s internal data on assortment, pricing, local availability and fulfilment options.3 Strategically, it positions Walmart to participate in – and influence – the universal protocols that might govern how agents transact across merchants, platforms and services in the coming decade.

From web search to agent-led commerce: why this is a step-change

To understand why Furner describes this as “the next great evolution in retail”, it is useful to place agent-led commerce in a longer history of digital retail evolution.

1. Catalogue search and the era of the query box

The first wave of e-commerce was built around catalogue search: customers navigated static product hierarchies or typed keywords into a search box. Relevance was determined by text matching and basic filters. Power resided in whoever controlled the dominant search interface or marketplace.

This model mapped well onto traditional retail metaphors – aisles, departments, categories – and it assumed that the customer knew roughly what they were looking for. Retailers competed on breadth of assortment, price transparency, delivery speed and user interface design.

2. Personalisation and recommendation

The second wave saw retailers deploy recommendation engines, collaborative filtering and behavioural targeting to personalise product suggestions. Here, algorithmic theories drawn from machine learning and statistics began to shape retail experiences, but the core unit remained the search query or product page.

Recommendations were adaptively presented around known products and purchase history, nudging customers to complementary or higher-margin items. Many of the leading ideas came from research in recommender systems, one of the most commercially influential branches of applied machine learning.

3. Conversational interfaces and agentic commerce

Agent-led commerce represents a third wave. Instead of asking customers to break down their needs into discrete product searches, it allows them to:

  • Express goals (“host a birthday party for ten-year-olds”), constraints (“under £100, dietary restrictions, limited time”) and context (“small flat, no oven”).
  • Delegate the planning and selection process to an AI agent that operates across categories, channels and services.
  • Iterate interactively, with the agent updating recommendations and baskets as the conversation evolves.

In this model, the agent becomes a co-pilot for both discovery and decision-making. It can optimise not only for price and relevance, but also for timing, delivery logistics, dietary requirements, compatibility across items and even sustainability preferences, depending on the data and constraints it is given. The underlying technologies draw on advances in large language models, planning algorithms and multi-agent coordination.

For retailers, the shift is profound:

  • It moves the locus of competition from web page design and keyword bidding to who supplies the most capable and trustworthy agents.
  • It elevates operational capabilities – inventory accuracy, fulfilment reliability, returns processing – because an agent that cannot deliver on its promises will quickly lose trust.
  • It opens the door to autonomous or semi-autonomous shopping flows, such as automatic replenishment, anticipatory shipping or continuous cart management, where the agent monitors needs and executes under defined guardrails.

Furner’s assertion that Walmart is “driving” the shift needs to be understood against this backdrop. Internally, Walmart has already invested in a family of “super agents” for shoppers, associates, partners and developers, including Sparky (customer assistant), My Assistant (associate productivity), Marty (partner and advertising support) and WIBEY (developer tooling).1,4 Externally, initiatives like integrating with ChatGPT for “instant checkout” and partnering with Google on Gemini experiences demonstrate a strategy of meeting customers inside the agents they already use.1,3,4

Agent-led commerce inside Walmart: from vision to practice

Agent-led commerce is not just a phrase in a press release for Walmart. The company has been progressively building the capabilities required to make it a practical reality.

AI-native shopping journeys

Walmart has rolled out AI-powered search experiences that allow customers to describe occasions or problems rather than individual items – for example, planning a party or organising a kitchen.1 The system then infers needs across multiple categories and pre-populates baskets or recommendations accordingly.

At the same time, the company has been piloting “replenishment” features that create suggested baskets based on past purchases, letting customers approve, modify or decline the auto-generated order.1 This is an early expression of agentic behaviour: the system anticipates needs and does the heavy lifting of basket formation.

Super agents as an organisational pattern

Internally, Walmart has articulated a vision of multiple domain-specific “super agents” that share core capabilities but specialise in particular user groups.1,4

  • Sparky supports customers, operating as a front-end conversational assistant for shopping journeys.
  • My Assistant helps associates draft documents, summarise information and interact with data, freeing them from repetitive tasks.1,4
  • Marty works with partners and increasingly underpins the advertising business, helping brands navigate Walmart’s ecosystem.4
  • WIBEY accelerates developer productivity, contributing to the internal fabric of AI tooling.4

Additionally, Walmart has built a generative AI assistant called Wally for merchandising tasks, using AI to support complex assortment, pricing and space decisions.4

Operational AI as the foundation

Critically, Walmart has recognised that agent-led commerce cannot function if the operational substrate is weak. AI agents that promise two-hour delivery on items that are out of stock will immediately erode trust. As a result, the company has deployed AI and automation deep into its supply chain and fulfilment network.1,4

This includes large-scale investment in warehouse automation (for example, through partnerships with Symbotic), sensor-based tracking to improve inventory accuracy, and forecasting models that help move products closer to expected demand.1 The philosophy is that data quality is strategy: without reliable, granular data about where products are and how they move, agentic experiences will fail at the last mile.

The intellectual backstory: the theorists behind agents, recommendations and AI commerce

While Walmart and Google are prominent practitioners, the transition Furner describes rests on decades of work by researchers and theorists in several overlapping fields: information retrieval, recommender systems, artificial intelligence agents, behavioural economics and commerce design. A brief backstory of these fields helps illuminate what is now converging under the label “agent-led commerce”.

Information retrieval and the search paradigm

The idea of representing information needs through queries and ranking results based on relevance traces back to mid-20th century information retrieval research. Early work by scholars such as Gerard Salton introduced the vector space model of documents and queries, which underpinned term-weighting schemes like tf-idf (term frequency – inverse document frequency). These ideas influenced both academic search engines and, eventually, commercial web search.

As web content exploded, researchers in IR refined ranking algorithms, indexing structures and relevance feedback mechanisms. The prevailing paradigm assumed that users could express needs in terms of keywords or structured queries, and that the system’s job was to approximate relevance as accurately as possible given those inputs.

Agent-led commerce departs from this model by treating language not as a set of keywords but as an interface for describing goals, constraints and preferences in natural form. Instead of mapping queries to documents, agents must map intentions to actions and sequences of actions – choose, bundle, schedule, pay, deliver.

Recommender systems and personalisation pioneers

The science of recommending products, films or content to users based on their behaviour has roots in the 1990s and early 2000s. Key theorists and practitioners include:

  • John Riedl and colleagues, whose work on collaborative filtering and the GroupLens project showed how crowd data could be used to predict individual preferences.
  • Yehuda Koren, whose contributions to matrix factorisation methods during the Netflix Prize competition demonstrated the power of latent factor models in recommendation.
  • Joseph Konstan and others who explored user experience and trust in recommender systems, highlighting that perceived transparency and control can be as important as accuracy.

These researchers established that it is possible – and commercially powerful – to infer what customers might want, even before they search. Their theories informed the design of recommendation engines across retail, streaming and social platforms.

Agent-led commerce builds on this tradition but extends it. Instead of recommending within a narrow context (“people who bought this also bought”), agents must manage multi-step goals, cross-category constraints and time-sensitive logistics. This requires integrating recommender logic with planning algorithms and conversational interfaces.

Software agents and multi-agent systems

The concept of a software agent – an autonomous entity that perceives its environment, makes decisions and acts on a user’s behalf – has deep roots in AI research. Theorists in this area include:

  • Michael Wooldridge, whose work on multi-agent systems formalised how agents can reason, cooperate and compete in complex environments.
  • Nick Jennings, who explored practical applications of autonomous agents in business, including negotiation, resource allocation and supply chain management.
  • Stuart Russell and Peter Norvig, whose widely adopted AI textbook set out the rational agent framework, defining intelligent behaviour as actions that maximise expected utility given beliefs about the world.

In this tradition, agents are not simply chat interfaces; they are decision-making entities with objectives, models of the environment and policies for action. Many of the recent ideas around “agentic” systems – where software components can autonomously plan, call tools, execute workflows and coordinate with other agents – derive conceptually from this line of research.

In retail, agentic commerce can be seen as a large-scale deployment of these ideas: shopper-facing agents negotiate between customer preferences, product availability, pricing, promotions and logistics, while back-end agents manage inventory, routing and labour scheduling.

Conversational AI and natural language understanding

The move from query-driven search to conversational agents has been enabled by advances in natural language processing (NLP), particularly large language models (LLMs). Theorists and practitioners in this domain include researchers who developed transformer architectures, attention mechanisms and large-scale pre-training techniques.

These models provide the linguistic and semantic fluency required for agents to engage in open-ended dialogue. However, in commerce they must be grounded in reliable data and constrained by business rules. Walmart’s AI strategy, for example, combines general-purpose language models with retail-specific systems like Wallaby, which is tuned to Walmart’s own data on catalogues, substitutions and seasonality.1

Behavioural economics and choice architecture

The design of agent-led experiences also draws on insights from behavioural economics and psychology. Researchers such as Daniel Kahneman, Amos Tversky, Richard Thaler and Cass Sunstein have shown how framing, defaults and choice architecture influence decisions.

In an agentic commerce environment, the agent effectively becomes the architect of the customer’s choice set. It decides which alternatives to present, how to explain trade-offs and what defaults to propose. The ethical and strategic implications are significant: the same technologies that can reduce friction and cognitive load can also be used to steer behaviour in subtle ways.

Leading thinkers in digital ethics and AI governance have therefore argued for transparency, contestability and human oversight in agentic systems. For retailers, this becomes a trust question: customers need to believe that the agent is working in their interests, not solely maximising short-term conversion or margin.

Google, Gemini and open standards for agentic commerce

On the technology platform side, Google has been a central theorist and practitioner in both search and AI. With Gemini, its family of multimodal models, Google is positioning AI not just as a backend enhancement to search results but as a front-end conversational partner.

In the joint Walmart – Google initiative, the companies highlight a “Universal Commerce Protocol” designed to let agents interact with merchants in a standardised way.3 While technical details continue to evolve, the ambition reflects a broader movement towards open or semi-open standards for how agents discover, price, bundle and purchase across multiple commerce ecosystems.

Sundar Pichai, Google’s CEO, has spoken of AI improving every step of the consumer journey, from discovery to delivery, and has explicitly framed the Walmart partnership as a step toward making “agentic commerce” a reality.3 This aligns with the longer arc of Google’s evolution from ten blue links to rich results, shopping tabs and now conversational, transaction-capable agents.

Strategic implications: trust, control and the future of retail interfaces

Furner’s quote hints at the strategic contest that agent-led commerce will intensify. Key questions include:

  • Who owns the interface? If customers increasingly begin journeys inside a small number of dominant agents (Gemini, ChatGPT, other assistants), traditional notions of direct traffic, branded apps and search engine optimisation will be reconfigured.
  • Who sets the rules? Universal protocols for agentic commerce could distribute power more widely, but the entities that define and maintain those protocols will have disproportionate influence.
  • How is trust earned and maintained? Mistakes in retail – wrong products, failed deliveries, billing errors – have tangible consequences. Agent-led systems must combine probabilistic AI outputs with robust guardrails, validation checks and escalation paths to humans.
  • How does work change? As McMillon has noted, and Furner will now operationalise, AI will touch every job in the organisation.4 Theorists of work and automation have long debated the balance between augmentation and substitution; agentic commerce will be one of the most visible test cases of those theories in practice.

Walmart’s own AI roadmap suggests a disciplined approach: build AI into the fabric of operations, prioritise store-first use cases, move carefully from assistants to agents with strict guardrails and develop platforms that can be standardised and scaled globally.1 Furner’s quote can thus be read as both a declaration of intent and a statement of competitive philosophy: in a world where AI agents mediate more and more of daily life, retailers must choose whether to be controlled by those agents or to help design them.

For customers, the promise is compelling: less time on search and comparison, more time on what the purchases enable in their lives. For retailers and technologists, the challenge is to build agents that are not only powerful and convenient but also aligned, transparent and worthy of long-term trust. That is the deeper context behind Furner’s assertion that the move from web and app search to agent-led commerce is not just another technology upgrade, but the “next great evolution in retail”.

References

1. https://www.mcmillandoolittle.com/walmarts-big-ai-bet-and-what-might-change-under-new-ceo-john-furner/

2. https://pulse2.com/walmart-and-google-turn-ai-discovery-into-effortless-shopping-experiences/

3. https://corporate.walmart.com/news/2026/01/11/walmart-and-google-turn-ai-discovery-into-effortless-shopping-experiences

4. https://www.digitalcommerce360.com/2026/01/08/how-walmart-is-using-ai/

5. https://www.nasdaq.com/press-release/walmart-and-google-turn-ai-discovery-effortless-shopping-experiences-2026-01-11

6. https://www.emarketer.com/content/walmart-tech-first-strategy-shapes-growth

7. https://www.futurecommerce.com/podcasts/predictions-2026-prepare-for-the-age-of-autonomy

read more
Quote: Associated Press – On AI shopping

Quote: Associated Press – On AI shopping

“Google, OpenAI and Amazon all are racing to create tools that would allow for seamless AI-powered shopping.” – Associated Press

When the Associated Press observes that “Google, OpenAI and Amazon all are racing to create tools that would allow for seamless AI-powered shopping”, it is capturing a pivotal moment in the evolution of retail and of the internet itself. The quote sits at the intersection of several long-running trends: the shift from search to conversation, from static websites to intelligent agents, and from one-size-fits-all retail to deeply personalised, data-driven commerce.

Behind this single sentence lies a complex story of technological breakthroughs, strategic rivalry between the worlds largest technology platforms, and a reimagining of how people discover, evaluate and buy what they need. It also reflects the culmination of decades of research in artificial intelligence, recommendation systems, human-computer interaction and digital economics.

The immediate context: AI agents meet the shopping basket

The Associated Press line comes against the backdrop of a wave of partnerships between AI platforms and major retailers. Google has been integrating its Gemini AI assistant with large retail partners such as Walmart and Sams Club, allowing users to move from a conversational query directly to tailored product recommendations and frictionless checkout.

Instead of typing a product name into a search bar, a shopper can describe a situation or a goal, such as planning a camping trip or furnishing a first flat. Gemini then uses natural language understanding and retailer catalogues to surface relevant items, combine them into coherent baskets and arrange rapid delivery, in some cases within hours.1,3 The experience is meant to feel less like using a website and more like speaking to a highly knowledgeable personal shopper.

Walmart leaders have described this shift as a move from traditional search-based ecommerce to what they call “agent-led commerce” – shopping journeys mediated not by menus and filters but by AI agents that understand intent, context and personal history.1,2,3 For Google, this integration is both a way to showcase the capabilities of its Gemini models and a strategic response to OpenAIs work with retailers like Walmart, Etsy and a wide range of Shopify merchants through tools such as Instant Checkout.2,3

OpenAI, in parallel, has enabled users to browse and buy directly within ChatGPT, turning the chatbot into a commercial surface as well as an information tool.2,3 Amazon, for its part, has been weaving generative AI into its core marketplace, logistics and voice assistant, using AI models to improve product discovery, summarise reviews, optimise pricing and automate seller operations. Each company is betting that the next era of retail will be shaped by AI agents that can orchestrate entire end-to-end journeys from inspiration to doorstep.

From web search to agentic commerce

The core idea behind “seamless AI-powered shopping” is the replacement of fragmented, multi-step customer journeys with coherent, adaptive experiences guided by AI agents. Historically, online shopping has been built around search boxes, category trees and static product pages. The burden has been on the consumer to know what they want, translate that into search terms, sift through results and manually assemble baskets.

Agentic commerce reverses this burden. The AI system becomes an active participant: interpreting vague goals, proposing options, remembering preferences, coordinating logistics and handling payments, often across multiple merchants. Google and OpenAI have both underpinned their efforts with new open protocols designed to let AI agents communicate with a wide ecosystem of retailers, payment providers and loyalty systems.3,5

Google refers to its initiative as a Universal Commerce Protocol and describes it as a new standard that allows agents and systems to talk to each other across each step of the shopping journey.3,5 OpenAI, in turn, introduced the Agentic Commerce Protocol in partnership with Stripe, enabling ChatGPT and other agents to complete purchases from Etsy and millions of Shopify merchants.3 The technical details differ, but the strategic goal is shared: create an infrastructure layer that allows any capable AI agent to act as a universal shopping front end.

In practice, this means that a single conversation might involve discovering a new product, joining a retailers loyalty scheme, receiving personalised offers, adding related items and completing payment – without ever visiting a conventional website or app. The Associated Press quote calls out the intensity of the competition between the major platforms to control this new terrain.

The Associated Press as observer and interpreter

The Associated Press (AP), the attributed source of the quote, has a distinctive role in this story. Founded in 1846, AP is one of the worlds oldest and most widely used news agencies. It operates as a non-profit cooperative, producing reporting that is syndicated globally and used as a baseline for coverage by broadcasters, newspapers and digital platforms.

AP has long been known for its emphasis on factual, neutral reporting, and over the past decade it has also become notable for its early adoption of AI in news production. It has experimented with automated generation of corporate earnings summaries, sports briefs and other data-heavy stories, while also engaging in partnerships with technology companies around synthetic media and content labelling.

By framing the competition between Google, OpenAI and Amazon as a “race” to build seamless AI shopping, AP is doing more than simply documenting product launches. It is drawing attention to the structural stakes: the question of who will mediate the everyday economic decisions of billions of people. APs wording underscores both the speed of innovation and the concentration of power in a handful of technology giants.

APs technology and business correspondents, in covering this domain, typically triangulate between company announcements, analyst commentary and academic work on AI and markets. The quote reflects that blend: it is rooted in concrete developments such as the integration of Gemini with major retailers and the emergence of new commerce protocols, but it also hints at broader theoretical debates about platforms, data and consumer autonomy.

Intellectual roots: from recommendation engines to intelligent agents

The idea of seamless, AI-mediated shopping is the visible tip of an intellectual iceberg that stretches back decades. Several overlapping fields contribute to the current moment: information retrieval, recommender systems, multi-sided platforms, behavioural economics and conversational AI. The leading theorists in these areas laid the groundwork for the systems now shaping retail.

Search and information retrieval

Long before conversational agents, the central challenge of online commerce was helping people find relevant items within vast catalogues. Researchers in information retrieval, such as Gerard Salton in the 1960s and 1970s, developed foundational models for document ranking and term weighting that later underpinned web search.

In the context of commerce, the key innovation was the integration of relevance ranking with commercial signals such as click-through rates, purchase behaviour and sponsored listings. Googles original PageRank algorithm, associated with Larry Page and Sergey Brin, revolutionised how information was organised on the web and provided the basis for search advertising – itself a driver of modern retail. As search became the dominant gateway to online shopping, the line between information retrieval and marketing blurred.

The move to AI-powered shopping agents extends this lineage. Instead of ranking static pages, large language models interpret natural language queries, generate synthetic descriptions and orchestrate actions such as adding items to a basket. The theoretical challenge shifts from simply retrieving documents to modelling context, intent and dialogue.

Recommender systems and personalisation

Much of seamless AI-powered shopping depends on the ability to personalise offers and predict what a particular consumer is likely to want. This traces back to work on recommender systems in the 1990s and 2000s. Pioneers such as John Riedl and Joseph Konstan developed early collaborative filtering systems that analysed user ratings to make personalised suggestions.

The famous Netflix Prize in the mid-2000s catalysed work on matrix factorisation and latent factor models, with researchers like Yehuda Koren demonstrating how to predict preferences from sparse interaction data. Amazon itself became synonymous with recommender systems, popularising the idea that “customers who bought this also bought” could drive significant incremental revenue.

Over time, recommendation theory has expanded to consider not just accuracy but diversity, serendipity and fairness. Work by researchers such as Gediminas Adomavicius and Alexander Tuzhilin analysed trade-offs between competing objectives in recommender systems, while others explored issues of filter bubbles and echo chambers.

In AI-powered shopping, these theoretical concerns are amplified. When a single conversational agent mediates choices across many domains, its recommendation logic effectively becomes a form of personalised market design. It can nudge users towards particular brands, balance commercial incentives with user welfare, and shape long-term consumption habits. The underlying theories of collaborative filtering, contextual bandits and reinforcement learning now operate in a more visible, consequential arena.

Multi-sided platforms and the economics of marketplaces

The race between Google, OpenAI and Amazon is also a contest between different platform models. Economists such as Jean-Charles Rochet and Jean Tirole provided the canonical analysis of multi-sided platforms – markets where intermediaries connect distinct groups of users, such as buyers and sellers, advertisers and viewers.

The theory of platform competition explains why network effects and data accumulation can produce powerful incumbents, and why controlling the interface through which users access multiple services confers strategic advantages. Amazon Marketplace, Google Shopping and ad networks, and now AI agents embedded in operating systems or browsers, can all be seen through this lens.

Further work by David Evans, Andrei Hagiu and others explored platform governance, pricing structures and the strategic choice between being a neutral intermediary or a competitor to ones own participants. These ideas are highly relevant when AI agents choose which merchants or products to recommend and on what terms.

Seamless AI shopping turns the agent itself into a platform. It connects consumers, retailers, payment services, logistics providers and loyalty schemes through a conversational interface. The Universal Commerce Protocol and the Agentic Commerce Protocol can be understood as attempts to standardise interactions within this multi-sided ecosystem.3,5 The underlying tensions – between openness and control, neutrality and self-preferencing – are illuminated by platform economics.

Behavioural economics, choice architecture and digital nudging

While traditional economics often assumes rational agents and transparent markets, the reality of digital commerce has always been shaped by design: the ordering of search results, the framing of options, the use of defaults, and the timing of prompts. Behavioural economists like Daniel Kahneman, Amos Tversky and Richard Thaler have demonstrated how real-world decision-making deviates from rational models and how “choice architecture” can influence outcomes.

In online retail, this has manifested as a rich literature on digital nudging: subtle interface choices that steer behaviour. Researchers in human-computer interaction and behavioural science have documented how factors such as social proof, scarcity cues and personalised messaging affect conversion.

AI-powered shopping agents add another layer. Instead of static designs, the conversation itself becomes the choice architecture. The way an AI agent frames options, in what order it presents them, how it responds to hesitation and how it explains trade-offs, all shape consumer welfare. Theorists working at the intersection of AI and behavioural economics are now grappling with questions of transparency, autonomy and manipulation in agentic environments.

Conversational AI and human-computer interaction

The ability to shop by talking to an AI depends on advances in natural language processing, dialogue modelling and user-centred design. The early work of Joseph Weizenbaum (ELIZA) and the subsequent development of chatbots provided the conceptual foundations, but the major leap came with deep learning and large language models.

Researchers such as Yoshua Bengio, Geoffrey Hinton and Yann LeCun advanced the neural network architectures that underpin todays generative models. Within natural language processing, work by many teams on sequence-to-sequence learning, attention mechanisms and transformer architectures led to systems capable of understanding and generating human-like text.

OpenAI popularised the transformer-based large language model with the GPT series, while Google researchers contributed foundational work on transformers and later developed models like BERT and its successors. These advances turned language interfaces from novelties into robust tools capable of handling complex, multi-turn interactions.

Human-computer interaction specialists, meanwhile, studied how people form mental models of conversational agents, how trust is built or undermined, and how to design dialogues that feel helpful rather than intrusive. The combination of technical capability and design insight has made it plausible for people to rely on an AI agent to curate shopping choices.

Autonomous agents and “agentic” AI

The term “agentic commerce” used by Walmart and Google points to a broader intellectual shift: viewing AI systems not just as passive tools but as agents capable of planning and executing sequences of actions.1,5 In classical AI, agent theory has its roots in work on autonomous systems, reinforcement learning and decision-making under uncertainty.

Reinforcement learning theorists such as Richard Sutton and Andrew Barto formalised the idea of an agent learning to act in an environment to maximise reward. In ecommerce, this can translate into systems that learn how best to present options, when to offer discounts or how to balance immediate sales with long-term customer satisfaction.

Recent research on tool-using agents goes further, allowing language models to call external APIs, interact with databases and coordinate services. In commerce settings, that means an AI can check inventory, query shipping options, apply loyalty benefits and complete payments – all within a unified reasoning loop. Googles and OpenAIs protocols effectively define the “environment” in which such agents operate and the “tools” they can use.3,5

The theoretical questions now concern safety, alignment and control: how to ensure that commercially motivated agents act in ways that are consistent with user interests and regulatory frameworks, and how to audit their behaviour when their decision-making is both data-driven and opaque.

Corporate protagonists: Google, OpenAI and Amazon

The Associated Press quote names three central actors, each with a distinct history and strategic posture.

Google: from search to Gemini-powered commerce

Google built its business on organising the worlds information and selling targeted advertising against search queries. Its dominance in web search made it the default starting point for many online shopping journeys. As user behaviour has shifted towards conversational interfaces and specialised shopping experiences, Google has sought to extend its role from search engine to AI companion.

Gemini, Googles family of large language models and AI assistants, sits at the heart of this effort. By integrating Gemini into retail scenarios, Google is attempting to ensure that when people ask an AI for help – planning a project, solving a problem or buying a product – it is their agent, not a competitors, that orchestrates the journey.1,3,5

Partnerships with retailers such as Walmart, Target, Shopify, Wayfair and others, combined with the Universal Commerce Protocol, are strategic levers in this competition.1,3,4,5 They allow Google to showcase Gemini as a shopping concierge while making it easier for merchants to plug into the ecosystem without bespoke integrations for each AI platform.

OpenAI: from research lab to commerce gateway

OpenAI began as a research-focused organisation with a mission to ensure that artificial general intelligence benefits humanity. Over time, it has commercialised its work through APIs and flagship products such as ChatGPT, which rapidly became one of the fastest-growing consumer applications in history.

As users started to rely on ChatGPT not just for information but for planning and decision-making, the platform became an attractive entry point for commerce. OpenAIs Instant Checkout feature and the Agentic Commerce Protocol reflect an attempt to formalise this role. By enabling users to buy directly within ChatGPT from merchants on platforms like Shopify and Etsy, OpenAI is turning its assistant into a transactional hub.2,3

In this model, the AI agent can browse catalogues, compare options and present distilled choices, collapsing the distance between advice and action. The underlying theory draws on both conversational AI and platform economics: OpenAI positions itself as a neutral interface layer connecting consumers and merchants, while also shaping how information and offers are presented.

Amazon: marketplace, infrastructure and the invisible AI layer

While the provided context focuses more explicitly on Google and OpenAI, Amazon is an equally significant player in AI-powered shopping. Its marketplace already acts as a giant, data-rich environment where search, recommendation and advertising interact.

Amazon has deployed AI across its operations: in demand forecasting, warehouse robotics, delivery routing, pricing optimisation and its Alexa voice assistant. It has also invested heavily in generative AI to enhance product search, summarise reviews and assist sellers with content creation.

From a theoretical standpoint, Amazon exemplifies the vertically integrated platform: it operates the marketplace, offers its own branded products, controls logistics and, increasingly, provides the AI services that mediate discovery. Its approach to AI shopping is therefore as much about improving internal efficiency and customer experience as about creating open protocols.

In the race described by AP, Amazons strengths lie in its end-to-end control of the commerce stack and its granular data on real-world purchasing behaviour. As conversational and agentic interfaces become more common, Amazon is well placed to embed them deeply into its existing shopping flows.

Retailers as co-architects of AI shopping

Although the quote highlights technology companies, retailers such as Walmart, Target and others are not passive recipients of AI tools. They are actively shaping how agentic commerce unfolds. Walmart, for example, has worked with both OpenAI and Google, enabling Instant Checkout in ChatGPT and integrating its catalogue and fulfilment options into Gemini.1,2,3

Walmart executives have spoken about “rewriting the retail playbook” and closing the gap between “I want it” and “I have it” using AI.2 The company has also launched its own AI assistant, Sparky, within its app, and has been candid about how AI will transform roles across its workforce.2

These moves reflect a broader theoretical insight from platform economics: large retailers must navigate their relationships with powerful technology platforms carefully, balancing the benefits of reach and innovation against the risk of ceding too much control over customer relationships. By participating in open protocols and engaging multiple AI partners, retailers seek to maintain some leverage and avoid lock-in.

Other retailers and adjacent companies are exploring similar paths. Home Depot, for instance, has adopted Gemini-based agents to provide project planning and aisle-level guidance in stores, while industrial partners like Honeywell are using AI to turn physical spaces into intelligent, sensor-rich environments.5 These developments blur the line between online and offline shopping, extending the idea of seamless AI-powered commerce into bricks-and-mortar settings.

The emerging theory of AI-mediated markets

As AI agents become more entwined with commerce, several theoretical threads are converging into what might be called the theory of AI-mediated markets:

  • Information symmetry and asymmetry: AI agents can, in principle, reduce information overload and help consumers navigate complex choices. But they also create new asymmetries, as platform owners may know far more about aggregate behaviour than individual users.
  • Algorithmic transparency and accountability: When an AI agent chooses which products to recommend, the criteria may include relevance, profit margins, sponsorship and long-term engagement. Understanding and governing these priorities is an active area of research and regulation.
  • Competition and interoperability: The existence of multiple commerce protocols and agent ecosystems raises questions about interoperability, switching costs and the potential for AI-mediated markets to become more or less competitive than their predecessors.
  • Personalisation versus autonomy: Enhanced personalisation can make shopping more efficient and enjoyable but may also narrow exposure to alternatives or gently steer behaviour in ways that users do not fully perceive.
  • Labour and organisational change: As AI takes on more of the cognitive labour of retail – from customer service to merchandising – the roles of human workers evolve. The theoretical work on technology and labour markets gains a new frontier in AI-augmented retail operations.

Researchers from economics, computer science, law and sociology are increasingly studying these dynamics, building on the earlier theories of platforms, recommendations and behavioural biases but extending them into a world where the primary interface to the market is itself an intelligent agent.

Why this moment matters

The Associated Press quote distils a complex, multi-layered transformation into a single observation: the most powerful technology firms are in a race to define how we shop in an age of AI. The endpoint of that race is not just faster checkout or more targeted ads. It is a restructuring of the basic relationship between consumers, merchants and the digital intermediaries that connect them.

Search boxes and product grids are giving way to conversations. Static ecommerce sites are being replaced or overlaid by agents that can understand context, remember preferences and act on our behalf. The theories of information retrieval, recommendation, platforms and behavioural economics that once described separate facets of digital commerce are converging in these agents.

Understanding the backstory of this quote – the intellectual currents, corporate strategies and emerging protocols behind it – is essential for grasping the stakes of AI-powered shopping. It is not merely a technological upgrade; it is a shift in who designs, controls and benefits from the everyday journeys that connect intention to action in the digital economy.

References

1. https://pulse2.com/walmart-and-google-turn-ai-discovery-into-effortless-shopping-experiences/

2. https://www.thefinance360.com/walmart-partners-with-googles-gemini-to-offer-ai-shopping-assistant-to-shoppers/

3. https://www.businessinsider.com/gemini-chatgpt-openai-google-competition-walmart-deal-2026-1

4. https://retail-insider.com/retail-insider/2026/01/google-expands-ai-shopping-with-walmart-shopify-wayfair/

5. https://cloud.google.com/transform/a-new-era-agentic-commerce-retail-ai

6. https://winningwithwalmart.com/walmart-teams-up-with-google-gemini-what-it-means-for-shoppers-and-suppliers/

read more
Quote: Pitchbook

Quote: Pitchbook

“Much of the market continues to find it difficult to raise venture capital funding. Non-AI companies have accounted for just 35% of deal value through Q3 2025, while representing more than 60% of completed deals.” – Pitchbook

PitchBook’s data through Q3 2025 reveals a stark disparity in venture capital (VC) funding, where non-AI companies captured just 35% of total deal value despite comprising over 60% of deals, underscoring investor preference for AI-driven opportunities amid market caution.1,4,5

Context of the Quote

This statistic, sourced from PitchBook’s Q3 2025 Venture Monitor (in collaboration with the National Venture Capital Association), highlights the “flight to quality” trend dominating VC dealmaking. Through the first nine months of 2025, overall deal counts reached 3,990 in Q1 alone (up 11% quarter-over-quarter), with total value hitting $91.5 billion—a post-2022 high driven largely by AI sectors.4,5 However, smaller and earlier-stage non-AI startups received only 36% of total value, the decade’s lowest share, as investors prioritized larger, AI-focused rounds amid uncertainties like tariffs, market volatility, and subdued consumer sentiment.3,4 Fundraising for VC funds also plummeted, with Q1 2025 seeing just 87 vehicles close at $10 billion—the lowest activity in over a decade—and dry powder nearing $300 billion but deploying slowly.4 Exit activity hinted at recovery ($56 billion in Q1 from 385 deals) but faltered due to paused IPOs (e.g., Klarna, StubHub) and reliance on outliers like Coreweave’s IPO, which accounted for nearly 40% of value.4 PitchBook’s H1 2025 VC Tech Survey of 32 investors confirmed this shift: 52% see AI disrupting fintech (up from 32% in H2 2024), with healthcare, enterprise tech, and cybersecurity following suit, while VC outlooks soured (only 38% expect rising funding, down from 58%).1 The quote encapsulates a market where volume persists but value concentrates in AI, leaving non-AI firms struggling for capital in a selective environment.

Backstory on PitchBook

PitchBook, founded in 2007 by John Gabbert in Seattle, emerged as a leading data provider for private capital markets from humble origins as a simple Excel-based tool for tracking VC and private equity deals. Acquired by Morningstar in 2016 for $225 million, it has grown into an authoritative platform aggregating data on over 3 million companies, 1.5 million funds, and millions of deals worldwide, powering reports like the PitchBook-NVCA Venture Monitor.3,4,5 Its Q3 2025 analysis draws from proprietary datasets as of late 2025, offering granular insights into deal counts, values, sector breakdowns, and fundraising—essential for investors navigating post-2022 VC normalization. PitchBook’s influence stems from its real-time tracking and predictive modeling, cited across industry reports for benchmarking trends like AI dominance and liquidity pressures.1,2,4

Leading Theorists on VC Market Dynamics and AI Concentration

The quote aligns with foundational theories on VC cycles, power laws, and technological disruption. Key thinkers include:

  • Bill Janeway (author of Doing Capitalism in the Innovation Economy, 2012): A veteran VC at Warburg Pincus, Janeway theorized VC as a “three-legged stool” of government R&D, entrepreneurial risk-taking, and financial engineering. He predicted funding concentration in breakthrough tech like AI during downturns, as investors seek “moonshots” amid capital scarcity—mirroring 2025’s non-AI value drought.1,4

  • Peter Thiel (co-founder of PayPal, Founders Fund; Zero to One, 2014): Thiel’s “definite optimism” framework argues VCs favor monopolistic, tech-dominant firms (e.g., AI) over competitive commoditized ones, enforcing power-law distributions where 80-90% of returns come from 1-2% of deals. This explains non-AI firms’ deal volume without value, as Thiel warns against “indefinite optimism” in crowded sectors.4

  • Andy Kessler (author of Venture Capital Deals, 1986; Wall Street Journal columnist): Kessler formalized the VC “spray and pray” model evolving into selective bets during liquidity crunches, predicting AI-like waves would eclipse legacy sectors—evident in 2025’s fintech AI disruption forecasts.1

  • Scott Kupor (a16z managing partner; Secrets of Sand Hill Road, 2019): Kupor analyzes LP-VC dynamics, noting how dry powder buildup (nearing $300B in 2025) leads to extended fund timelines and AI favoritism, as LPs demand outsized returns amid low distributions.1,2,4

  • Diane Mulcahy (former Providence Equity; The New World of Entrepreneurship, 2013): Mulcahy critiqued VC overfunding bubbles, advocating “patient capital” for non-hyped sectors; her warnings resonate in 2025’s fundraising cliff and non-AI funding gaps.4

These theorists collectively frame 2025’s trends as a power-law amplification of AI amid cyclical caution, building on historical VC patterns from the dot-com bust to post-2008 recovery.

References

1. https://www.foley.com/insights/publications/2025/06/investor-insights-overview-pitchbook-h1-2025-vc-tech-survey/

2. https://www.sganalytics.com/blog/us-venture-capital-outlook-2025/

3. https://www.deloitte.com/us/en/services/audit-assurance/articles/trends-in-venture-capital.html

4. https://www.junipersquare.com/blog/vc-q1-2025

5. https://nvca.org/wp-content/uploads/2025/10/Q3-2025-PitchBook-NVCA-Venture-Monitor.pdf

read more
Quote: Nathaniel Whittemore – AI Daily Brief

Quote: Nathaniel Whittemore – AI Daily Brief

“If you want to get a preview of what everyone else is going to be dealing with six months from now, there’s basically not much better you can do than watching what developers are talking about right now.” – Nathaniel Whittemore – AI Daily Brief – On: Tailwind CSS and AI disruption

This observation captures a pattern that has repeated itself through every major technology wave of the past half-century. The people who live closest to the tools – the engineers, open source maintainers and framework authors – are usually the first to encounter both the power and the problems that the rest of the world will later experience at scale. In the current artificial intelligence cycle, that dynamic is especially clear: developers are experimenting with new models, agents and workflows months before they become mainstream in business, design and everyday work.

Nathaniel Whittemore and the AI Daily Brief

The quote comes from Nathaniel Whittemore, better known in technology circles as NLW, the host of The AI Daily Brief: Artificial Intelligence News and Analysis (formerly The AI Breakdown).4,7,9 His show has emerged as a daily digest and analytical lens on the rapid cascade of AI announcements, research papers, open source projects and enterprise case studies. Rather than purely cataloguing news, Whittemore focuses on how AI is reshaping business models, labour, creative work and the broader economy.4

Whittemore has built a reputation as an interpreter between worlds: the fast-moving communities of AI engineers and builders on the one hand, and executives, policymakers and non-technical leaders on the other. Episodes range from detailed walkthroughs of specific tools and models to long-read analyses of how organisations are actually deploying AI in the field.1,5 His recurring argument is that the most important AI stories are not just technical; they are about context, incentives and the way capabilities diffuse into real workflows.1,4

On his show and in talks, Whittemore frequently returns to the idea that AI is best understood through its users: the people who push tools to their limits, improvise around their weaknesses and discover entirely new categories of use. In recent years, that has meant tracking developers who integrate AI into code editors, build autonomous agents, or restructure internal systems around AI-native processes.3,8 The quote about watching developers is, in effect, a mental model for anyone trying to see around the next corner.

Tailwind CSS as the context for the quote

The quote lands inside a very specific story: Tailwind CSS as a case study in AI-enabled demand with AI-damaged monetisation.

Tailwind is an open-source, utility-first CSS framework that became foundational to modern front-end development. It is widely adopted by developers and heavily used by AI coding tools. Tailwind’s commercial model, however, depends on a familiar open-source pattern: the core framework is free, and revenue comes from paid add-ons (the “Plus” tier). Critically, the primary channel to market for those paid offerings was the documentation.

AI broke that channel.

As AI coding tools improved, many developers stopped visiting documentation pages. Instead, they asked the model and got the answer immediately—often derived from scraped docs and community content. Usage of Tailwind continued to grow, but the discovery path for paid offerings weakened because humans no longer needed to read the docs. In plain terms: the product stayed popular, but the funnel collapsed.

That is why this story resonated beyond CSS. It shows a broader pattern: AI can remove the need for the interface you monetise—even while it increases underlying adoption. For any business that relies on “users visit our site, then convert,” Tailwind is not a niche developer drama. It is a preview.

Tailwind’s episode makes the mechanism of disruption uncomfortably clear: AI tools boosted adoption, but also removed the need for humans to visit Tailwind’s documentation. That mattered because the documentation was Tailwind’s primary channel to market—where users discovered the paid “Plus” offerings that funded maintenance. Once AI started answering questions directly from scraped content, the funnel broke: fewer doc visits meant fewer conversions, and a widely used framework suddenly struggled to monetise the very popularity AI helped accelerate.

AI Disruption Seen from the Builder Front Line

In the AI era, this pattern is amplified. AI capabilities roll out as research models, APIs and open source libraries long before they are wrapped in polished consumer interfaces. Developers are often the first group to:

  • Benchmark new models, probing their strengths and failure modes.
  • Integrate them into code editors, data pipelines, content tools and internal dashboards.
  • Build specialised agents tuned to niche workflows or industry-specific tasks.6,8
  • Stress-test the economics of running models at scale and find where they can genuinely replace or augment existing systems.3,5

Whittemore’s work sits precisely at this frontier. Episodes dissect the emergence of coding agents, the economics of inference, the rise of AI-enabled “tiny teams”, and the way reasoning models are changing expectations around what software can autonomously do.3,8 He tracks how new agentic capabilities go from developer experiments to production deployments in enterprises, often in less than a year.3,5

His quote reframes this not as a curiosity but as a practical strategy: if you want to understand what your organisation or industry will be wrestling with in six to twelve months – from new productivity plateaus to unfamiliar risks – you should look closely at what AI engineers and open source maintainers are building and debating now.

Developers as Lead Users: Theoretical Roots

Behind Whittemore’s intuition sits a substantial body of innovation research. Long before AI, scholars studied why certain groups seemed to anticipate the needs and behaviours of the wider market. Several theoretical strands help explain why watching developers is so powerful.

Eric von Hippel and Lead User Theory

MIT innovation scholar Eric von Hippel developed lead user theory to describe how some users experience needs earlier and more intensely than the general market. These lead users frequently innovate on their own, building or modifying products to solve their specific problems. Over time, their solutions diffuse and shape commercial offerings.

Developers often fit this lead user profile in technology markets. They are:

  • Confronted with cutting-edge problems first – scaling systems, integrating new protocols, or handling novel data types.
  • Motivated to create tools and workflows that relieve their own bottlenecks.
  • Embedded in communities where ideas, snippets and early projects can spread quickly and be iterated upon.

Tailwind CSS itself reflects this: it emerged as a developer-centric solution to recurring front-end pain points, then radiated outward to reshape how teams approach design systems. In AI, developer-built tooling often precedes large commercial platforms, as seen with early AI coding assistants, monitoring tools and evaluation frameworks.3,8

Everett Rogers and the Diffusion of Innovations

Everett Rogers’ classic work on the diffusion of innovations describes how new ideas spread through populations in phases: innovators, early adopters, early majority, late majority and laggards. Developers often occupy the innovator or early adopter categories for digital technologies.

Rogers stressed that watching these early groups offers a glimpse of future mainstream adoption. Their experiments reveal not only whether a technology is technically possible, but how it will be framed, understood and integrated into social systems. In AI, the debates developers have about safety, guardrails, interpretability and tooling are precursors to the regulatory, ethical and organisational questions that follow at scale.4,5

Clayton Christensen and Disruptive Innovation

Clayton Christensen’s theory of disruptive innovation emphasises how new technologies often begin in niches that incumbents overlook. Early adopters tolerate rough edges because they value new attributes – lower cost, flexibility, or a different performance dimension – that established customers do not yet prioritise.

AI tools and frameworks frequently begin life like this: half-finished interfaces wrapped around powerful primitives, attractive primarily to technical users who can work around their limitations. Developers discover where these tools are genuinely good enough, and in doing so, they map the path by which a once-nascent capability becomes a serious competitive threat.

Open Source Communities and Collective Foresight

Another important line of thinking comes from research on open source software and user-driven innovation. Scholars such as Steven Weber and Yochai Benkler have explored how distributed communities coordinate to build complex systems without traditional firm structures.

These communities act as collective sensing networks. Bug reports, pull requests, issue threads and design discussions form a live laboratory where emerging practices are tested and refined. In AI, this is visible in the rapid evolution of open weights models, fine-tuning techniques, evaluation harnesses and orchestration frameworks. The tempo of progress in these spaces often sets the expectations which commercial vendors then have to match or exceed.6,8

AI-Specific Perspectives: From Labs to Production

Beyond general innovation theory, several contemporary AI thinkers and practitioners shed light on why developer conversations are such powerful predictors.

Andrej Karpathy and the Software 2.0 Vision

Former Tesla AI director Andrej Karpathy popularised the term “Software 2.0” to describe a shift from hand-written rules to learned neural networks. In this paradigm, developers focus less on explicit logic and more on data curation, model selection and feedback loops.

Under a Software 2.0 lens, developers are again early indicators. They experiment with prompt engineering, fine-tuning, retrieval-augmented generation and multi-agent systems. Their day-to-day struggles – with context windows, hallucinations, latency and cost-performance trade-offs – foreshadow the operational questions businesses later face when they automate processes or embed AI in products.

Ian Goodfellow, Yoshua Bengio and Deep Learning Pioneers

Deep learning pioneers such as Ian Goodfellow, Yoshua Bengio and Geoffrey Hinton illustrated how research breakthroughs travel from lab settings into practical systems. What began as improvements on benchmark datasets and academic competitions became, within a few years, the foundation for translation services, recommendation engines, speech recognition and image analysis.

Developers building on these techniques were the bridge between research and industry. They discovered how to deploy models at scale, handle real-world data, and integrate AI into existing stacks. In today’s generative AI landscape, the same dynamic holds: frontier models and architectures are translated into frameworks, SDKs and reference implementations by developer communities, and only then absorbed into mainstream tools.

AI Engineers and the Rise of Agents

Recent work at the intersection of AI and software engineering has focused on agents: AI systems that can plan, call tools, write and execute code, and iteratively refine their own outputs. Industry reports summarised on The AI Daily Brief highlight how executives are beginning to grasp the impact of these agents on workflows and organisational design.5

Yet developers have been living with these systems for longer. They are the ones:

  • Embedding agents into CI/CD pipelines and testing regimes.
  • Using them to generate and refactor large codebases.3,6
  • Designing guardrails and permissions to keep them within acceptable bounds.
  • Developing evaluation harnesses to measure quality, robustness and reliability.8

Their experiments and post-mortems provide an unvarnished account of both the promise and the fragility of agentic systems. When Whittemore advises watching what developers are talking about, this is part of what he means: the real-world friction points that will later surface as board-level concerns.

Context, Memory and Business Adoption

Whittemore has also emphasised how advances in context and memory – the ability of AI systems to integrate and recall large bodies of information – are changing what is possible in the enterprise.1 He highlights features such as:

  • Tools that allow models to access internal documents, code repositories and communication platforms securely, enabling organisation-specific reasoning.1
  • Modular context systems that let AI draw on different knowledge packs depending on the task.1
  • Emerging expectations that AI should “remember” ongoing projects, preferences and constraints rather than treating each interaction as isolated.1

Once again, developers are at the forefront. They are wiring these systems into data warehouses, knowledge graphs and production applications. They see early where context systems break, where privacy models need strengthening, and where the productivity gains are real rather than speculative.

From there, insights filter into broader business discourse: about data governance, AI strategy, vendor selection and the design of AI-native workflows. The lag between developer experience and executive recognition is, in Whittemore’s estimate, often measured in months – hence his six-month framing.

From Developer Talk to Strategic Foresight

The core message behind the quote is a practical discipline for anyone thinking about AI and software-driven change:

  • Follow where developers invest their time. Tools that inspire side projects, plugin ecosystems and community events often signal deeper shifts in how work will be done.
  • Listen to what frustrates them. Complaints about context limits, flaky APIs or insufficient observability reveal where new infrastructure, standards or governance will be needed.
  • Pay attention to what they take for granted. When a capability stops being exciting and becomes expected – instant code search, semantic retrieval, AI-assisted refactoring – it is often a sign that broader expectations in the market will soon adjust.
  • Watch the crossovers. When developer patterns show up in no-code tools, productivity suites or design platforms, the wave is moving from early adopters to the early majority.

Nathaniel Whittemore’s work with The AI Daily Brief is, in many ways, a structured practice of this approach. By curating, analysing and contextualising what builders are doing and saying in real time, he offers a way for non-technical leaders to see the outlines of the future before it is evenly distributed.4,7,9 The Tailwind CSS example is one case; the ongoing wave of AI disruption is another. The constant, across both, is that if you want to know what is coming next, you start by watching the people building it.

 

References

1. https://pod.wave.co/podcast/the-ai-daily-brief-formerly-the-ai-breakdown-artificial-intelligence-news-and-analysis/ai-context-gets-a-major-upgrade

2. https://www.youtube.com/watch?v=MdfYA3xv8jw

3. https://www.youtube.com/watch?v=0EDdQchuWsA

4. https://podcasts.apple.com/us/podcast/the-ai-daily-brief-artificial-intelligence-news/id1680633614

5. https://www.youtube.com/watch?v=nDDWWCqnR60

6. https://www.youtube.com/watch?v=f34QFs7tVjg

7. https://open.spotify.com/show/7gKwwMLFLc6RmjmRpbMtEO

8. https://podcasts.apple.com/us/podcast/the-biggest-trends-from-the-ai-engineer-worlds-fair/id1680633614?i=1000711906377

9. https://www.audible.com/podcast/The-AI-Breakdown-Daily-Artificial-Intelligence-News-and-Discussions/B0C3Q4BG17

 

read more
Quote: Blackrock

Quote: Blackrock

“AI’s buildout is also happening at a potentially unprecedented speed and scale. This shift to capital-intensive growth from capital-light, is profoundly changing the investment environment – and pushing limits on multiple fronts, physical, financial and socio-political.” – Blackrock

The quote highlights BlackRock’s observation that artificial intelligence (AI) infrastructure development is advancing at an extraordinary pace and magnitude, shifting economic growth models from low-capital-intensity (e.g., software-driven scalability) to high-capital demands, while straining physical infrastructure like power grids, financial systems through massive leverage needs, and socio-political frameworks amid geopolitical tensions.1,2

Context of the Quote

This statement emerges from BlackRock’s 2026 Investment Outlook, published by the BlackRock Investment Institute (BII), the firm’s research arm focused on macro trends and portfolio strategy. It encapsulates discussions from BlackRock’s internal 2026 Outlook Forum in late 2025, where AI’s “buildout”—encompassing data centers, chips, and energy infrastructure—dominated debates among portfolio managers.2 Key concerns included front-loaded capital expenditures (capex) estimated at $5-8 trillion globally through 2030, creating a “financing hump” as revenues lag behind spending, potentially requiring increased leverage in an already vulnerable financial system.1,3,5 Physical limits like compute capacity, materials, and especially U.S. power grid strain were highlighted, with AI data centers projected to drive massive electricity demand amid U.S.-China strategic competition.2 Socio-politically, it ties into “mega forces” like geopolitical fragmentation, blurring public-private boundaries (e.g., via stablecoins), and policy shifts from inflation control to neutral stances, fostering market dispersion where only select AI beneficiaries thrive.2,4 BlackRock remains pro-risk, overweighting U.S. AI-exposed stocks, active strategies, private credit, and infrastructure while underweighting long-term Treasuries.1,5

BlackRock and the Quoted Perspective

BlackRock, the world’s largest asset manager with nearly $14 trillion in assets under management as of late 2025, issues annual outlooks to guide institutional and retail investors.3 The quote aligns with BII’s framework of “mega forces”—structural shifts like AI, geopolitics, and financial evolution—launched years prior to frame investments in a fragmented macro environment.2 Key voices include Rick Rieder, BlackRock’s Chief Investment Officer of Fixed Income, who in related 2026 insights emphasized AI as a “cost and margin story,” potentially slashing labor costs (55% of business expenses) by 5%, unlocking $1.2 trillion in annual U.S. savings and $82 trillion in present-value corporate profits.4 BII analysts note AI’s speed surpasses prior tech waves, with capex ambitions making “micro macro,” though uncertainties persist on revenue capture by tech giants versus broader dispersion.1,3

Backstory on Leading Theorists of AI’s Economic Transformation

The quote draws on decades of economic theory about technological revolutions, capital intensity, and growth limits, pioneered by thinkers who analyzed how innovations like electrification and computing reshaped productivity, investment, and society.

  • Robert Gordon (The Rise and Fall of American Growth, 2016): Gordon, an NBER economist, argues U.S. productivity growth has stagnated since 1970 (averaging ~2% annually over 150 years) due to diminishing returns from past innovations like electricity and sanitation, contrasting AI’s potential but warning of “hump”-like front-loaded costs without guaranteed back-loaded gains—mirroring BlackRock’s financing concerns.3,4

  • Erik Brynjolfsson and Andrew McAfee (The Second Machine Age, 2014; Machine, Platform, Crowd, 2017): MIT scholars at the Initiative on the Digital Economy posit AI enables exponential productivity via automation of cognitive tasks, shifting from capital-light digital scaling to infrastructure-heavy buildouts (e.g., data centers), but predict “recombination” winners amid labor displacement and inequality—echoing BlackRock’s dispersion and socio-political strains.4

  • Daron Acemoglu and Simon Johnson (Power and Progress, 2023): MIT economists critique tech optimism, asserting AI’s direction depends on institutional choices; undirected buildouts risk elite capture and gridlock (physical/financial limits), not broad prosperity, aligning with BlackRock’s U.S.-China rivalry and policy debates.2

  • Nicholas Crafts (historical productivity scholar): Building on 20th-century analyses, Crafts documented electrification’s 1920s-1930s “productivity paradox”—decades of heavy capex before payoffs—paralleling AI’s current phase, where investments outpace adoption.1

  • Jensen Huang (NVIDIA CEO, practitioner-theorist): While not academic, Huang’s 2024-2025 forecasts of $1 trillion+ annual AI capex by 2030 popularized the “buildout” narrative, influencing BlackRock’s scale estimates and energy focus.3,5

These theorists underscore AI as a capital-intensive pivot akin to the Second Industrial Revolution, but accelerated, with BlackRock synthesizing their ideas into actionable investment amid 2025-2026 market highs (e.g., Nasdaq peaks) and volatility (e.g., tech routs).2,3

References

1. https://www.blackrock.com/americas-offshore/en/insights/blackrock-investment-institute/outlook

2. https://www.medirect.com.mt/updates/news/all-news/blackrock-commentary-ai-front-and-center-at-our-2026-forum/

3. https://www.youtube.com/watch?v=Ww7Zy3MAWAs

4. https://www.blackrock.com/us/financial-professionals/insights/investing-in-2026

5. https://www.blackrock.com/us/financial-professionals/insights/ai-stocks-alternatives-and-the-new-market-playbook-for-2026

6. https://www.blackrock.com/corporate/insights/blackrock-investment-institute/publications/outlook

read more
Quote: Blackrock

Quote: Blackrock

“AI is not only an innovation itself but has the potential to accelerate other innovation.” – Blackrock

This quote originates from BlackRock’s 2026 Investment Outlook published by its Investment Institute, emphasizing AI’s dual role as a transformative technology and a catalyst for broader innovation across sectors like connectivity, security, and physical automation.6 BlackRock positions AI as a “mega force” driving digital disruption, with potential to automate tasks, enhance productivity, and unlock economic growth by enabling faster advancements in other fields.5,6

Context of the Quote

The statement reflects BlackRock’s strategic focus on AI as a cornerstone of long-term investment opportunities amid rapid technological evolution. In the 2026 Investment Outlook, BlackRock highlights AI’s capacity to go beyond task automation, fostering an “intelligence revolution” that amplifies innovation in interconnected technologies.1,6 This aligns with BlackRock’s actions, including launching active ETFs like the iShares A.I. Innovation and Tech Active ETF (BAI), which targets 20-40 global AI companies across infrastructure, models, and applications to capture growth in the AI stack.1,8 Tony Kim, head of BlackRock’s fundamental equities technology group, described this as seizing “outsized and overlooked investment opportunities across the full stack of AI and advanced technologies.”1 Similarly, the firm views active ETFs as the “next frontier in investment innovation,” expanding access to AI-driven returns.1

BlackRock’s commitment extends to massive infrastructure investments. In 2024, it co-founded the Global AI Infrastructure Investment Partnership (GAIIP, later AIP) with Global Infrastructure Partners (GIP), Microsoft, and MGX, aiming to mobilize up to $100 billion for U.S.-focused data centers and power infrastructure to support AI scaling.2,3,9 Larry Fink, BlackRock’s Chairman and CEO, stated these investments “will help power economic growth, create jobs, and drive AI technology innovation,” underscoring AI’s role in revitalizing economies.2 By 2025, NVIDIA and xAI joined AIP, reinforcing its open-architecture approach to accelerate AI factories and supply chains.3 BlackRock executives like Alex Brazier argue AI investments face no bubble risk; instead, capacity constraints in computing power and data centers demand more capital.4

BlackRock’s Backstory and Leadership

BlackRock, the world’s largest asset manager with $11.5 trillion in assets, evolved from a fixed-income specialist founded in 1988 by Larry Fink and partners at Blackstone into a global powerhouse after its 1994 spin-off and 2009 Barclays acquisition.2 Under Fink’s leadership since inception, BlackRock pioneered ETFs via iShares (acquired 2009) and Aladdin risk-management software, managing $32 billion in U.S. active ETFs.1 Its AI strategy integrates proprietary insights from the BlackRock Investment Institute, which identifies AI as interplaying with other “mega forces” like geopolitics and sustainability.5,6 Fink, a mortgage-backed securities innovator during the 1980s savings-and-loan crisis, has championed infrastructure and tech since steering BlackRock public in 1999; his AIP comments frame AI as a multi-trillion-dollar opportunity.2,3

Leading Theorists on AI as an Innovation Accelerator

The idea of AI accelerating other innovations traces to foundational thinkers in technology diffusion, general-purpose technologies (GPTs), and computational economics:

  • Erik Brynjolfsson and Andrew McAfee (MIT): In The Second Machine Age (2014) and subsequent works, they argue AI as a GPT—like electricity—initially boosts productivity slowly but then accelerates innovation across industries by enabling data-driven decisions and automation.5,6 Their research quantifies AI’s “exponential” complementarity, where it amplifies human ingenuity in fields like biotech and materials science.

  • Bengt Holmström and Paul Milgrom (Nobel 2019): Their principal-agent theories underpin AI’s role in aligning incentives for innovation; AI reduces information asymmetries, speeding R&D in multi-agent systems like supply chains.2

  • Jensen Huang (NVIDIA CEO): A practitioner-theorist, Huang describes accelerated computing and generative AI as powering the “next industrial revolution,” converting data into intelligence to propel every industry—echoed in his AIP role.2,3

  • Satya Nadella (Microsoft CEO): Frames AI as driving “growth across every sector,” with infrastructure as the enabler for breakthroughs, aligning with BlackRock’s partnerships.2

  • Historical roots: Building on Solow’s productivity paradox (1987)—why computers took decades to boost growth—theorists like Robert Gordon contrast narrow tech impacts with AI’s potential for broad acceleration, as BlackRock’s outlook affirms.6

These perspectives inform BlackRock’s view: AI isn’t isolated but a multiplier, demanding infrastructure to realize its full accelerative power.1,2,6

References

1. https://www.investmentnews.com/etfs/blackrock-broadens-active-etf-shelf-with-ai-and-tech-funds/257815

2. https://news.microsoft.com/source/2024/09/17/blackrock-global-infrastructure-partners-microsoft-and-mgx-launch-new-ai-partnership-to-invest-in-data-centers-and-supporting-power-infrastructure/

3. https://ir.blackrock.com/news-and-events/press-releases/press-releases-details/2025/BlackRock-Global-Infrastructure-Partners-Microsoft-and-MGX-Welcome-NVIDIA-and-xAI-to-the-AI-Infrastructure-Partnership-to-Drive-Investment-in-Data-Centers-and-Enabling-Infrastructure/default.aspx

4. https://getcoai.com/news/blackrock-exec-says-ai-investments-arent-in-a-bubble-capacity-is-the-real-problem/

5. https://www.blackrock.com/corporate/insights/blackrock-investment-institute/publications/mega-forces/artificial-intelligence

6. https://www.blackrock.com/corporate/insights/blackrock-investment-institute/publications/outlook

7. https://www.blackrock.com/uk/individual/products/339936/blackrock-ai-innovation-fund

8. https://www.blackrock.com/us/financial-professionals/products/339081/ishares-a-i-innovation-and-tech-active-etf

9. https://www.global-infra.com/news/mgx-blackrock-global-infrastructure-partners-and-microsoft-welcome-kuwait-investment-authority-kia-to-the-ai-infrastructure-partnership/

read more
Quote: Kaoutar El Maghraoui

Quote: Kaoutar El Maghraoui

“We can’t keep scaling compute, so the industry must scale efficiency instead.” – Kaoutar El Maghraoui – IBM Principal Research Scientist

“We can’t keep scaling compute, so the industry must scale efficiency instead.” – Kaoutar El Maghraoui, IBM Principal Research Scientist

This quote underscores a pivotal shift in AI development: as raw computational power reaches physical and economic limits, the focus must pivot to efficiency through optimized hardware, software co-design, and novel architectures like analog in-memory computing.1,2

Backstory and Context of Kaoutar El Maghraoui

Dr. Kaoutar El Maghraoui is a Principal Research Scientist at IBM’s T.J. Watson Research Center in Yorktown Heights, NY, where she leads the AI testbed at the IBM Research AI Hardware Center—a global hub advancing next-generation accelerators and systems for AI workloads.1,2 Her work centers on the intersection of systems research and artificial intelligence, including distributed systems, high-performance computing (HPC), and AI hardware-software co-design. She drives open-source development and cloud experiences for IBM’s digital and analog AI accelerators, emphasizing operationalization of AI in hybrid cloud environments.1,2

El Maghraoui’s career trajectory reflects deep expertise in scalable systems. She earned her PhD in Computer Science from Rensselaer Polytechnic Institute (RPI) in 2007, following a Master’s in Computer Networks (2001) and Bachelor’s in General Engineering from Al Akhawayn University, Morocco. Early roles included lecturing at Al Akhawayn and research on IBM’s AIX operating system—covering performance tuning, multi-core scheduling, Flash SSD storage, and OS diagnostics using IBM Watson cognitive tech.2,6 In 2017, she co-led IBM’s Global Technology Outlook, shaping the company’s AI leadership vision across labs and units.1,2

The quote emerges from her lectures and research on efficient AI deployment, such as “Powering the Future of Efficient AI through Approximate and Analog In-Memory Computing,” which addresses performance bottlenecks in deep neural networks (DNNs), and “Platform for Next-Generation Analog AI Hardware Acceleration,” highlighting Analog In-Memory Computing (AIMC) to reduce energy losses in DNN inference and training.1 It aligns with her 2026 co-authored paper “STARC: Selective Token Access with Remapping and Clustering for Efficient LLM Decoding on PIM Systems” (ASPLOS 2026), targeting efficiency in large language models via processing-in-memory (PIM).2 With over 2,045 citations on Google Scholar, her contributions span AI hardware optimization and performance.8

Beyond research, El Maghraoui is an ACM Distinguished Member and Speaker, Senior IEEE Member, and adjunct professor at Columbia University. She holds awards like the 2021 Best of IBM, IBM Eminence and Excellence for advancing women in tech, 2021 IEEE TCSVC Women in Service Computing, and 2022 IBM Technical Corporate Award. Leadership roles include global vice-chair of Arab Women in Computing (ArabWIC), co-chair of IBM Research Watson Women Network (2019-2021), and program/general co-chair for Grace Hopper Celebration (2015-2016).1,2

Leading Theorists in AI Efficiency and Compute Scaling Limits

The quote resonates with foundational theories on compute scaling limits and efficiency paradigms, pioneered by key figures challenging Moore’s Law extensions in AI hardware.

Theorist Key Contributions Relevance to Quote
Cliff Young & Contributors (Google) Co-authored “Scaling Laws for Neural Language Models” (2020, arXiv) and MLPerf benchmarks; advanced hardware-aware neural architecture search (NAS) for DNN optimization on edge devices.1 Demonstrates efficiency gains via NAS, directly echoing El Maghraoui’s lectures on hardware-specific DNN design to bypass compute scaling.1
Bill Dally (NVIDIA) Pioneer of processing-in-memory (PIM) and tensor cores; authored works on energy-efficient architectures amid “end of Dennard scaling” (power density limits post-2000s).2 Warns against endless compute scaling; promotes PIM and sparsity, aligning with El Maghraoui’s STARC paper and analog accelerators.2
Jeff Dean (Google) Formulated Chinchilla scaling laws (2022), showing optimal compute allocation balances parameters and data; co-developed TensorFlow and TPUs for efficiency.2 Highlights diminishing returns of pure compute scaling, urging efficiency in training/inference—core to IBM’s AI Hardware Center focus.1,2
Hadi Esmaeilzadeh (Georgia Tech) Introduced neurocube and analog in-memory computing (AIMC) concepts (e.g., “Navigating the Energy Wall” papers); quantified AI’s “memory wall” and von Neumann bottlenecks.1 Foundational for El Maghraoui’s AIMC advocacy, proving analog methods boost DNN efficiency by 10-100x over digital compute scaling.1
Song Han (MIT) Developed pruning, quantization, and NAS (e.g., TinyML, HAWQ frameworks); showed 90%+ parameter reduction without accuracy loss.1 Enables “scale efficiency” for real-world deployment, as in El Maghraoui’s “Optimizing Deep Learning for Real-World Deployment” lecture.1

These theorists collectively established that post-Moore’s Law (transistor density doubling every ~2 years, slowing since 2010s), AI progress demands efficiency multipliers: sparsity, analog compute, co-design, and beyond-von Neumann architectures. El Maghraoui’s work operationalizes these at IBM scale, from cloud-native DL platforms to PIM for LLMs.1,2,6

References

1. https://speakers.acm.org/speakers/el_maghraoui_19271

2. https://research.ibm.com/people/kaoutar-el-maghraoui

3. https://github.com/kaoutar55

4. https://orcid.org/0000-0002-1967-8749

5. https://www.sharjah.ac.ae/-/media/project/uos/sites/uos/research/conferences/wirf2025/webinars/dr-kaoutar-el-maghraoui-_webinar.pdf

6. https://s3.us.cloud-object-storage.appdomain.cloud/res-files/1843-Kaoutar_ElMaghraoui_CV_Dec2022.pdf

7. https://www.womentech.net/speaker/all/all/69100

8. https://scholar.google.com/citations?user=yDp6rbcAAAAJ&hl=en

“We can’t keep scaling compute, so the industry must scale efficiency instead.” - Quote: Kaoutar El Maghraoui

read more
Quote: Andrew Yeung

Quote: Andrew Yeung

“The first explicitly anti-AI social network will emerge. No AI-generated posts, no bots, no synthetic engagement, and proof-of-person required. People are already revolting against AI ‘slop’” – Andrew Yeung – Tech investor

Andrew Yeung: Tech Investor and Community Builder

Andrew Yeung is a prominent tech investor, entrepreneur, and events host known as the “Gatsby of Silicon Alley” by Business Insider for curating exclusive tech gatherings that draw founders, CEOs, investors, and operators.1,2,4 After 20 years in China, he moved to the U.S., leading products at Facebook and Google before pivoting to startups, investments, and community-building.2 As a partner at Next Wave NYC—a pre-seed venture fund backed by Flybridge—he has invested in over 20 early-stage companies, including Hill.com (real estate tech), Superpower (health tech), Othership (wellness), Carry (logistics), and AI-focused ventures like Natura (naturaumana.ai), Ruli (ruli.ai), Otis AI (meetotis.com), and Key (key.ai).2

Yeung hosts high-profile events through Fibe, his events company and 50,000+ member tech community, including Andrew’s Mixers (1,000+ person rooftop parties), The Junto Series (C-suite dinners), and Lumos House (multi-day mansion experiences across 8 cities like NYC, LA, Toronto, and San Francisco).1,2,4 Over 50,000 attendees, including billion-dollar founders, media figures, and Olympic athletes, have participated, with sponsors like Fidelity, J.P. Morgan, Perplexity, Silicon Valley Bank, Techstars, and Notion.2,4 His platform reaches 120,000+ tech leaders monthly and 1M+ people, aiding hundreds of founders in fundraising, hiring, and scaling.1,2 Yeung writes for Business Insider, his blog (andrew.today with 30,000+ readers), and has spoken at Princeton, Columbia Business School, SXSW, AdWeek, and Jason Calacanis’ This Week in Startups podcast on tech careers, networking, and entrepreneurship.1,2,4

Context of the Quote

The quote—”The first explicitly anti-AI social network will emerge. No AI-generated posts, no bots, no synthetic engagement, and proof-of-person required. People are already revolting against AI ‘slop’”—originates from Yeung’s newsletter post “11 Predictions for 2026 & Beyond,” published on andrew.today.3 It is prediction #9, forecasting a 2026 platform that bans AI content, bots, and fake interactions, enforcing human verification to restore authentic connections.3 Yeung cites rising backlash against AI “slop”—low-quality synthetic media—with studies showing 20%+ of YouTube recommendations for new users as such content.3 He warns of the “dead internet theory” (the idea that much online activity is bot-driven) becoming reality without human-only spaces, driven by demand for genuine interaction amid AI dominance.3

This prediction aligns with Yeung’s focus on human-centric tech: his investments blend AI tools (e.g., Otis AI, Ruli) with platforms enhancing real-world connections (e.g., events, networking advice emphasizing specific intros, follow-ups, and clarity in asks).1,2 In podcasts, he stresses high-value networking via precise value exchanges, like linking founders to niche investors, mirroring his vision for “proof-of-person” authenticity over synthetic engagement.1,4

Backstory on Leading Theorists and Concepts

The quote draws from established ideas on AI’s societal impact, particularly the Dead Internet Theory. Originating in online forums around 2021, it posits that post-2016 internet content is increasingly AI-generated, bot-amplified, and human-free, eroding authenticity—evidenced by studies like a 2024 analysis finding 20%+ of YouTube videos as low-effort AI slop, as Yeung notes.3 Key proponents include:

  • Ignas (u/illuminoATX): The pseudonymous 4chan user who formalized the theory in 2021, arguing algorithms prioritize engagement-farming bots over humans, citing examples like identical comment patterns and ghost towns on social platforms.

  • Zach Vorhies (ex-Google whistleblower): Popularized it via Twitter (now X) and interviews, analyzing YouTube’s algorithm favoring synthetic content; his 2022 claims align with Yeung’s YouTube stats.

  • Media Amplifiers: The Atlantic (2023 article “Maybe You Missed It, but the Internet Died Five Years Ago”) and New York Magazine substantiated it with data on bot proliferation (e.g., 40-50% of web traffic as bots per Imperva reports).

Related theorists on AI slop and authenticity revolts include:

  • Ethan Mollick (Wharton professor, author of Co-Intelligence): Critiques AI’s “hallucinated” mediocrity flooding culture; warns of “enshittification” (Cory Doctorow’s term for platform decay via AI spam), predicting user flight to verified-human spaces.[Inference: Mollick’s 2024 writings echo Yeung’s revolt narrative.]

  • Cory Doctorow: Coined “enshittification” (2023), describing how platforms degrade via ad-driven AI content; advocates decentralized, human-verified alternatives.

  • Jaron Lanier (VR pioneer, You Are Not a Gadget): Early critic of social media’s dehumanization; in 2024’s There Is No Antimemetics Division, pushes “humane tech” rejecting synthetic engagement.

These ideas fuel real-world responses: platforms like Bluesky and Mastodon emphasize human moderation, while proof-of-person tech (e.g., Worldcoin’s iris scans, though controversial) tests Yeung’s vision. His prediction positions him as a connector spotting unmet needs in a bot-saturated web.3

References

1. https://www.youtube.com/watch?v=uO0dI_tCvUU

2. https://www.andrewyeung.co

3. https://www.andrew.today/p/11-predictions-for-2026-and-beyond

4. https://www.youtube.com/watch?v=MdI0RhGhySI

5. https://www.andrew.today/p/my-ai-productivity-stack

“The first explicitly anti-AI social network will emerge. No AI-generated posts, no bots, no synthetic engagement, and proof-of-person required. People are already revolting against AI ‘slop’” - Quote: Andrew Yeung

read more
Quote: Blackrock

Quote: Blackrock

“The AI builders are leveraging up: investment is front-loaded while revenues are back-loaded. Along with highly indebted governments, this creates a more levered financial system vulnerable to shocks like bond yield spikes.” – Blackrock – 2026 Outlook

The AI Financing Paradox: How Front-Loaded Investment and Back-Loaded Returns are Reshaping Global Financial Risk

The Quote in Context

BlackRock’s 2026 Investment Outlook identifies a critical structural vulnerability in global markets: the massive capital requirements of AI infrastructure are arriving years before the revenue benefits materialize1. This temporal mismatch creates what the firm describes as a financing “hump”—a period of intense leverage accumulation across both the private sector and government balance sheets, leaving financial systems exposed to potential shocks from rising bond yields or credit market disruptions1,2.

The quote reflects BlackRock’s core thesis that AI’s economic impact will be transformational, but the path to that transformation is fraught with near-term financial risks. As the world’s largest asset manager, overseeing nearly $14 trillion in assets, BlackRock’s assessment carries significant weight in shaping investment strategy and market expectations3.

The Investment Spend-Revenue Gap

The scale of the AI buildout is staggering. BlackRock projects $5-8 trillion in AI-related capital expenditure through 20305, with annual spending estimated at 5-8 trillion dollars globally until that date3. This represents the fastest technological buildout in recent centuries, yet the economics are unconventional: companies are committing enormous capital today with the expectation that productivity gains and revenue growth will materialize later2.

BlackRock notes that while the overall revenues AI eventually generates could theoretically justify the spending at a macroeconomic level, it remains unclear how much of that value will accrue to the tech companies actually building the infrastructure1,2. This uncertainty creates a critical vulnerability—if AI deployment proves less profitable than anticipated, or if adoption rates slow, highly leveraged companies may struggle to service their debt obligations.

The Leverage Imperative

The financing structure is not optional; it is inevitable. AI spending necessarily precedes benefits and revenues, creating an unavoidable need for long-term financing and greater leverage2. Tech companies and infrastructure providers cannot wait years to recoup their investments—they must borrow in capital markets today to fund construction, equipment, and operations.

This creates a second layer of risk. As companies issue bonds to finance AI capex, they increase corporate debt levels. Simultaneously, governments worldwide remain highly indebted from pandemic stimulus and ongoing fiscal pressures. The combination produces what BlackRock identifies as a “more levered financial system”—one where both public and private sector balance sheets are stretched1.

The Vulnerability to Shocks

BlackRock’s warning about vulnerability to “shocks like bond yield spikes” is particularly prescient. In a highly leveraged environment, rising interest rates have cascading effects:

  • Refinancing costs increase: Companies and governments face higher borrowing costs when existing bonds mature and must be renewed.
  • Debt service burden rises: Higher yields directly increase the cost of servicing existing debt, reducing profitability and fiscal flexibility.
  • Credit spreads widen: Investors demand higher risk premiums, making debt more expensive across the board.
  • Forced deleveraging: Companies unable to service debt at higher rates may need to cut spending, sell assets, or restructure obligations.

The AI buildout amplifies this risk because so much spending is front-loaded. If yield spikes occur before significant productivity gains materialize, companies may lack the cash flow to manage higher borrowing costs, creating potential defaults or forced asset sales that could trigger broader financial instability.

BlackRock’s Strategic Response

Rather than abandoning risk, BlackRock has taken a nuanced approach: the firm remains pro-risk and overweight U.S. stocks on the AI theme1, betting that the long-term benefits will justify near-term leverage accumulation. However, the firm has also shifted toward tactical underweighting of long-term Treasuries and identified opportunities in both public and private credit markets to manage risk while maintaining exposure1.

This reflects a sophisticated view: the financial system’s increased leverage is a real concern, but the AI opportunity is too significant to avoid. Instead, active management and diversification across asset classes become essential.

Broader Economic Context

The leverage dynamic intersects with broader macroeconomic shifts. BlackRock emphasizes that inflation is no longer the central issue driving markets; instead, labor dynamics and the distributional effects of AI now matter more4. The firm projects that AI could generate roughly $1.2 trillion in annual labor cost savings, translating into about $878 billion in incremental after-tax corporate profits each year, with a present value on the order of $82 trillion for corporations and another $27 trillion for AI providers4.

These enormous potential gains justify the current spending—on a macro level. Yet for individual investors and companies, dispersion and default risk are rising4. The benefits of AI will be highly concentrated among successful implementers, while laggards face obsolescence. This uneven distribution of gains and losses adds another layer of risk to a more levered financial system.

Historical and Theoretical Parallels

The AI financing paradox echoes historical technology cycles. During the dot-com boom of the late 1990s, massive capital investment in internet infrastructure preceded revenue generation by years, creating similar leverage vulnerabilities. The subsequent crash revealed how vulnerable highly leveraged systems are to disappointment about future growth rates.

However, this cycle differs in scale and maturity. Unlike the dot-com era, AI is already demonstrating productivity benefits across multiple sectors. The question is not whether AI creates value, but whether the timeline and magnitude of value creation justify the financial risks being taken today.


BlackRock’s insight captures a fundamental tension in modern finance: transformative technological change requires enormous upfront capital, yet highly leveraged financial systems are fragile. The path forward depends on whether productivity gains materialize quickly enough to validate the investment and reduce leverage before external shocks test the system’s resilience.

References

1. https://www.blackrock.com/americas-offshore/en/insights/blackrock-investment-institute/outlook

2. https://www.youtube.com/watch?v=eFBwyu30oTU

3. https://www.youtube.com/watch?v=Ww7Zy3MAWAs

4. https://www.blackrock.com/us/financial-professionals/insights/investing-in-2026

5. https://www.blackrock.com/us/financial-professionals/insights/ai-stocks-alternatives-and-the-new-market-playbook-for-2026

6. https://www.blackrock.com/corporate/insights/blackrock-investment-institute/publications/outlook

7. https://www.blackrock.com/institutions/en-us/insights/2026-macro-outlook

read more
Quote: Hari Vasudevan – Utility Dive

Quote: Hari Vasudevan – Utility Dive

“Data centers used 4% of U.S. electricity two years ago and are on track to devour three times that by 2028.” – Hari Vasudevan – Utility Dive

Hari Vasudevan is the founder and CEO of KYRO AI, an AI-powered platform designed to streamline operations in utilities, vegetation management, disaster response, and critical infrastructure projects, supporting over $150 billion in program value by enhancing safety, efficiency, and cost savings for contractors and service providers.1,3,4

Backstory and Context of the Quote

The quote—”Utilities that embrace artificial intelligence will set reliability and affordability standards for decades to come”—originates from Vasudevan’s November 26, 2025, opinion piece in Utility Dive titled “Data centers are breaking the old grid. Let AI build the new one.”1,6 In it, he addresses the grid’s strain from surging data center demand fueled by AI, exemplified by Georgia regulators’ summer 2025 rules to protect residential customers from related cost hikes.6 Vasudevan argues that the U.S. power grid faces an “inflection point,” where clinging to a reactive 20th-century model leads to higher bills and outages, while AI adoption enables a resilient system balancing homes, businesses, and digital infrastructure.1,6 This piece builds on his November 2025 Energy Intelligence article urging utilities and hyperscalers (e.g., tech giants building data centers) to collaborate via dynamic load management, on-site generation, and shared capital risks to avoid burdening ratepayers.5 The context reflects escalating challenges: data centers are driving grid overloads, extreme weather has caused $455 billion in U.S. storm damage since 1980 (one-third in the last five years), and utility rate disallowances have risen to 35-40% from 2019-2023 amid regulatory scrutiny.4,5,6

Vasudevan’s perspective stems from hands-on experience. He founded Think Power Solutions to provide construction management and project oversight for electric utilities, managing multi-billion-dollar programs nationwide and achieving a 100% increase in working capital turns alongside 57% growth by improving billing accuracy, reducing delays, and bridging field-office gaps in thin-margin industries.3 After exiting as CEO, he launched KYRO AI to apply these efficiencies at scale, particularly for storm response—where AI optimizes workflows for linemen, fleets, and regulators amid rising billion-dollar weather events—and infrastructure buildouts like transmission lines powering data centers.3,4 In a CCCT podcast, he emphasized AI’s role in powering the economy during uncertain times, closing gaps that erode profits, and aiding small construction businesses.3

Leading Theorists in AI for Grid Modernization and Utility Resilience

Vasudevan’s advocacy aligns with pioneering work in AI applications for energy systems. Key theorists include:

  • Amory Lovins: Co-founder of Rocky Mountain Institute, Lovins pioneered “soft path” energy theory in the 1970s, advocating distributed resources over centralized grids—a concept echoed in maximizing home/business energy assets for resilience, as Vasudevan supports via AI orchestration.1
  • Massoud Amin: Often called the “father of the smart grid,” Amin (University of Minnesota) developed early frameworks for AI-driven, self-healing grids in the 2000s, integrating sensors and automation to prevent blackouts and enhance reliability amid data center loads.4,6
  • Andrew Ng: Stanford professor and AI pioneer (co-founder of Coursera, former Baidu chief scientist), Ng has theorized AI’s role in predictive grid maintenance and demand forecasting since 2010s deep learning breakthroughs, directly influencing tools like KYRO for storm response and vegetation management.3,4
  • Bri-Mathias Hodge: NREL researcher advancing AI/ML for renewable integration and grid stability, with models optimizing distributed energy resources—core to Vasudevan’s push against “breaking the old grid.”1,5

These theorists provide the intellectual foundation: Lovins for decentralization, Amin for smart infrastructure, Ng for scalable AI, and Hodge for optimization, all converging on AI as essential for affordable, resilient grids facing AI-driven demand.1,4,5,6

 

References

1. https://www.utilitydive.com/opinion/

2. https://www.utilitydive.com/?page=1&p=505

3. https://www.youtube.com/watch?v=g8q16BWXk4o

4. https://www.utilitydive.com/news/ai-utility-storm-response-kyro/752172/

5. https://www.energyintel.com/0000019b-2712-d02f-adfb-e7932e490000

6. https://www.utilitydive.com/news/ai-utilities-reliability-cost/805224/

 

Data centers used 4% of U.S. electricity two years ago and are on track to devour three times that by 2028. - Quote: Hari Vasudevan - Utility Dive

read more
Quote: Professor Anil Bilgihan – Florida Atlantic University Business

Quote: Professor Anil Bilgihan – Florida Atlantic University Business

“AI agents will be the new gatekeepers of loyalty, The question is no longer just ‘How do we win a customer’s heart?’ but ‘How do we win the trust of the algorithms that are advising them?’” – Professor Anil Bilgihan – Florida Atlantic University Business

Professor Anil Bilgihan: Academic and Research Profile

Professor Anil Bilgihan is a leading expert in services marketing and hospitality information systems at Florida Atlantic University’s College of Business, where he serves as a full Professor in the Marketing Department with a focus on Hospitality Management.1,2,4 He holds the prestigious Harry T. Mangurian Professorship and previously the Dean’s Distinguished Research Fellowship, recognizing his impactful work at the intersection of technology, consumer behavior, and the hospitality industry.2,3

Education and Early Career

Bilgihan earned his PhD in 2012 from the University of Central Florida’s Rosen College of Hospitality Management, specializing in Education/Hospitality Education Track.1,2 He holds an MS in Hospitality Information Management (2009) from the University of Delaware and a BS in Computer Technology and Information Systems (2007) from Bilkent University in Turkey.1,2,4 His technical foundation in computer systems laid the groundwork for his research in digital technologies applied to services.

Before joining FAU in 2013, he was a faculty member at The Ohio State University.2,4 At FAU, based in Fleming Hall Room 316 (Boca Raton), he teaches courses in hotel marketing and revenue management while directing research efforts.1,2

Research Contributions and Expertise

Bilgihan’s scholarship centers on how technology transforms hospitality and tourism, including e-commerce, user experience, digital marketing, online social interactions, and emerging tools like artificial intelligence (AI).2,3,4 With over 70 refereed journal articles, 80 conference proceedings, an h-index of 38, and i10-index of 68—resulting in more than 18,000 citations—he is a prolific influencer in the field.2,4,7

Key recent publications highlight his forward-looking focus on generative AI:

  • Co-authored a 2025 framework for generative AI in hospitality and tourism research (Journal of Hospitality and Tourism Research).1
  • Developed a 2025 systematic review on AI awareness and employee outcomes in hospitality (International Journal of Hospitality Management).1
  • Explored generative AI’s implications for academic research in tourism and hospitality (2024, Tourism Economics).1

Earlier works include agent-based modeling for eWOM strategies (2021), AI assessment frameworks for hospitality (2021), and online community building for brands (2018).1 His research appears in top journals such as Tourism Management, International Journal of Hospitality Management, Computers in Human Behavior, and Journal of Service Management.2,4

Bilgihan co-authored the textbook Hospitality Information Technology: Learning How to Use It, widely used in the field.2,4 He serves on editorial boards (e.g., International Journal of Contemporary Hospitality Management), as associate editor of Psychology & Marketing, and co-editor of Journal of International Hospitality Management.2

Awards and Leadership Roles

Recognized with the Cisco Extensive Research Award, FAU Scholar of the Year Award, and Highly Commended Award from the Emerald/EFMD Outstanding Doctoral Research Awards.2,4 He contributes to FAU’s Behavioral Insights Lab, developing AI-digital marketing frameworks for customer satisfaction, and the Center for Services Marketing.3,5

Leading Theorists in Hospitality Technology and AI

Bilgihan’s work builds on foundational theorists in services marketing, technology adoption, and AI in hospitality. Key figures include:

  • Jill Kandampully (co-author on brand communities, 2018): Pioneer in services marketing and customer loyalty; her relational co-creation theory emphasizes technology’s role in value exchange (Journal of Hospitality and Tourism Technology).1
  • Peter Ricci (frequent collaborator): Expert in hospitality revenue management and digital strategies; advances real-time data analytics for tourism marketing.1,5
  • Ye Zhang (collaborator): Focuses on agent-based modeling and social media’s impact on travel; extends motivation theories for accessibility in tourism.1
  • Fred Davis (Technology Acceptance Model, TAM, 1989): Core influence on Bilgihan’s user experience research; TAM explains technology adoption via perceived usefulness and ease-of-use, widely applied in hospitality e-commerce.2 (Inferred from Bilgihan’s tech adoption focus.)
  • Viswanath Venkatesh (Unified Theory of Acceptance and Use of Technology, UTAUT, 2003): Builds on TAM for AI and digital tools; Bilgihan’s AI frameworks align with UTAUT’s performance expectancy in service contexts.3 (Inferred from AI decision-making emphasis.)
  • Ming-Hui Huang and Roland T. Rust: Leaders in AI-service research; their “AI substitution” framework (2018) informs Bilgihan’s hospitality AI assessments, predicting AI’s role in frontline service transformation.1 (Directly cited in Bilgihan’s 2021 AI paper.)

These theorists provide the theoretical backbone for Bilgihan’s empirical frameworks, bridging behavioral economics, information systems, and hospitality operations amid digital disruption.1,2,3,4

 

References

1. https://business.fau.edu/faculty-research/faculty-profiles/profile/abilgihan.php

2. https://www.madintel.com/team/anil-bilgihan

3. https://business.fau.edu/centers/behavioral-insights-lab/meet-behavioral-insights-experts/

4. https://sites.google.com/view/anil-bilgihan/

5. https://business.fau.edu/centers/center-for-services-marketing/center-faculty/

6. https://business.fau.edu/departments/marketing/hospitality-management/meet-faculty/

7. https://scholar.google.com/citations?user=5pXa3OAAAAAJ&hl=en

 

AI agents will be the new gatekeepers of loyalty, The question is no longer just ‘How do we win a customer’s heart?’ but ‘How do we win the trust of the algorithms that are advising them?’ - Quote: Professor Anil Bilgihan - Florida Atlantic University Business

read more
Quote: Naval Ravikant – Venture Capitalist

Quote: Naval Ravikant – Venture Capitalist

“UI is pre-AI.” – Naval Ravikant – Venture Capitalist

Naval Ravikant stands as one of Silicon Valley’s most influential yet unconventional thinkers—a figure who bridges the gap between pragmatic entrepreneurship and philosophical inquiry. His observation that “UI is pre-AI” reflects a distinctive perspective on technological evolution that warrants careful examination, particularly given his track record as an early-stage investor in transformative technologies.

The Architect of Modern Startup Infrastructure

Ravikant’s influence on the technology landscape extends far beyond individual company investments. As co-founder, chairman, and former CEO of AngelList, he fundamentally altered how early-stage capital flows through the startup ecosystem. AngelList democratised access to venture funding, creating infrastructure that connected aspiring entrepreneurs with angel investors and venture capital firms on an unprecedented scale. This wasn’t merely a business achievement; it represented a structural shift in how innovation gets financed globally.

His investment portfolio reflects prescient timing and discerning judgement. Ravikant invested early in companies including Twitter, Uber, Foursquare, Postmates, Yammer, and Stack Overflow—investments that collectively generated over 70 exits and more than 10 unicorn companies. This track record positions him not as a lucky investor, but as someone with genuine pattern recognition capability regarding which technologies would matter most.

Beyond the Venture Capital Thesis

What distinguishes Ravikant from conventional venture capitalists is his deliberate rejection of the traditional founder mythology. He explicitly advocates against the “hustle mentality” that dominates startup culture, instead promoting a more holistic conception of wealth that encompasses time, freedom, and peace of mind alongside financial returns. This philosophy shapes how he evaluates opportunities and mentors founders—he considers not merely whether a business will scale, but whether it will scale without scaling stress.

His broader intellectual contributions extend through multiple channels. With more than 2.4 million followers on Twitter (X), Ravikant regularly shares aphoristic insights blending practical wisdom with Eastern philosophical traditions. His appearances on influential podcasts, particularly the Tim Ferriss Show and Joe Rogan Experience, have introduced his thinking to audiences far beyond Silicon Valley. Most notably, his “How to Get Rich (without getting lucky)” thread has become foundational reading across technology and business communities, articulating principles around leverage through code, capital, and content.

Understanding “UI is Pre-AI”

The quote “UI is pre-AI” requires interpretation within Ravikant’s broader intellectual framework and the contemporary technological landscape. The statement operates on multiple levels simultaneously.

The Literal Interpretation: User interface design and development necessarily precedes artificial intelligence implementation in most technology products. This reflects a practical observation about product development sequencing—one must typically establish how users interact with systems before embedding intelligent automation into those interactions. In this sense, the UI is the foundational layer upon which AI capabilities are subsequently layered.

The Philosophical Dimension: More provocatively, the statement suggests that how we structure human-computer interaction through interface design fundamentally shapes the possibilities for what artificial intelligence can accomplish. The interface isn’t merely a presentation layer; it represents the primary contact point between human intent and computational capability. Before AI can be genuinely useful, the interface must make that utility legible and accessible to end users.

The Investment Perspective: For Ravikant specifically, this observation carries investment implications. It suggests that companies solving user experience problems will likely remain valuable even as AI capabilities evolve, whereas companies that focus purely on algorithmic sophistication without considering user interaction may find their innovations trapped in laboratory conditions rather than deployed in markets.

The Theoretical Lineage

Ravikant’s observation sits within a longer intellectual tradition examining the relationship between interface, interaction, and technological capability.

Don Norman and Human-Centered Design: The foundational modern work on this subject emerged from Don Norman’s research at the University of California, San Diego, particularly his seminal work on design of everyday things. Norman argued that excellent product design requires intimate understanding of human cognition, perception, and behaviour. Before any technological system—intelligent or otherwise—can create value, it must accommodate human limitations and leverage human strengths through thoughtful interface design.

Douglas Engelbart and Augmentation Philosophy: Douglas Engelbart’s mid-twentieth-century work on human-computer augmentation established that technology’s primary purpose should be extending human capability rather than replacing human judgment. His thinking implied that interfaces represent the crucial bridge between human cognition and computational power. Without well-designed interfaces, the most powerful computational systems remain inert.

Alan Kay and Dynabook Vision: Alan Kay’s vision of personal computing—articulated through concepts like the Dynabook—emphasised that technology’s democratising potential depends entirely on interface accessibility. Kay recognised that computational power matters far less than whether ordinary people can productively engage with that power through intuitive interaction models.

Contemporary HCI Research: Modern human-computer interaction research builds on these foundations, examining how interface design shapes which problems users attempt to solve and how they conceptualise solutions. Researchers like Shneiderman and Plaisant have demonstrated empirically that interface design decisions have second-order effects on what users believe is possible with technology.

The Contemporary Context

Ravikant’s statement carries particular resonance in the current artificial intelligence moment. As organisations rush to integrate large language models and other AI systems into products, many commit what might be called “technology-first” errors—embedding sophisticated algorithms into user experiences that haven’t been thoughtfully designed to accommodate them.

Meaningful user interface design for AI-powered systems requires addressing distinct challenges: How do users understand what an AI system can and cannot do? How is uncertainty communicated? How are edge cases handled? What happens when the AI makes errors? These questions cannot be answered through better algorithms alone; they require interface innovation.

Ravikant’s observation thus functions as a corrective to the current technological moment. It suggests that the companies genuinely transforming industries through artificial intelligence will likely be those that simultaneously innovate in both algorithmic capability and user interface design. The interface becomes pre-AI not merely chronologically but causally—shaping what artificial intelligence can accomplish in practice rather than merely in principle.

Investment Philosophy Integration

This observation aligns with Ravikant’s broader investment thesis emphasising leverage and scalability. An excellent user interface represents exactly this kind of leverage—it scales human attention and human decision-making without requiring proportional increases in effort or resources. Similarly, artificial intelligence scaled through well-designed interfaces amplifies this effect, allowing individual users or organisations to accomplish work that previously required teams.

Ravikant’s focus on investments at seed and Series A stages across media, content, cloud infrastructure, and AI reflects implicit confidence that the foundational layer of how humans interact with technology remains unsettled terrain. Rather than assuming interface design has been solved, he appears to recognise that each new technological capability—whether cloud infrastructure or artificial intelligence—creates new design challenges and opportunities.

The quote ultimately encapsulates a distinctive investment perspective: that attention to human interaction, to aesthetics, to usability, represents not secondary ornamentation but primary technological strategy. In an era of intense focus on algorithmic sophistication, Ravikant reminds us that the interface through which those algorithms engage with human needs and human judgment represents the true frontier of technological value creation.

read more
Quote: Ilya Sutskever – Safe Superintelligence

Quote: Ilya Sutskever – Safe Superintelligence

“The robustness of people is really staggering.” – Ilya Sutskever – Safe Superintelligence

This statement, made in his November 2025 conversation with Dwarkesh Patel, comes from someone uniquely positioned to make such judgments: co-founder and Chief Scientist of Safe Superintelligence Inc., former Chief Scientist at OpenAI, and co-author of AlexNet—the 2012 paper that launched the modern deep learning era.

Sutskever’s claim about robustness points to something deeper than mere durability or fault tolerance. He is identifying a distinctive quality of human learning: the ability to function effectively across radically diverse contexts, to self-correct without explicit external signals, to maintain coherent purpose and judgment despite incomplete information and environmental volatility, and to do all this with sparse data and limited feedback. These capacities are not incidental features of human intelligence. They are central to what makes human learning fundamentally different from—and vastly superior to—current AI systems.

Understanding what Sutskever means by robustness requires examining not just human capabilities but the specific ways in which AI systems are fragile by comparison. It requires recognising what humans possess that machines do not. And it requires understanding why this gap matters profoundly for the future of artificial intelligence.

What Robustness Actually Means: Beyond Mere Reliability

In engineering and systems design, robustness typically refers to a system’s ability to continue functioning when exposed to perturbations, noise, or unexpected conditions. A robust bridge continues standing despite wind, earthquakes, or traffic loads beyond its design specifications. A robust algorithm produces correct outputs despite noisy inputs or computational errors.

But human robustness operates on an entirely different plane. It encompasses far more than mere persistence through adversity. Human robustness includes:

  1. Flexible adaptation across domains: A teenager learns to drive after ten hours of practice and then applies principles of vehicle control, spatial reasoning, and risk assessment to entirely new contexts—motorcycles, trucks, parking in unfamiliar cities. The principles transfer because they have been learned at a level of abstraction and generality that allows principled application to novel situations.
  2. Self-correction without external reward: A learner recognises when they have made an error not through explicit feedback but through an internal sense of rightness or wrongness—what Sutskever terms a “value function” and what we experience as intuition, confidence, or unease. A pianist knows immediately when they have struck a wrong note; they do not need external evaluation. This internal evaluative system enables rapid, efficient learning.
  3. Judgment under uncertainty: Humans routinely make decisions with incomplete information, tolerating ambiguity whilst maintaining coherent action. A teenager drives defensively not because they can compute precise risk probabilities but because they possess an internalized model of danger, derived from limited experience but somehow applicable to novel situations.
  4. Stability across time scales: Human goals, values, and learning integrate across vastly different temporal horizons. A person may pursue long-term education goals whilst adapting to immediate challenges, and these different time scales cohere into a unified, purposeful trajectory. This temporal integration is largely absent from current AI systems, which optimise for immediate reward signals or fixed objectives.
  5. Learning from sparse feedback: Humans learn from remarkably little data. A child sees a dog once or twice and thereafter recognises dogs in novel contexts, even in stylised drawings or unfamiliar breeds. This learning from sparse examples contrasts sharply with AI systems requiring thousands or millions of examples to achieve equivalent recognition.

This multifaceted robustness is what Sutskever identifies as “staggering”—not because it is strong but because it operates across so many dimensions simultaneously whilst remaining stable, efficient, and purposeful.

The Fragility of Current AI: Why Models Break

The contrast becomes clear when examining where current AI systems are fragile. Sutskever frequently illustrates this through the “jagged behaviour” problem: models that perform superhuman on benchmarks yet fail in elementary ways during real-world deployment.

A language model can score in the 88th percentile on the bar examination yet, when asked to debug code, introduces new errors whilst fixing previous ones. It cycles between mistakes even when provided clear feedback. It lacks the internal evaluative sense that tells a human programmer, “This approach is leading nowhere; I should try something different.” The model lacks robust value functions—internal signals that guide learning and action.

This fragility manifests across multiple dimensions:

  1. Distribution shift fragility: Models trained on one distribution of data often fail dramatically when confronted with data that differs from training distribution, even slightly. A vision system trained on images with certain lighting conditions fails on images with different lighting. A language model trained primarily on Western internet text struggles with cultural contexts it has not heavily encountered. Humans, by contrast, maintain competence across remarkable variation—different languages, accents, cultural contexts, lighting conditions, perspectives.
  2. Benchmark overfitting: Contemporary AI systems achieve extraordinary performance on carefully constructed evaluation tasks yet fail at the underlying capability the benchmark purports to measure. This occurs because models have been optimised (through reinforcement learning) specifically to perform well on benchmarks rather than to develop robust understanding. Sutskever has noted that this reward hacking is often unintentional—companies genuinely seeking to improve models inadvertently create RL environments that optimise for benchmark performance rather than genuine capability.
  3. Lack of principled abstraction: Models often memorise patterns rather than developing principled understanding. This manifests as inability to apply learned knowledge to genuinely novel contexts. A model may solve thousands of addition problems yet fail on a slightly different formulation it has not encountered. A human, having understood addition as a principle, applies it to any context where addition is relevant.
  4. Absence of internal feedback mechanisms: Current reinforcement learning typically provides feedback only at the end of long trajectories. A model can pursue 1,000 steps of reasoning down an unpromising path, only to receive a training signal after the trajectory completes. Humans, by contrast, possess continuous internal feedback—emotions, intuition, confidence levels—that signal whether reasoning is productive or should be redirected. This enables far more efficient learning.

The Value Function Hypothesis: Emotions as Robust Learning Machinery

Sutskever’s analysis points toward a crucial hypothesis: human robustness depends fundamentally on value functions—internal mechanisms that provide continuous, robust evaluation of states and actions.

In machine learning, a value function is a learned estimate of expected future reward or utility from a given state. In human neurobiology, value functions are implemented, Sutskever argues, through emotions and affective states. Fear signals danger. Confidence signals competence. Boredom signals that current activity is unproductive. Satisfaction signals that effort has succeeded. These emotional states, which evolution has refined over millions of years, serve as robust evaluative signals that guide learning and behaviour.

Sutskever illustrates this with a striking neurological case: a person who suffered brain damage affecting emotional processing. Despite retaining normal IQ, puzzle-solving ability, and articulate cognition, this person became radically incapable of making even trivial decisions. Choosing which socks to wear would take hours. Financial decisions became catastrophically poor. This person could think but could not effectively decide or act—suggesting that emotions (and the value functions they implement) are not peripheral to human cognition but absolutely central to effective agency.

What makes human value functions particularly robust is their simplicity and stability. They are not learned during a person’s lifetime through explicit training. They are evolved, hard-coded by billions of years of biological evolution into neural structures that remain remarkably consistent across human populations and contexts. A person experiences hunger, fear, social connection, and achievement similarly whether in ancient hunter-gatherer societies or modern industrial ones—because these value functions were shaped by evolutionary pressures that remained relatively stable.

This evolutionary hardcoding of value functions may be crucial to human learning robustness. Imagine trying to teach a child through explicit reward signals alone: “Do this task and receive points; optimise for points.” This would be inefficient and brittle. Instead, humans learn through value functions that are deeply embedded, emotionally weighted, and robust across contexts. A child learns to speak not through external reward optimisation but through intrinsic motivation—social connection, curiosity, the inherent satisfaction of communication. These motivations persist across contexts and enable robust learning.

Current AI systems largely lack this. They optimise for explicitly defined reward signals or benchmark metrics. These are fragile by comparison—vulnerable to reward hacking, overfitting, distribution shift, and the brittle transfer failures Sutskever observes.

Why This Matters Now: The Transition Point

Sutskever’s observation about human robustness arrives at a precise historical moment. As of November 2025, the AI industry is transitioning from what he terms the “age of scaling” (2020–2025) to what will be the “age of research” (2026 onward). This transition is driven by recognition that scaling alone is reaching diminishing returns. The next advances will require fundamental breakthroughs in understanding how to build systems that learn and adapt robustly—like humans do.

This creates an urgent research agenda: How do you build AI systems that possess human-like robustness? This is not a question that scales with compute or data. It is a research question—requiring new architectures, learning algorithms, training procedures, and conceptual frameworks.

Sutskever’s identification of robustness as the key distinguishing feature of human learning sets the research direction for the next phase of AI development. The question is not “how do we make bigger models” but “how do we build systems with value functions that enable efficient, self-correcting, context-robust learning?”

The Research Frontier: Leading Theorists Addressing Robustness

Antonio Damasio: The Somatic Marker Hypothesis

Antonio Damasio, neuroscientist at USC and authority on emotion and decision-making, has developed the somatic marker hypothesis—a framework explaining how emotions serve as rapid evaluative signals that guide decisions and learning. Damasio’s work provides neuroscientific grounding for Sutskever’s hypothesis that value functions (implemented as emotions) are central to effective agency. Damasio’s case studies of patients with emotional processing deficits closely parallel Sutskever’s neurological example—demonstrating that emotional value functions are prerequisites for robust, adaptive decision-making.

Judea Pearl: Causal Models and Robust Reasoning

Judea Pearl, pioneer in causal inference and probabilistic reasoning, has argued that correlation-based learning has fundamental limits and that robust generalisation requires learning causal structure—the underlying relationships between variables that remain stable across contexts. Pearl’s work suggests that human robustness derives partly from learning causal models rather than mere patterns. When a human understands how something works (causally), that understanding transfers to novel contexts. Current AI systems, lacking robust causal models, fail at transfer—a key component of robustness.

Karl Friston: The Free Energy Principle

Karl Friston, neuroscientist at University College London, has developed the free energy principle—a unified framework explaining how biological systems, including humans, maintain robustness by minimising prediction error and maintaining models of their environment and themselves. The principle suggests that what makes humans robust is not fixed programming but a general learning mechanism that continuously refines internal models to reduce surprise. This framework has profound implications for building robust AI: rather than optimising for external rewards, systems should optimise for maintaining accurate models of reality, enabling principled generalisation.

Stuart Russell: Learning Under Uncertainty and Value Alignment

Stuart Russell, UC Berkeley’s leading AI safety researcher, has emphasised that robust AI systems must remain genuinely uncertain about objectives and learn from interaction rather than operating under fixed goal specifications. Russell’s work suggests that rigidity about objectives makes systems fragile—vulnerable to reward hacking and context-specific failure. Robustness requires systems that maintain epistemic humility and adapt their understanding of what matters based on continued learning. This directly parallels how human value systems are robust: they are not brittle doctrines but evolving frameworks that integrate experience.

Demis Hassabis and DeepMind’s Continual Learning Research

Demis Hassabis, CEO of DeepMind, has invested substantial effort into systems that learn continuously from environmental interaction rather than through discrete offline training phases. DeepMind’s research on continual reinforcement learning, meta-learning, and adaptive systems reflects the insight that robustness emerges not from static pre-training but from ongoing interaction with environments—enabling systems to refine their models and value functions continuously. This parallels human learning, which is fundamentally continual rather than episodic.

Yann LeCun: Self-Supervised Learning and World Models

Yann LeCun, Meta’s Chief AI Scientist, has advocated for learning approaches that enable systems to build internal models of how the world works—what he terms world models—through self-supervised learning. LeCun argues that robust generalisation requires systems that understand causal structure and dynamics, not merely correlations. His work on self-supervised learning suggests that systems trained to predict and model their environments develop more robust representations than systems optimised for specific external tasks.

The Evolutionary Basis: Why Humans Have Robust Value Functions

Understanding human robustness requires appreciating why evolution equipped humans with sophisticated, stable value function systems.

For millions of years, humans and our ancestors faced fundamentally uncertain environments. The reward signals available—immediate sensory feedback, social acceptance, achievement, safety—needed to guide learning and behaviour across vast diversity of contexts. Evolution could not hard-code specific solutions for every possible situation. Instead, it encoded general-purpose value functions—emotions and motivational states—that would guide adaptive behaviour across contexts.

Consider fear. Fear is a robust value function signal that something is dangerous. This signal evolved in environments full of predators and hazards. Yet the same fear response that protected ancestral humans from predators also keeps modern humans safe from traffic, heights, and social rejection. The value function is robust because it operates on a general principle—danger—rather than specific memorised hazards.

Similarly, social connection, curiosity, achievement, and other human motivations evolved as general-purpose signals that, across millions of years, correlated with survival and reproduction. They remain remarkably stable across radically different modern contexts—different cultures, technologies, and social structures—because they operate at a level of abstraction robust to context change.

Current AI systems, by contrast, lack this evolutionary heritage. They are trained from scratch, often on specific tasks, with reward signals explicitly engineered for those tasks. These reward signals are fragile by comparison—vulnerable to distribution shift, overfitting, and context-specificity.

Implications for Safe AI Development

Sutskever’s emphasis on human robustness carries profound implications for safe AI development. Robust systems are safer systems. A system with genuine value functions—robust internal signals about what matters—is less vulnerable to reward hacking, specification gaming, or deployment failures. A system that learns continuously and maintains epistemic humility is more likely to remain aligned as its capabilities increase.

Conversely, current AI systems’ lack of robustness is dangerous. Systems optimised for narrow metrics can fail catastrophically when deployed in novel contexts. Systems lacking robust value functions cannot self-correct or maintain appropriate caution. Systems that cannot learn from deployment feedback remain brittle.

Building AI systems with human-like robustness is therefore not merely an efficiency question—though efficiency matters greatly. It is fundamentally a safety question. The development of robust value functions, continual learning capabilities, and general-purpose evaluative mechanisms is central to ensuring that advanced AI systems remain beneficial as they become more powerful.

The Research Direction: From Scaling to Robustness

Sutskever’s observation that “the robustness of people is really staggering” reorients the entire research agenda. The question is no longer primarily “how do we scale?” but “how do we build systems with robust value functions, efficient learning, and genuine adaptability across contexts?”

This requires:

  • Architectural innovation: New neural network structures that embed or can learn robust evaluative mechanisms—value functions analogous to human emotions.
  • Training methodology: Learning procedures that enable systems to develop genuine self-correction capabilities, learn from sparse feedback, and maintain robustness across distribution shift.
  • Theoretical understanding: Deeper mathematical and conceptual frameworks explaining what makes value functions robust and how to implement them in artificial systems.
  • Integration of findings from neuroscience, evolutionary biology, and decision theory: Drawing on multiple fields to understand the principles underlying human robustness and translating them into machine learning.

Conclusion: Robustness as the Frontier

When Sutskever identifies human robustness as “staggering,” he is not offering admiration but diagnosis. He is pointing out that current AI systems fundamentally lack what makes humans effective learners: robust value functions, efficient learning from sparse feedback, genuine self-correction, and adaptive generalisation across contexts.

The next era of AI research—the age of research beginning in 2026—will be defined largely by attempts to solve this problem. The organisation or research group that successfully builds AI systems with human-like robustness will not merely have achieved technical progress. They will have moved substantially closer to systems that learn efficiently, generalise reliably, and remain aligned to human values even as they become more capable.

Human robustness is not incidental. It is fundamental—the quality that makes human learning efficient, adaptive, and safe. Replicating it in artificial systems represents the frontier of AI research and development.

read more
Quote: Ilya Sutskever – Safe Superintelligence

Quote: Ilya Sutskever – Safe Superintelligence

“These models somehow just generalize dramatically worse than people. It’s super obvious. That seems like a very fundamental thing.” – Ilya Sutskever – Safe Superintelligence

Sutskever, as co-founder and Chief Scientist of Safe Superintelligence Inc. (SSI), has emerged as one of the most influential voices in AI strategy and research direction. His trajectory illustrates the depth of his authority: co-author of AlexNet (2012), the paper that ignited the deep learning revolution; Chief Scientist at OpenAI during the development of GPT-2 and GPT-3; and now directing a $3 billion research organisation explicitly committed to solving the generalisation problem rather than pursuing incremental scaling.

His assertion about generalisation deficiency is not rhetorical flourish. It represents a fundamental diagnostic claim about why current AI systems, despite superhuman performance on benchmarks, remain brittle, unreliable, and poorly suited to robust real-world deployment. Understanding this claim requires examining what generalisation actually means, why it matters, and what the gap between human and AI learning reveals about the future of artificial intelligence.

What Generalisation Means: Beyond Benchmark Performance

Generalisation, in machine learning, refers to the ability of a system to apply knowledge learned in one context to novel, unfamiliar contexts it has not explicitly encountered during training. A model that generalises well can transfer principles, patterns, and capabilities across domains. A model that generalises poorly becomes a brittle specialist—effective within narrow training distributions but fragile when confronted with variation, novelty, or real-world complexity.

The crisis Sutskever identifies is this: contemporary large language models and frontier AI systems achieve extraordinary performance on carefully curated evaluation tasks and benchmarks. GPT-4 scores in the 88th percentile of the bar exam. O1 solves competition mathematics problems at elite levels. Yet these same systems, when deployed into unconstrained real-world workflows, exhibit what Sutskever terms “jagged” behaviour—they repeat errors, introduce new bugs whilst fixing previous ones, cycle between mistakes even with clear corrective feedback, and fail in ways that suggest fundamentally incomplete understanding rather than mere data scarcity.

This paradox reveals a hidden truth: benchmark performance and deployment robustness are not tightly coupled. An AI system can memorise, pattern-match, and perform well on evaluation metrics whilst failing to develop the kind of flexible, transferable understanding that enables genuine competence.

The Sample Efficiency Question: Orders of Magnitude of Difference

Underlying the generalisation crisis is a more specific puzzle: sample efficiency. Why does it require vastly more training data for AI systems to achieve competence in a domain than it takes humans?

A human child learns to recognise objects through a few thousand exposures. Contemporary vision models require millions. A teenager learns to drive in approximately ten hours of practice; AI systems struggle to achieve equivalent robustness with orders of magnitude more training. A university student learns to code, write mathematically, and reason about abstract concepts—domains that did not exist during human evolutionary history—with remarkably few examples and little explicit feedback.

This disparity points to something fundamental: humans possess not merely better priors or more specialised knowledge, but better general-purpose learning machinery. The principle underlying human learning efficiency remains largely unexpressed in mathematical or computational terms. Current AI systems lack it.

Sutskever’s diagnostic claim is that this gap reflects not engineering immaturity or the need for more compute, but the absence of a conceptual breakthrough—a missing principle of how to build systems that learn as efficiently as humans do. The implication is stark: you cannot scale your way out of this problem. More data and more compute, applied to existing methodologies, will not solve it. The bottleneck is epistemic, not computational.

Why Current Models Fail at Generalisation: The Competitive Programming Analogy

Sutskever illustrates the generalisation problem through an instructive analogy. Imagine two competitive programmers:

Student A dedicates 10,000 hours to competitive programming. They memorise every algorithm, every proof technique, every problem pattern. They become exceptionally skilled within competitive programming itself—one of the very best.

Student B spends only 100 hours on competitive programming but develops deeper, more flexible understanding. They grasp underlying principles rather than memorising solutions.

When both pursue careers in software engineering, Student B typically outperforms Student A. Why? Because Student A has optimised for a narrow domain and lacks the flexible transfer of understanding that Student B developed through lighter but more principled engagement.

Current frontier AI models, in Sutskever’s assessment, resemble Student A. They are trained on enormous quantities of narrowly curated data—competitive programming problems, benchmark evaluation tasks, reinforcement learning environments explicitly designed to optimise for measurable performance. They have been “over-trained” on carefully optimised domains but lack the flexible, generalised understanding that enables robust performance in novel contexts.

This over-optimisation problem is compounded by a subtle but crucial factor: reinforcement learning optimisation targets. Companies designing RL training environments face substantial degrees of freedom in how to construct reward signals. Sutskever observes that there is often a systematic bias: RL environments are subtly shaped to ensure models perform well on public benchmarks at release time, creating a form of unintentional reward hacking where the system becomes highly tuned to evaluation metrics rather than genuinely robust to real-world variation.

The Deeper Problem: Pre-Training’s Limits and RL’s Inefficiency

The generalisation crisis reflects deeper structural issues within contemporary AI training paradigms.

Pre-training’s opacity: Large-scale language model pre-training—trained on internet text data—provides models with an enormous foundation of patterns. Yet the way models rely on this pre-training data is poorly understood. When a model fails, it is unclear whether the failure reflects insufficient statistical support in the training distribution or whether something more fundamental is missing. Pre-training provides scale but at the cost of reasoning about what has actually been learned.

RL’s inefficiency: Current reinforcement learning approaches provide training signals only at the end of long trajectories. If a model spends thousands of steps reasoning about a problem and arrives at a dead end, it receives no signal until the trajectory completes. This is computationally wasteful. A more efficient learning system would provide intermediate evaluative feedback—signals that say, “this direction of reasoning is unpromising; abandon it now rather than after 1,000 more steps.” Sutskever hypothesises that this intermediate feedback mechanism—what he terms a “value function” and what evolutionary biology has encoded as emotions—is crucial to sample-efficient learning.

The gap between how humans and current AI systems learn suggests that human learning operates on fundamentally different principles: continuous, intermediate evaluation; robust internal models of progress and performance; the ability to self-correct and redirect effort based on internal signals rather than external reward.

Generalisation as Proof of Concept: What Human Learning Reveals

A critical move in Sutskever’s argument is this: the fact that humans generalise vastly better than current AI systems is not merely an interesting curiosity—it is proof that better generalisation is achievable. The existence of human learners demonstrates, in principle, that a learning system can operate with orders of magnitude less data whilst maintaining superior robustness and transfer capability.

This reframes the research challenge. The question is no longer whether better generalisation is possible (humans prove it is) but rather what principle or mechanism underlies it. This principle could arise from:

  • Architectural innovations: new ways of structuring neural networks that embody better inductive biases for generalisation
  • Learning algorithms: different training procedures that more efficiently extract principles from limited data
  • Value function mechanisms: intermediate feedback systems that enable more efficient learning trajectories
  • Continual learning frameworks: systems that learn continuously from interaction rather than through discrete offline training phases

What matters is that Sutskever’s claim shifts the research agenda from “get more compute” to “discover the missing principle.”

The Strategic Implications: Why This Matters Now

Sutskever’s diagnosis, articulated in November 2025, arrives at a crucial moment. The AI industry has operated under the “age of scaling” paradigm since approximately 2020. During this period, the scaling laws discovered by OpenAI and others suggested a remarkably reliable relationship: larger models trained on more data with more compute reliably produced better performance.

This created a powerful strategic imperative: invest capital in compute, acquire data, build larger systems. The approach was low-risk from a research perspective because the outcome was relatively predictable. Companies could deploy enormous resources confident they would yield measurable returns.

By 2025, however, this model shows clear strain. Data is approaching finite limits. Computational resources, whilst vast, are not unlimited, and marginal returns diminish. Most importantly, the question has shifted: would 100 times more compute actually produce a qualitative transformation or merely incremental improvement? Sutskever’s answer is clear: the latter. This fundamentally reorients strategic thinking. If 100x scaling yields only incremental gains, the bottleneck is not compute but ideas. The competitive advantage belongs not to whoever can purchase the most GPUs but to whoever discovers the missing principle of generalisation.

Leading Theorists and Related Research Programs

Yann LeCun: World Models and Causal Learning

Yann LeCun, Meta’s Chief AI Scientist and a pioneer of deep learning, has long emphasized that current supervised learning approaches are fundamentally limited. His work on “world models”—internal representations that capture causal structure rather than mere correlation—points toward learning mechanisms that could enable better generalisation. LeCun’s argument is that humans learn causal models of how the world works, enabling robust generalisation because causal understanding is stable across contexts in a way that statistical correlation is not.

Geoffrey Hinton: Neuroscience-Inspired Learning

Geoffrey Hinton, recipient of the 2024 Nobel Prize in Physics for foundational deep learning work, has increasingly emphasized that neuroscience holds crucial clues for improving AI learning efficiency. His recent work on biological plausibility and learning mechanisms reflects conviction that important principles of how neural systems efficiently extract generalised understanding remain undiscovered. Hinton has expressed support for Sutskever’s research agenda, recognizing that the next frontier requires fundamental conceptual breakthroughs rather than incremental scaling.

Stuart Russell: Learning Under Uncertainty

Stuart Russell, UC Berkeley’s leading AI safety researcher, has articulated that robust AI alignment requires systems that remain genuinely uncertain about objectives and learn from interaction. This aligns with Sutskever’s emphasis on continual learning. Russell’s work highlights that systems designed to optimise fixed objectives without capacity for ongoing learning and adjustment tend to produce brittle, misaligned outcomes—a dynamic that improves when systems maintain epistemic humility and learn continuously.

Demis Hassabis and DeepMind’s Continual Learning Research

Demis Hassabis, CEO of DeepMind, has invested substantial research effort into systems that learn continually from environmental interaction rather than through discrete offline training phases. DeepMind’s work on continual reinforcement learning, meta-learning, and systems that adapt to new tasks reflects recognition that learning efficiency depends on how feedback is structured and integrated over time—not merely on total data quantity.

Judea Pearl: Causality and Abstraction

Judea Pearl, pioneering researcher in causal inference and probabilistic reasoning, has long argued that correlation-based learning has fundamental limits and that causal reasoning is necessary for genuine understanding and generalisation. His work on causal models and graphical representation of dependencies provides theoretical foundations for why systems that learn causal structure (rather than mere patterns) achieve better generalisation across domains.

The Research Agenda Going Forward

Sutskever’s claim that generalisation is the “very fundamental thing” reorients the entire research agenda. This shift has profound implications:

From scaling to methodology: Research emphasis moves from “how do we get more compute” to “what training procedures, architectural innovations, or learning algorithms enable human-like generalisation?”

From benchmarks to robustness: Evaluation shifts from benchmark performance to deployment reliability—how systems perform on novel, unconstrained tasks rather than carefully curated evaluations.

From monolithic pre-training to continual learning: The training paradigm shifts from discrete offline phases (pre-train, then RL, then deploy) toward systems that learn continuously from real-world interaction.

From scale as differentiator to ideas as differentiator: Competitive advantage in AI development becomes less about resource concentration and more about research insight—the organisation that discovers better generalisation principles gains asymmetric advantage.

The Deeper Question: What Humans Know That AI Doesn’t

Beneath Sutskever’s diagnostic claim lies a profound question: What do humans actually know about learning that AI systems don’t yet embody?

Humans learn efficiently because they:

  • Develop internal models of their own performance and progress (value functions)
  • Self-correct through continuous feedback rather than awaiting end-of-trajectory rewards
  • Transfer principles flexibly across domains rather than memorising domain-specific patterns
  • Learn from remarkably few examples through principled understanding rather than statistical averaging
  • Integrate feedback across time scales and contexts in ways that build robust, generalised knowledge

These capabilities do not require superhuman intelligence or extraordinary cognitive resources. A fifteen-year-old possesses them. Yet current AI systems, despite vastly larger parameter counts and more data, lack equivalent ability.

This gap is not accidental. It reflects that current AI development has optimised for the wrong targets—benchmark performance rather than genuine generalisation, scale rather than efficiency, memorisation rather than principled understanding. The next breakthrough requires not more of the same but fundamentally different approaches.

Conclusion: The Shift from Scaling to Discovery

Sutskever’s assertion that “these models somehow just generalize dramatically worse than people” is, at first glance, an observation of inadequacy. But reframed, it is actually a statement of profound optimism about what remains to be discovered. The fact that humans achieve vastly better generalisation proves that better generalisation is possible. The task ahead is not to accept poor generalisation as inevitable but to discover the principle that enables human-like learning efficiency.

This diagnostic shift—from “we need more compute” to “we need better understanding of generalisation”—represents the intellectual reorientation of AI research in 2025 and beyond. The age of scaling is ending not because scaling is impossible but because it has approached its productive limits. The age of research into fundamental learning principles is beginning. What emerges from this research agenda may prove far more consequential than any previous scaling increment.

read more
Quote: Ilya Sutskever – Safe Superintelligence

Quote: Ilya Sutskever – Safe Superintelligence

“Is the belief really, ‘Oh, it’s so big, but if you had 100x more, everything would be so different?’ It would be different, for sure. But is the belief that if you just 100x the scale, everything would be transformed? I don’t think that’s true. So it’s back to the age of research again, just with big computers.” – Ilya Sutskever – Safe Superintelligence

Ilya Sutskever stands as one of the most influential figures in modern artificial intelligence—a scientist whose work has fundamentally shaped the trajectory of deep learning over the past decade. As co-author of the seminal 2012 AlexNet paper, he helped catalyse the deep learning revolution that transformed machine vision and launched the contemporary AI era. His influence extends through his role as Chief Scientist at OpenAI, where he played a pivotal part in developing GPT-2 and GPT-3, the models that established large-scale language model pre-training as the dominant paradigm in AI research.

In late 2024, Sutskever departed OpenAI and co-founded Safe Superintelligence Inc. (SSI) alongside Daniel Gross and Daniel Levy, positioning the company as the world’s “first straight-shot SSI lab”—an organisation with a single focus: developing safe superintelligence without distraction from product development or revenue generation. The company has since raised $3 billion and reached a $32 billion valuation, reflecting investor confidence in Sutskever’s strategic vision and reputation.

The Context: The Exhaustion of Scaling

Sutskever’s quoted observation emerges from a moment of genuine inflection in AI development. For roughly five years—from 2020 to 2025—the AI industry operated under what he terms the “age of scaling.” This era was defined by a simple, powerful insight: that scaling pre-training data, computational resources, and model parameters yielded predictable improvements in model performance. Organisations could invest capital with low perceived risk, knowing that more compute plus more data plus larger models would reliably produce measurable gains.

This scaling paradigm was extraordinarily productive. It yielded GPT-3, GPT-4, and an entire generation of frontier models that demonstrated capabilities that astonished both researchers and the public. The logic was elegant: if you wanted better AI, you simply scaled the recipe. Sutskever himself was instrumental in validating this approach. The word “scaling” became conceptually magnetic, drawing resources, attention, and organisational focus toward a single axis of improvement.

Yet by 2024–2025, that era began showing clear signs of exhaustion. Data is finite—the amount of high-quality training material available on the internet is not infinite, and organisations are rapidly approaching meaningful constraints on pre-training data supply. Computational resources, whilst vast, are not unlimited, and the economic marginal returns on compute investment have become less obvious. Most critically, the empirical question has shifted: if current frontier labs have access to extraordinary computational resources, would 100 times more compute actually produce a qualitative transformation in capabilities, or merely incremental improvement?

Sutskever’s answer is direct: incremental, not transformative. This reframing is consequential because it redefines where the bottleneck actually lies. The constraint is no longer the ability to purchase more GPUs or accumulate more data. The constraint is ideas—novel technical approaches, new training methodologies, fundamentally different recipes for building AI systems.

The Jaggedness Problem: Theory Meeting Reality

One critical observation animates Sutskever’s thinking: a profound disconnect between benchmark performance and real-world robustness. Current models achieve superhuman performance on carefully constructed evaluation tasks—yet in deployment, they exhibit what Sutskever calls “jagged” behaviour. They repeat errors, introduce new bugs whilst fixing old ones, and cycle between mistakes even when given clear corrective feedback.

This apparent paradox suggests something deeper than mere data or compute insufficiency. It points to inadequate generalisation—the inability to transfer learning from narrow, benchmark-optimised domains into the messy complexity of real-world application. Sutskever frames this through an analogy: a competitive programmer who practises 10,000 hours on competition problems will be highly skilled within that narrow domain but often fails to transfer that knowledge flexibly to broader engineering challenges. Current models, in his assessment, resemble that hyper-specialised competitor rather than the flexible, adaptive learner.

The Core Insight: Generalisation Over Scale

The central thesis animating Sutskever’s work at SSI—and implicit in his quote—is that human-like generalisation and learning efficiency represent a fundamentally different ML principle than scaling, one that has not yet been discovered or operationalised within contemporary AI systems.

Humans learn with orders of magnitude less data than large models yet generalise far more robustly to novel contexts. A teenager learns to drive in roughly ten hours of practice; current AI systems struggle to acquire equivalent robustness with vastly more training data. This is not because humans possess specialised evolutionary priors for driving (a recent activity that evolution could not have optimized for); rather, it suggests humans employ a more general-purpose learning principle that contemporary AI has not yet captured.

Sutskever hypothesises that this principle is connected to what he terms “value functions”—internal mechanisms akin to emotions that provide continuous, intermediate feedback on actions and states, enabling more efficient learning than end-of-trajectory reward signals alone. Evolution appears to have hard-coded robust value functions—emotional and evaluative systems—that make humans viable, adaptive agents across radically different environments. Whether an equivalent principle can be extracted purely from pre-training data, rather than built into learning architecture, remains uncertain.

The Leading Theorists and Related Work

Yann LeCun and Data Efficiency

Yann LeCun, Meta’s Chief AI Scientist and a pioneer of deep learning, has long emphasised the importance of learning efficiency and the role of what he terms “world models” in understanding how agents learn causal structure from limited data. His work highlights that human vision achieves remarkable robustness from developmental data scarcity—children recognise cars after seeing far fewer exemplars than AI systems require—suggesting that the brain employs inductive biases or learning principles that current architectures lack.

Geoffrey Hinton and Neuroscience-Inspired AI

Geoffrey Hinton, winner of the 2024 Nobel Prize in Physics for his work on deep learning, has articulated concerns about AI safety and expressed support for Sutskever’s emphasis on fundamentally rethinking how AI systems learn and align. Hinton’s career-long emphasis on biologically plausible learning mechanisms—from Boltzmann machines to capsule networks—reflects a conviction that important principles for efficient learning remain undiscovered and that neuroscience offers crucial guidance.

Stuart Russell and Alignment Through Uncertainty

Stuart Russell, UC Berkeley’s leading AI safety researcher, has emphasised that robust AI alignment requires systems that remain genuinely uncertain about human values and continue learning from interaction, rather than attempting to encode fixed objectives. This aligns with Sutskever’s thesis that safe superintelligence requires continual learning in deployment rather than monolithic pre-training followed by fixed RL optimisation.

Demis Hassabis and Continual Learning

Demis Hassabis, CEO of DeepMind and a co-developer of AlphaGo, has invested significant research effort into systems that learn continually rather than through discrete training phases. This work recognises that biological intelligence fundamentally involves interaction with environments over time, generating diverse signals that guide learning—a principle SSI appears to be operationalising.

The Paradigm Shift: From Offline to Online Learning

Sutskever’s thinking reflects a broader intellectual shift visible across multiple frontiers of AI research. The dominant pre-training + RL framework assumes a clean separation: a model is trained offline on fixed data, then post-trained with reinforcement learning, then deployed. Increasingly, frontier researchers are questioning whether this separation reflects how learning should actually work.

His articulation of “age of research” signals a return to intellectual plurality and heterodox experimentation—the opposite of the monoculture that scaling paradigm created. When everyone is racing to scale the same recipe, innovation becomes incremental. When new recipes are required, diversity of approach becomes an asset rather than liability.

The Stakes and Implications

This reframing carries significant strategic implications. If the bottleneck is truly ideas rather than compute, then smaller, more cognitively coherent organisations with clear intellectual direction may outpace larger organisations constrained by product commitments, legacy systems, and organisational inertia. If the key innovation is a new training methodology—one that achieves human-like generalisation through different mechanisms—then the first organisation to discover and validate it may enjoy substantial competitive advantage, not through superior resources but through superior understanding.

Equally, this framing challenges the common assumption that AI capability is primarily a function of computational spend. If methodological innovation matters more than scale, the future of AI leadership becomes less a question of capital concentration and more a question of research insight—less about who can purchase the most GPUs, more about who can understand how learning actually works.

Sutskever’s quote thus represents not merely a rhetorical flourish but a fundamental reorientation of strategic thinking about AI development. The age of confident scaling is ending. The age of rigorous research into the principles of generalisation, sample efficiency, and robust learning has begun.

read more
Quote: Dr Eric Schmidt – Ex-Google CEO

Quote: Dr Eric Schmidt – Ex-Google CEO

“The win will be teaming between a human and their judgment and a supercomputer and what it can think.” – Dr Eric Schmidt – Former Google CEO

Dr Eric Schmidt is recognised globally as a principal architect of the modern digital era. He served as CEO of Google from 2001 to 2011, guiding its evolution from a fast-growing startup into a cornerstone of the tech industry. His leadership was instrumental in scaling Google’s infrastructure, accelerating product innovation, and instilling a model of data-driven culture that underpins contemporary algorithms and search technologies. After stepping down as CEO, Schmidt remained pivotal as Executive Chairman and later as Technical Advisor, shepherding Google’s transition to Alphabet and advocating for long-term strategic initiatives in AI and global connectivity.

Schmidt’s influence extends well beyond corporate leadership. He has played policy-shaping roles at the highest levels, including chairing the US National Security Commission on Artificial Intelligence and advising multiple governments on technology strategy. His career is marked by a commitment to both technical progress and the responsible governance of innovation, positioning him at the centre of debates on AI’s promises, perils, and the necessity of human agency in the face of accelerating machine intelligence.

Context of the Quotation: Human–AI Teaming

Schmidt’s statement emerged during high-level discussions about the trajectory of AI, particularly in the context of autonomous systems, advanced agents, and the potential arrival of superintelligent machines. Rather than portraying AI as a force destined to replace humans, Schmidt advocates a model wherein the greatest advantage arises from joint endeavour: humans bring creativity, ethical discernment, and contextual understanding, while supercomputers offer vast capacity for analysis, pattern recognition, and iterative reasoning.

This principle is visible in contemporary AI deployments. For example:

  • In drug discovery, AI systems can screen millions of molecular variants in a day, but strategic insights and hypothesis generation depend on human researchers.
  • In clinical decision-making, AI augments the observational scope of physicians—offering rapid, precise diagnoses—but human judgement is essential for nuanced cases and values-driven choices.
  • Schmidt points to future scenarios where “AI agents” conduct scientific research, write code by natural-language command, and collaborate across domains, yet require human partnership to set objectives, interpret outcomes, and provide oversight.
  • He underscores that autonomous AI agents, while powerful, must remain under human supervision, especially as they begin to develop their own procedures and potentially opaque modes of communication.

Underlying this vision is a recognition: AI is a multiplier, not a replacement, and the best outcomes will couple human judgement with machine cognition.

Relevant Leading Theorists and Critical Backstory

This philosophy of human–AI teaming aligns with and is actively debated by several leading theorists:

  • Stuart Russell
    Professor at UC Berkeley, Russell is renowned for his work on human-compatible AI. He contends that the long-term viability of artificial intelligence requires that systems are designed to understand and comply with human preferences and values. Russell has championed the view that human oversight and interpretability are non-negotiable as intelligence systems become more capable and autonomous.
  • Fei-Fei Li
    Stanford Professor and co-founder of AI4ALL, Fei-Fei Li is a major advocate for “human-centred AI.” Her research highlights that AI should augment human potential, not supplant it, and she stresses the critical importance of interdisciplinary collaboration. She is a proponent of AI systems that foster creativity, support decision-making, and preserve agency and dignity.
  • Demis Hassabis
    Founder and CEO of DeepMind, Hassabis’s group famously developed AlphaGo and AlphaFold. DeepMind’s work demonstrates the principle of human–machine teaming: AI systems solve previously intractable problems, such as protein folding, that can only be understood and validated with strong human scientific context.
  • Gary Marcus
    A prominent AI critic and academic, Marcus warns against overestimating current AI’s capacity for judgment and abstraction. He pursues hybrid models where symbolic reasoning and statistical learning are paired with human input to overcome the limitations of “black-box” models.
  • Eric Schmidt’s own contributions reflect active engagement with these paradigms, from his advocacy for AI regulatory frameworks to public warnings about the risks of unsupervised AI, including “unplugging” AI systems that operate beyond human understanding or control.

Structural Forces and Implications

Schmidt’s perspective is informed by several notable trends:

  • Expansion of infinite context windows: Models can now process millions of words and reason through intricate problems with humans guiding multi-step solutions, a paradigm shift for fields like climate research, pharmaceuticals, and engineering.
  • Proliferation of autonomous agents: AI agents capable of learning, experimenting, and collaborating independently across complex domains are rapidly becoming central; their effectiveness maximised when humans set goals and interpret results.
  • Democratisation paired with concentration of power: As AI accelerates innovation, the risk of centralised control emerges; Schmidt calls for international cooperation and proactive governance to keep objectives aligned with human interests.
  • Chain-of-thought reasoning and explainability: Advanced models can simulate extended problem-solving, but meaningful solutions depend on human guidance, interpretation, and critical thinking.

Summary

Eric Schmidt’s quote sits at the intersection of optimistic technological vision and pragmatic governance. It reflects decades of strategic engagement with digital transformation, and echoes leading theorists’ consensus: the future of AI is collaborative, and its greatest promise lies in amplifying human judgment with unprecedented computational support. Realising this future will depend on clear policies, interdisciplinary partnership, and an unwavering commitment to ensuring technology remains a tool for human advancement—and not an unfettered automaton beyond our reach.

read more

Download brochure

Introduction brochure

What we do, case studies and profiles of some of our amazing team.

Download

Our latest podcasts on Spotify

Sign up for our newsletters - free

Global Advisors | Quantified Strategy Consulting