Select Page

Global Advisors | Quantified Strategy Consulting

Artificial Intelligence
Quote: Jensen Huang – CEO Nvidia

Quote: Jensen Huang – CEO Nvidia

“Oftentimes, if you reason about things from first principles, what’s working today incredibly well — if you could reason about it from first principles and ask yourself on what foundation that first principle is built and how that would change over time — it allows you to hopefully see around corners.” – Jensen Huang – CEO Nvidia

Jensen Huang’s quote was delivered in the context of an in-depth dialogue with institutional investors on the trajectory of Nvidia, the evolution of artificial intelligence, and strategies for anticipating and shaping the technological future.

Context of the Quote

The quote was made during an interview at a Citadel Securities event in October 2025, hosted by Konstantine Buhler, a partner at Sequoia Capital. The dialogue’s audience consisted of leading institutional investors, all seeking avenues for sustainable advantage or ‘edge’. The conversation explored the founding moments of Nvidia in the early 1990s, through the reinvention of the graphics processing unit (GPU), the creation of new computing markets, and the subsequent rise of Nvidia as the platform underpinning the global AI boom. The question of how to ‘see around corners’ — to anticipate technology and industry shifts before they crystallise for others — was at the core of the discussion. Huang’s answer, invoking first-principles reasoning, linked Nvidia’s success to its ability to continually revisit and challenge foundational assumptions, and to methodically project how they will be redefined by progress in science and technology.

Jensen Huang: Profile and Approach

Jensen Huang, born in Tainan, Taiwan in 1963, immigrated to the United States as a child, experiencing the formative challenges of cultural dislocation, financial hardship, and adversity. He obtained his undergraduate degree in electrical engineering from Oregon State University and a master’s from Stanford University. After working at AMD and LSI Logic, he co-founded Nvidia in 1993 at 30, reportedly at a Denny’s restaurant. From the outset, the company faced daunting odds — neither established market nor assured funding, and frequent existential risk in the initial years.

Huang is distinguished not only by technical fluency — he is deeply involved in hardware and software architecture — but also by an ability to translate complexity for diverse audiences. He eschews corporate formality in favour of trademark leather jackets and a focus on product. His leadership style is marked by humility, a willingness to bet on emerging ideas, and what he describes as “urgent innovation” born of early near-failure. This disposition has been integral to Nvidia’s progress, especially as the company repeatedly “invented markets” and defined entirely new categories, such as accelerated computing and AI infrastructure.

By 2024, Nvidia became the world’s most valuable public company, with its GPUs foundational to gaming, scientific computing, and, critically, the rise of AI. Huang’s awards — from the IEEE Founder’s Medal to listing among Time Magazine’s 100 most influential — underscore his reputation as a technologist and strategic thinker. He is widely recognised for being able to establish technical direction well before it becomes market consensus, an approach reflected in the quote.

First-Principles Thinking: Theoretical Foundations

Huang’s endorsement of “first principles” echoes a method of problem-solving and innovation associated with thinkers as diverse as Aristotle, Isaac Newton, and, in the modern era, entrepreneurs and strategists such as Elon Musk. The essence of first-principles thinking is to break down complex systems to their most fundamental truths — concepts that cannot be deduced from anything simpler — and to reason forward from those axioms, unconstrained by traditional assumptions, analogies, or received wisdom.

  • Aristotle was the first to coin the term “first principles”, distinguishing knowledge derived from irreducible foundational truths from knowledge obtained through analogy or precedent.
  • René Descartes advocated for systematic doubt and logical rebuilding of knowledge from foundational elements.
  • Richard Feynman, the physicist, was famous for urging students to “understand from first principles”, encouraging deep understanding and avoidance of rote memorisation or mere pattern recognition.
  • Elon Musk is often cited as a contemporary example, applying first-principles thinking to industries as varied as automotive (Tesla), space (SpaceX), and energy. Musk has described the technique as “boiling things down to the most fundamental truths and then reasoning up from there,” directly influencing not just product architectures but also cost models and operational methods.

Application in Technology and AI

First-principles thinking is particularly powerful in periods of technological transition:

  • In computing, first principles were invoked by Carver Mead and Lynn Conway, who reimagined the semiconductor industry in the 1970s by establishing the foundational laws for microchip design, known as Mead-Conway methodology. This approach was cited by Huang as influential for predicting the physical limitations of transistor miniaturisation and motivating Nvidia’s focus on accelerated computing.
  • Clayton Christensen, cited by Huang as an influence, introduced the idea of disruptive innovation, arguing that market leaders must question incumbent logic and anticipate non-linear shifts in technology. His books on disruption and innovation strategy have shaped how leaders approach structural shifts and avoid the “innovator’s dilemma”.
  • The leap from von Neumann architectures to parallel, heterogenous, and ultimately AI-accelerated computing frameworks — as pioneered by Nvidia’s CUDA platform and deep learning libraries — was possible because leaders at Nvidia systematically revisited underlying assumptions about how computation should be structured for new workloads, rather than simply iterating on the status quo.
  • The AI revolution itself was catalysed by the “deep learning” paradigm, championed by Geoffrey Hinton, Yann LeCun, and Andrew Ng. Each demonstrated that previous architectures, which had reached plateaus, could be superseded by entirely new approaches, provided there was willingness to reinterpret the problem from mathematical and computational fundamentals.

Backstory of the Leading Theorists

The ecosystem that enabled Nvidia’s transformation is shaped by a series of foundational theorists:

  • Mead and Conway: Their 1979 textbook and methodologies codified the “first-principles” approach in chip design, allowing for the explosive growth of Silicon Valley’s fabless innovation model.
  • Gordon Moore: Moore’s Law, while originally an empirical observation, inspired decades of innovation, but its eventual slow-down prompted leaders such as Huang to look for new “first principles” to govern progress, beyond mere transistor scaling.
  • Clayton Christensen: His disruption theory is foundational in understanding why entire industries fail to see the next shift — and how those who challenge orthodoxy from first principles are able to “see around corners”.
  • Geoffrey Hinton, Yann LeCun, Andrew Ng: These pioneers directly enabled the deep learning revolution by returning to first principles on how learning — both human and artificial — could function at scale. Their work with neural networks, widely doubted after earlier “AI winters”, was vindicated with landmark results like AlexNet (2012), enabled by Nvidia GPUs.

Implications

Jensen Huang’s quote is neither idle philosophy nor abstract advice — it is a methodology proven repeatedly by his own journey and by the history of technology. It is a call to scrutinise assumptions, break complex structures to their most elemental truths, and reconstruct strategy consciously from the bedrock of what is not likely to change, but also to ask: on what foundation do these principles rest, and how will these foundations themselves evolve.

Organisations and individuals who internalise this approach are equipped not only to compete in current markets, but to invent new ones — to anticipate and shape the next paradigm, rather than reacting to it.

read more
Quote: Andrej Karpathy – Ex-OpenAI, Ex-Tesla AI

Quote: Andrej Karpathy – Ex-OpenAI, Ex-Tesla AI

“What I think we have to do going forward…is figure out ways to remove some of the knowledge and to keep what I call this cognitive core. It’s this intelligent entity that is stripped from knowledge but contains the algorithms and contains the magic of intelligence and problem-solving and the strategies of it and all this stuff.” – Andrej Karpathy – Ex-OpenAI, Ex-Tesla AI

Andrej Karpathy’s observation about the need to “strip away knowledge whilst retaining the cognitive core” represents one of the most penetrating insights into contemporary artificial intelligence development. Speaking on Dwarkesh Patel’s podcast in October 2025, Karpathy—formerly a leading figure at both OpenAI and Tesla’s Autopilot programme—articulated a fundamental tension at the heart of modern AI: the current generation of large language models have become prodigious memorisers, yet this very capability may be constraining their potential for genuine intelligence.

The Paradox of Pre-training

To comprehend Karpathy’s thesis, one must first understand the architecture of contemporary AI systems. Large language models are trained on vast corpora—often 15 trillion tokens or more—through a process called pre-training. During this phase, models learn to predict the next token in a sequence, effectively compressing the entire internet into their neural networks. Karpathy describes this compressed representation as only “0.07 bits per token” for a model like Llama 3 70B, highlighting the extraordinary degree of compression occurring.

This compression serves two distinct functions, which Karpathy carefully delineates. First, models accumulate factual knowledge—the content of Wikipedia articles, the specifics of historical events, the details of scientific papers. Second, and more crucially, they develop what Karpathy terms “algorithmic patterns”—the capacity for in-context learning, the ability to recognise and complete patterns, the fundamental mechanisms of reasoning itself.

The problem, as Karpathy sees it, is that contemporary models have become too adept at the former whilst the latter remains the true seat of intelligence. When a model can regurgitate passages verbatim or recite obscure facts, it demonstrates remarkable memory. But this same capability creates what he calls a “distraction”—the model becomes reliant on its hazy recollections of training data rather than developing robust reasoning algorithms that could operate independently of specific factual knowledge.

The Cognitive Core Concept

Karpathy’s proposed solution is to isolate and preserve what he terms the “cognitive core”—an intelligent entity stripped of encyclopaedic knowledge but retaining the fundamental algorithms of problem-solving, the strategies of thought, and what he describes as “the magic of intelligence.” This concept represents a profound shift in how we conceptualise artificial intelligence.

Consider the analogy to human cognition. Humans are remarkably poor memorisers compared to AI systems. Present a human with a random sequence of numbers, and they’ll struggle after seven or eight digits. Yet this apparent limitation forces humans to develop robust pattern-recognition capabilities and abstract reasoning skills. We’re compelled to “see the forest for the trees” precisely because we cannot memorise every individual tree.

Karpathy suggests that AI systems would benefit from similar constraints. A model with less memory but stronger reasoning capabilities would be forced to look up factual information whilst maintaining sophisticated algorithms for processing that information. Such a system would more closely resemble human intelligence—not in its limitations, but in the way those limitations drive the development of generalisable cognitive strategies.

The implications extend beyond mere technical architecture. Karpathy envisions cognitive cores as compact as one billion parameters—potentially even smaller—that could operate as genuine reasoning engines rather than glorified databases. These systems would “know that they don’t know” when confronted with factual questions, prompting them to retrieve information whilst applying sophisticated analysis. The result would be AI that thinks more than it remembers, that reasons rather than recites.

From Evolution to Engineering: The Path Not Taken

Karpathy’s perspective on AI development diverges sharply from what he calls the “Richard Sutton viewpoint”—the notion that we should build AI systems analogous to biological intelligence, learning from scratch through reinforcement learning in the manner of animals. Instead, Karpathy argues we’re building what he evocatively describes as “ghosts” or “spirit entities”—ethereal intelligences that emerge from imitating human-generated text rather than evolving through environmental interaction.

This distinction illuminates a crucial divergence in AI philosophy. Biological intelligence, as embodied in animals, emerges from evolution operating over millions of years, with vast amounts of capability “baked in” to neural circuitry. A zebra foal runs within minutes of birth not through reinforcement learning but through evolutionary encoding. Humans similarly arrive with substantial cognitive machinery pre-installed, with lifetime learning representing maturation and refinement rather than learning from first principles.

By contrast, contemporary AI systems learn through what Karpathy terms “crappy evolution”—pre-training on internet documents serves as a compressed, accelerated alternative to evolutionary optimisation. This process creates entities fundamentally different from biological intelligence, optimised for different tasks through different mechanisms. The current approach imitates the products of human intelligence (text, code, conversations) rather than replicating the developmental process that creates intelligence.

The Limits of Current Learning Paradigms

Karpathy’s critique extends to reinforcement learning, which he describes with characteristic bluntness as “terrible.” His concerns illuminate deep problems in how AI systems currently learn from experience. In reinforcement learning, a model generates hundreds of solution attempts, and those that arrive at correct answers have every intermediate step up-weighted, whilst failed attempts are down-weighted. Karpathy calls this “sucking supervision through a straw”—extracting minimal learning signal from vast amounts of computational work.

The fundamental issue is noise. When a solution works, not every step along the way was necessarily correct or optimal. The model may have taken wrong turns, pursued dead ends, or stumbled upon the answer despite flawed reasoning. Yet reinforcement learning broadcasts the final reward across the entire trajectory, reinforcing both good and bad reasoning indiscriminately. Karpathy notes that “you may have gone down the wrong alleys until you arrived at the right solution,” yet every mistaken step gets marked as something to do more of.

Humans, by contrast, engage in sophisticated post-hoc analysis. After solving a problem, we reflect on which approaches worked, which didn’t, and why. We don’t simply label an entire problem-solving session as “good” or “bad”—we dissect our reasoning, identify productive and unproductive strategies, and refine our approach. Current AI systems lack this reflective capacity entirely.

This limitation connects to broader questions about how AI systems might achieve continual learning—the ability to genuinely learn from ongoing experience rather than requiring massive retraining. Karpathy suggests that humans engage in a nightly “distillation phase” during sleep, processing the day’s experiences and integrating insights into long-term knowledge. AI systems have no equivalent mechanism. They simply restart from the same state each time, unable to evolve based on individual experiences.

Model Collapse and the Entropy Problem

A subtle but critical concern in Karpathy’s analysis is what he terms “model collapse”—the tendency of AI systems to produce outputs that occupy “a very tiny manifold of the possible space of thoughts.” Ask ChatGPT to tell a joke repeatedly, and you’ll receive the same three jokes. Request reflection on a topic multiple times, and you’ll observe striking similarity across responses. The models are “silently collapsed,” lacking the entropy and diversity that characterises human thought.

This phenomenon creates profound challenges for synthetic data generation, a technique labs use to create additional training material. If models generate training data for themselves or subsequent models, this collapsed distribution gradually dominates the training corpus. Training on one’s own outputs creates a dangerous feedback loop—each generation becomes less diverse, more stereotyped, more “collapsed” than the last. Karpathy suggests this may not even be a solvable problem, noting that humans similarly “collapse over time,” becoming more rigid and less creative as they age, revisiting the same thoughts and patterns with decreasing learning rates.

The contrast with children is illuminating. Young minds, not yet “overfitted” to the world, produce shocking, creative, unexpected responses precisely because they haven’t collapsed into standard patterns of thought. This freshness, this maintenance of high entropy in cognitive processes, may be essential to genuine intelligence. Yet our current training paradigms actively work against it, rewarding convergence towards common patterns and penalising deviation.

The Decade of Agents: Why Progress Takes Time

When Karpathy states this will be “the decade of agents” rather than “the year of agents,” he draws on hard-won experience from five years leading Tesla’s Autopilot programme. His insights into why artificial intelligence deployment takes far longer than demonstrations suggest carry particular weight given this background.

The central concept is what Karpathy calls “the march of nines.” Getting something to work 90% of the time—the level typically showcased in demonstrations—represents merely the first nine in “99.9%.” Each additional nine requires equivalent effort. During his tenure at Tesla, the team progressed through perhaps two or three nines over five years. More crucially, numerous nines remain before self-driving cars achieve true autonomy at scale.

This pattern isn’t unique to autonomous vehicles. Karpathy argues it applies across safety-critical domains, including software engineering. When code errors can leak millions of Social Security numbers or create critical security vulnerabilities, the cost of failure becomes prohibitively high. The demo-to-product gap widens dramatically. What works impressively in controlled conditions fails in countless edge cases when confronting reality’s full complexity.

Waymo’s experience illustrates this challenge. Despite providing “perfect drives” as early as 2014, the company still operates limited deployments requiring elaborate teleoperation infrastructure and supervision. Humans haven’t been removed; they’ve been rendered invisible, beaming in remotely to handle edge cases. The technology lives in a “pulled-back future”—functional but not yet economical, capable but not yet scalable.

Contemporary AI agents face analogous challenges. Whilst Claude and GPT-5 Pro demonstrate remarkable capabilities, they remain what Karpathy characterises as “elementary school students”—savants with perfect memory but lacking robust reasoning across all necessary dimensions. They’re “cognitively deficient” in ways users intuitively recognise even if they can’t articulate precisely what’s missing.

The Software Engineering Puzzle

Perhaps no domain better illustrates the puzzling contours of current AI capabilities than software engineering. Karpathy notes, somewhat ruefully, that whilst these systems were meant to enable “any economically valuable task,” API revenue remains “dominated by coding.” This supposedly general intelligence overwhelmingly excels at one specific domain.

This concentration isn’t accidental. Code enjoys unique properties that make it ideal for current AI architectures. Software development has always operated through text—terminals, editors, version control systems all manipulate textual representations. LLMs, trained on internet text, encounter code as a native format. Moreover, decades of infrastructure exist for handling code textually: diff tools for showing changes, IDEs for navigation, testing frameworks for verification.

Contrast this with domains lacking such infrastructure. Creating presentations involves spatial arrangement and visual design—there’s no “diff” for slides that elegantly shows modifications. Many knowledge work tasks involve physical documents, in-person interactions, or tacit knowledge that resists textual representation. These domains haven’t been pre-optimised for AI interaction in the way software development has.

Yet even in coding, Karpathy remains sceptical of current capabilities for cutting-edge work. When building nanoChat, a repository implementing a complete ChatGPT clone in simplified form, he found AI tools valuable for autocomplete and handling familiar patterns but inadequate for novel architectural decisions. The models kept trying to impose standard approaches when he deliberately chose non-standard implementations. They couldn’t comprehend his custom solutions, constantly suggesting deprecated APIs and bloating code with unnecessary defensive programming.

This points to a deeper truth: current models excel at reproducing common patterns from their training data but struggle with code “that has never been written before”—precisely the domain of frontier AI research itself. The recursive self-improvement that some forecast, where AI systems rapidly enhance their own capabilities, founders on this limitation. Models can accelerate work within established paradigms but cannot yet pioneer truly novel approaches.

The Trajectory of Intelligence Explosion

Karpathy’s perspective on potential intelligence explosions diverges sharply from both pessimistic and optimistic extremes. He sees AI not as a discrete, alien technology but as a continuation of computing’s evolution—part of an ongoing automation trend stretching back through compilers, high-level programming languages, and computer-aided design tools. From this view, the “intelligence explosion” has already been occurring for decades, visible in the exponential GDP growth curve that represents accumulated automation across countless domains.

This framing leads to counterintuitive predictions. Rather than expecting AI to suddenly accelerate economic growth from 2% annually to 20%, Karpathy suggests it will enable continued progress along the existing exponential trajectory. Just as computers, the internet, and mobile phones transformed society without producing visible discontinuities in aggregate growth statistics, AI will diffuse gradually across industries, maintaining rather than disrupting established growth patterns.

This gradualism doesn’t imply insignificance. The compounding effects of sustained exponential growth produce extraordinary transformation over time. But it does suggest that simple extrapolations from impressive demonstrations to imminent superintelligence misunderstand how technology integrates into society. There will be no discrete moment when “AGI” arrives and everything changes. Instead, we’ll experience continuous advancement in capabilities, continuous expansion of automation, and continuous adaptation of society to new technological possibilities.

The analogy to the Industrial Revolution proves instructive. That transformation didn’t occur through a single breakthrough but through cascading improvements across multiple technologies and practices, gradually shifting society from 0.2% annual growth to 2%. Similarly, AI’s impact will emerge through countless incremental deployments, each automating specific tasks, enabling new workflows, and creating feedback loops that accelerate subsequent progress.

The Human Element: Education in an AI Future

Karpathy’s work on Eureka, his educational initiative, reveals his deepest concerns about AI’s trajectory. He fears not that AI will fail but that “humanity gets disempowered by it,” relegated to the sidelines like the portly, passive citizens of WALL-E. His solution lies in radically reimagining education around the principle that “pre-AGI education is useful; post-AGI education is fun.”

The analogy to fitness culture illuminates this vision. Nobody needs physical strength to manipulate heavy objects—we have machines for that. Yet gyms proliferate because exercise serves intrinsic human needs: health, aesthetics, the satisfaction of physical mastery. Similarly, even in a world where AI handles most cognitive labour, humans will pursue learning for its inherent rewards: the pleasure of understanding, the status of expertise, the deep satisfaction of mental cultivation.

But achieving this vision requires solving a technical problem: making learning genuinely easy and rewarding. Currently, most people abandon learning because they encounter material that’s too difficult or too trivial, bouncing between frustration and boredom. Karpathy describes the experience of working with an expert language tutor who maintained a perfect calibration—always presenting challenges at the edge of current capability, never boring, never overwhelming. This created a state where “I was the only constraint to learning,” with knowledge delivery perfectly optimised.

Replicating this experience at scale represents what Karpathy sees as education’s great technical challenge. Current AI tutors, despite their sophistication, remain far from this standard. They can answer questions but cannot probe understanding, identify gaps, or sequence material to create optimal learning trajectories. The capability exists in exceptional human tutors; the challenge lies in encoding it algorithmically.

Yet Karpathy sees this challenge as tractable. Just as AI has transformed coding through autocomplete and code generation, it will eventually transform education through personalised, responsive tutoring. When learning becomes “trivial”—not in the sense of requiring no effort but in the sense of encountering no artificial obstacles—humans will pursue it enthusiastically. Not everyone will become an expert in everything, but the ceiling on human capability will rise dramatically as the floor on accessibility descends.

The Physics of Understanding: Karpathy’s Pedagogical Philosophy

Karpathy’s approach to teaching reveals principles applicable far beyond AI. His background in physics instilled what he describes as finding “first-order terms”—identifying the essential, dominant factors in any system whilst recognising that second and third-order effects exist but matter less. This habit of abstraction, of seeing spherical cows where others see only messy complexity, enables the creation of minimal, illustrative examples that capture phenomena’s essence.

MicroGrad exemplifies this approach perfectly. In 100 lines of Python, Karpathy implements backpropagation—the fundamental algorithm underlying all neural network training. Everything else in modern deep learning frameworks, he notes, is “just efficiency”—optimisations for speed, memory management, numerical stability. But the intellectual core, the actual mechanism by which networks learn, fits in 100 comprehensible lines. This distillation makes the previously arcane accessible.

The broader principle involves “untangling knowledge”—reorganising understanding so each concept depends only on what precedes it. This creates “ramps to knowledge” where learners never encounter gaps or leaps that would require them to take claims on faith. The famous transformer tutorial embodies this, beginning with a simple bigram model (literally a lookup table) and progressively adding components, each motivated by solving a specific limitation of what came before.

This approach contrasts sharply with the standard academic practice of presenting solutions before establishing problems, or introducing abstractions before concrete examples. Karpathy sees such approaches as, in his words, “a dick move”—they rob learners of the opportunity to grapple with challenges themselves, to develop intuition about what solutions might work, and to appreciate why particular approaches succeed where alternatives fail. The pedagogical crime isn’t challenging students; it’s presenting answers without first establishing questions.

Leading Theorists: The Intellectual Lineage

Richard Sutton and the Bitter Lesson

Richard Sutton, a pioneering reinforcement learning researcher, articulated what became known as “the bitter lesson”—the observation that simple, scalable methods leveraging computation consistently outperform approaches incorporating human knowledge or structural assumptions. His perspective suggests that the path to artificial general intelligence lies through learning algorithms powerful enough to discover structure from experience, much as evolution discovered biological intelligence.

Sutton’s famous assertion that “if you got to the squirrel, you’d be most of the way to AGI” reflects this view. Animal intelligence, in his framework, represents the core achievement—the fundamental learning algorithms that enable organisms to navigate environments, solve problems, and adapt to challenges. Human language and culture, whilst impressive, represent relatively minor additions to this foundation.

Karpathy respectfully dissents. His “we’re building ghosts, not animals” formulation captures the divergence: current AI systems don’t replicate the learning processes that create biological intelligence. They imitate the products of human intelligence (text, code, reasoning traces) rather than replicating its developmental origins. This distinction matters profoundly for predicting AI’s trajectory and understanding its capabilities and limitations.

Geoffrey Hinton and the Neural Network Renaissance

Geoffrey Hinton, often termed the “godfather of AI,” pioneered the neural network approaches that underpin contemporary systems. His persistence through decades when neural networks were unfashionable, his development of backpropagation techniques, and his later work on capsule networks and other architectures established the foundation for today’s large language models.

Karpathy studied directly under Hinton at the University of Toronto, experiencing firsthand the intellectual ferment as deep learning began its ascent to dominance. Hinton’s influence appears throughout Karpathy’s thinking—the emphasis on learning from data rather than hand-crafted rules, the focus on representation learning, the conviction that scale and simplicity often trump elaborate architectural innovations.

Yet Karpathy’s view extends beyond his mentor’s. Where Hinton focused primarily on perception (particularly computer vision), Karpathy grapples with the full scope of intelligence—reasoning, planning, continual learning, multi-agent interaction. His work synthesises Hinton’s foundational insights with broader questions about cognitive architecture and the nature of understanding itself.

Yann LeCun and Convolutional Networks

Yann LeCun’s development of convolutional neural networks in 1989 represented the first successful application of gradient descent to real-world pattern recognition. His work on handwritten digit recognition established core principles: the power of hierarchical feature learning, the importance of translation invariance, the value of specialised architectures for specific domains.

Karpathy’s reconstruction of LeCun’s 1989 network, time-travelling 33 years of algorithmic improvements, reveals his appreciation for this lineage. He found that pure algorithmic advances—modern optimisers, better architectures, regularisation techniques—could halve error rates. But achieving further gains required more data and more computation. This trinity—algorithms, data, compute—advances in lockstep, with no single factor dominating.

This lesson shapes Karpathy’s predictions about AI’s future. He expects continued progress across all three dimensions, with the next decade bringing better algorithms, vaster datasets, more powerful hardware, and more efficient software. But no breakthrough in any single dimension will produce discontinuous acceleration. Progress emerges from the intersection of many incremental improvements.

The Broader Intellectual Context

The debate Karpathy engages extends beyond specific individuals to fundamental questions about intelligence itself. Does intelligence arise primarily from general learning algorithms (the Sutton view) or from accumulated structure and innate mechanisms (the evolutionary perspective)? Can we build intelligence by imitating its products (the current LLM approach) or must we replicate its developmental processes? Will artificial intelligence remain fundamentally tool-like, augmenting human capability, or evolve into genuinely autonomous agents pursuing their own goals?

These questions connect to century-old debates in psychology and cognitive science between behaviourists emphasising learning and nativists emphasising innate structure. They echo discussions in evolutionary biology about the relative roles of genetic determination and developmental plasticity. They parallel arguments in philosophy of mind about whether intelligence requires embodiment or can exist as pure information processing.

Karpathy’s position threads between extremes. He acknowledges both the power of learning from data and the necessity of architectural structure. He recognises both the distinctiveness of AI systems and their illuminating analogies to biological intelligence. He balances optimism about AI’s potential with realism about current limitations and the difficulty of translating demonstrations into robust, deployed systems.

The Cognitive Core in Context: A New Paradigm for Intelligence

The concept of a cognitive core stripped of factual knowledge represents more than a technical proposal—it’s a reconceptualisation of what intelligence fundamentally is. Rather than viewing intelligence as encompassing both reasoning algorithms and accumulated knowledge, Karpathy proposes treating these as separate, with reasoning capability as the essence and factual knowledge as external resources to be accessed rather than internalised.

This separation mirrors certain aspects of human cognition whilst diverging in others. Humans do maintain a distinction between knowing how to think and knowing specific facts—we can reason about novel situations without direct experience, applying general problem-solving strategies learned in one domain to challenges in another. Yet our factual knowledge isn’t purely external; it shapes the very structure of our reasoning, creating rich semantic networks that enable rapid, intuitive judgement.

The proposal to strip AI systems down to cognitive cores involves accepting tradeoffs. Such systems would need to perform external lookups for factual information, introducing latency and dependency on knowledge bases. They would lack the pattern-matching capabilities that arise from vast memorisation, potentially missing connections between superficially unrelated domains. They might struggle with tasks requiring seamless integration of many small facts, where lookup costs would dominate processing time.

Yet the gains could prove transformative. A genuine cognitive core—compact, efficient, focused on algorithmic reasoning rather than fact retrieval—could operate in settings where current models fail. Edge deployment becomes feasible when models don’t require storing terabytes of parameters. Personalisation becomes practical when core reasoning engines can be fine-tuned or adapted without retraining on entire knowledge corpora. Interpretability improves when reasoning processes aren’t obscured by retrieval of memorised patterns.

Most profoundly, genuine cognitive cores might avoid the collapse and loss of entropy that plagues current models. Freed from the burden of maintaining consistency with vast memorised datasets, such systems could explore more diverse solution spaces, generate more varied outputs, and maintain the creative flexibility that characterises human cognition at its best.

Implications for the Decade Ahead

Karpathy’s decade-long timeline for agentic AI reflects hard-earned wisdom about technology deployment. His experience with autonomous vehicles taught him that impressive demonstrations represent merely the beginning of a long productisation journey. Each additional “nine” of reliability—moving from 90% to 99% to 99.9% accuracy—requires comparable effort. Safety-critical domains demand many nines before deployment becomes acceptable.

This reality shapes expectations for AI’s economic impact. Rather than sudden disruption, we’ll witness gradual diffusion across domains with varying characteristics. Tasks that are repetitive, well-defined, purely digital, and allowing high error rates will automate first. Call centre work exemplifies this profile—short interaction horizons, clear success criteria, limited context requirements, tolerance for occasional failures that human supervisors can catch.

More complex knowledge work will resist automation longer. Radiologists, consultants, accountants—professionals whose work involves lengthy timescales, subtle judgements, extensive context, and high costs of error—will see AI augmentation before replacement. The pattern will resemble Waymo’s current state: AI handling routine cases whilst humans supervise, intervene in edge cases, and maintain ultimate responsibility.

This graduated deployment creates an “autonomy slider”—a continuous spectrum from pure human operation through various degrees of AI assistance to eventual full automation. Most jobs won’t flip discretely from human to machine. Instead, they’ll slide along this spectrum as AI capabilities improve and organisations develop confidence in delegation. This process will unfold over years or decades, not months.

The economic implications differ from both optimistic and pessimistic extremes. We won’t see overnight mass unemployment—the gradual nature of deployment, the persistence of edge cases requiring human judgement, and society’s adaptation through creating new roles all mitigate disruption. But neither will we see disappointing underutilisation—the compound effect of many small automations across countless tasks will produce genuine transformation.

The Path Forward: Research Priorities

Karpathy’s analysis suggests several critical research directions for developing robust, capable AI systems. First, developing methods to isolate cognitive cores from memorised knowledge whilst maintaining reasoning capability. This might involve novel training objectives that penalise rote memorisation whilst rewarding generalisation, or architectural innovations that separate knowledge storage from reasoning mechanisms.

Second, creating effective continual learning systems that can distil experience into lasting improvements without catastrophic forgetting or model collapse. This requires moving beyond simple fine-tuning toward something more akin to the reflection and consolidation humans perform during sleep—identifying patterns in experience, extracting lessons, and integrating insights whilst maintaining diversity.

Third, advancing beyond current reinforcement learning to richer forms of learning from experience. Rather than broadcasting sparse reward signals across entire trajectories, systems need sophisticated credit assignment that identifies which reasoning steps contributed to success and which didn’t. This might involve explicit review processes where models analyse their own problem-solving attempts, or meta-learning approaches that learn how to learn from experience.

Fourth, developing multi-agent systems with genuine culture—shared knowledge bases that agents collectively maintain and evolve, self-play mechanisms that drive capability improvement through competition, and organisational structures that enable collaboration without centralized control. Current systems remain fundamentally solitary; genuine agent economies will require breakthroughs in coordination and communication.

Fifth, and perhaps most ambitiously, maintaining entropy in AI systems—preventing the collapse toward stereotyped outputs that currently plagues even frontier models. This might involve explicit diversity penalties, adversarial training to prevent convergence, or inspiration from biological systems that maintain variation through mechanisms like mutation and recombination.

Conclusion: Intelligence as Engineering Challenge

Andrej Karpathy’s vision of the cognitive core represents a mature perspective on artificial intelligence—neither breathlessly optimistic about imminent superintelligence nor dismissively pessimistic about current limitations. He sees AI as an engineering challenge rather than a mystical threshold, requiring patient work across multiple dimensions rather than awaiting a single breakthrough.

This perspective derives from direct experience with the messy reality of deploying AI systems at scale. Self-driving cars that work perfectly in demonstrations still require years of refinement before handling edge cases reliably. Coding agents that generate impressive solutions for common problems still struggle with novel architectural challenges. Educational AI that answers questions adequately still falls far short of expert human tutors’ adaptive responsiveness.

Yet within these limitations lies genuine progress. Models continue improving along multiple dimensions simultaneously. Infrastructure for deploying and managing AI systems grows more sophisticated. Understanding of these systems’ capabilities and constraints becomes more nuanced. The path forward is visible, even if it stretches further than optimists anticipated.

The concept of stripping knowledge to reveal the cognitive core captures this mature vision perfectly. Rather than pursuing ever-larger models memorising ever-more data, we might achieve more capable intelligence through subtraction—removing the crutch of memorisation to force development of robust reasoning algorithms. Like humans compelled to abstract and generalise because we cannot remember everything, AI systems might benefit from similar constraints.

This vision offers hope not for sudden transformation but for steady progress—the kind that compounds over decades into revolutionary change. It suggests that the hard technical problems of intelligence remain tractable whilst acknowledging their genuine difficulty. Most importantly, it positions humans not as passive observers of AI’s ascent but as active participants in shaping its development and ensuring its integration enhances rather than diminishes human flourishing.

The decade ahead will test these ideas. We’ll discover whether cognitive cores can be effectively isolated, whether continual learning mechanisms can be made robust, whether the demo-to-product gap can be bridged across diverse domains. The answers will shape not just the trajectory of AI technology but the future of human society in an increasingly automated world. Karpathy’s contribution lies in framing these questions with clarity, drawing on hard-won experience to guide expectations, and reminding us that the most profound challenges often yield to patient, disciplined engineering rather than waiting for miraculous breakthroughs.

read more
Quote: Andrej Karpathy – Ex-OpenAI, Ex-Tesla AI

Quote: Andrej Karpathy – Ex-OpenAI, Ex-Tesla AI

“I feel like the [ AI ] problems are tractable, they’re surmountable, but they’re still difficult. If I just average it out, it just feels like a decade [ to AGI ] to me.” – Andrej Karpathy – Ex-OpenAI, Ex-Tesla AI

Andrej Karpathy’s reflection—“I feel like the [ AI ] problems are tractable, they’re surmountable, but they’re still difficult. If I just average it out, it just feels like a decade [ to AGI ] to me.”—encapsulates both a grounded optimism and a caution honed through years at the forefront of artificial intelligence research. Understanding this statement requires context about the speaker, the evolution of the field, and the intellectual landscape that shapes contemporary thinking on artificial general intelligence (AGI).

Andrej Karpathy: Technical Leadership and Shaping AI’s Trajectory

Karpathy is recognised as one of the most influential figures in modern AI. With a doctorate under Geoffrey Hinton, the so-called “godfather” of deep learning, Karpathy’s early career put him at the confluence of academic breakthroughs and industrial deployment. At Stanford, he helped launch the seminal CS231n course, which became a training ground for a generation of practitioners. He subsequently led critical efforts at OpenAI and Tesla, where he served as Director of AI, architecting large-scale deep learning systems for both language and autonomous driving.

From the earliest days of deep learning, Karpathy has witnessed—and helped drive—several “seismic shifts” that have periodically redefined the field. He recalls, for example, the transition from neural networks being considered a niche topic to their explosive relevance with the advent of AlexNet. At OpenAI, he observed the limitations of reinforcement learning when applied too soon to general agent-building and became an early proponent of focusing on practical, useful systems rather than chasing abstract analogies with biological evolution.

Karpathy’s approach is self-consciously pragmatic. He discounts analogies between AI and animal evolution, preferring to frame current efforts as “summoning ghosts,” i.e., building digital entities trained by imitation, not evolved intelligence. His career has taught him to discount industry hype cycles and focus on the “march of nines”—the painstaking work required to close the gap between impressive demos and robust, trustworthy products. This stance runs through his entire philosophy on AI progress.

Context for the Quote: Realism amidst Exponential Hype

The statement about AGI’s timeline emerges from Karpathy’s nuanced position between the extremes of utopian accelerationism and excessive scepticism. Against a backdrop of industry figures claiming near-term transformative breakthroughs, Karpathy advocates for a middle path: current models represent significant progress, but numerous “cognitive deficits” persist. Key limitations include the lack of robust continual learning, difficulties generalising out-of-distribution, and the absence of key memory and reasoning capabilities seen in human intelligence.

Karpathy classifies present-day AI systems as “competent, but not yet capable agents”—useful in narrow domains, such as code generation, but unable to function autonomously in open-ended, real-world contexts. He highlights how models exhibit an uncanny ability to memorise, yet often lack the generalisation skills required for truly adaptive behaviour; they’re powerful, but brittle. The hard problems left are not insurmountable, but solving them—including integrating richer memory, developing agency, and building reliable, context-sensitive learning—will take sustained, multi-year effort.

AGI and the Broader Field: Dialogue with Leading Theorists

Karpathy’s thinking exists in dialogue with several foundational theorists:

  • Geoffrey Hinton: Pioneered deep learning and neural network approaches that underlie all current large-scale AI. His early conviction in neural networks, once seen as fringe, is now mainstream, but Hinton remains open to new architectural breakthroughs.

  • Richard Sutton: A major proponent of reinforcement learning as a route to general intelligence. Sutton’s vision focuses on “building animals”—systems capable of learning from scratch via trial and error in complex environments—whereas Karpathy now sees this as less immediately relevant than imitation-based, practically grounded approaches.

  • Yann LeCun: Another deep learning pioneer, LeCun has championed the continuous push toward self-supervised learning and innovations within model architecture.

  • The Scaling Optimists: The school of thought, including some in the OpenAI and DeepMind circles, who argue that simply increasing model size and data, within current paradigms, will inexorably deliver AGI. Karpathy explicitly distances himself from this view, arguing for the necessity of algorithmic innovation and socio-technical integration.

Karpathy sees the arc of AI progress as analogous to general trends in automation and computing: exponential in aggregate, but marked by periods of over-prediction, gradual integration, and non-linear deployment. He draws lessons from the slow maturation of self-driving cars—a field he led at Tesla—where early demos quickly gave way to years of incremental improvement, ironing out “the last nines” to reach real-world reliability.

He also foregrounds the human side of the equation: as AI’s technical capability increases, the question becomes as much about organisational integration, legal and social adaptation, as it does about raw model performance.

In Summary: Surmountable Yet Difficult

Karpathy’s “decade to AGI” estimate is anchored in a sober appreciation of both technical tractability and practical difficulty. He is neither pessimistic nor a hype-driven optimist. Instead, he projects that AGI—defined as machines able to deliver the full spectrum of knowledge work at human levels—will require another decade of systematic progress spanning model architecture, algorithmic innovation, memory, continual learning, and above all, integration with the complex realities of the real world.

His perspective stands out for its blend of technical rigour, historical awareness, and humility in the face of both engineering constraints and the unpredictability of broader socio-technical systems. In this, Karpathy situates himself in conversation with a lineage of thinkers who have repeatedly recalibrated the AI field’s ambitions—and whose own varied predictions continue to shape the ongoing march toward general intelligence.

read more
Quote: David Solomon – Goldman Sachs CEO

Quote: David Solomon – Goldman Sachs CEO

“If the firm grows and you expand and you can invest in other areas for growth, we’ll wind up with more jobs… we have at every step along the journey for the last forty years as technology has made us more productive. I don’t think it’s different this time [with AI].” – David Solomon – Goldman Sachs CEO

David Michael Solomon, born in 1962 in Hartsdale, New York, is an American investment banker and DJ, currently serving as the CEO and Chairman of Goldman Sachs. His journey into the financial sector began after he graduated with a BA in political science from Hamilton College. Initially, Solomon worked at Irving Trust Company and Drexel Burnham before joining Bear Stearns. In 1999, he moved to Goldman Sachs as a partner and became co-head of the High Yield and Leveraged Loan Business.

Solomon’s rise within Goldman Sachs was swift and strategic. He became the co-head of the Investment Banking Division in 2006 and held this role for a decade. In 2017, he was appointed President and Chief Operating Officer, and by October 2018, he succeeded Lloyd Blankfein as CEO. He became Chairman in January 2019.

Beyond his financial career, Solomon is known for his passion for music, producing electronic dance music under the alias “DJ D-Sol”. He has performed at various venues, including nightclubs and music festivals in New York, Miami, and The Bahamas.

Context of the Quote

The quote highlights Solomon’s perspective on technology and job creation in the financial sector. He suggests that while technology, particularly AI, can enhance productivity and potentially lead to job reductions in certain areas, the overall growth of the firm will create more opportunities for employment. This view is rooted in his experience observing how technological advancements have historically led to increased productivity and growth for Goldman Sachs.

Leading Theorists on AI and Employment

Several leading theorists have explored the impact of AI on employment, with divergent views:

  • Joseph Schumpeter is famous for his theory of “creative destruction,” which suggests that technological innovations often lead to the destruction of existing jobs but also create new ones. This cycle is seen as essential for economic growth and innovation.

  • Klaus Schwab, founder of the World Economic Forum, has discussed the Fourth Industrial Revolution, emphasizing how AI and automation will transform industries. However, he also highlights the potential for new job creation in emerging sectors.

  • Economists Erik Brynjolfsson and Andrew McAfee have written extensively on how technology can lead to both job displacement and creation. They argue that while AI may reduce certain types of jobs, it also fosters economic growth and new opportunities.

These theorists provide a backdrop for understanding Solomon’s optimistic view on AI’s impact on employment, focusing on the potential for growth and innovation to offset job losses.

Conclusion

David Solomon’s quote encapsulates his optimism about the interplay between technology and job creation. Focusing on the strategic growth of Goldman Sachs, he believes that technological advancements will enhance productivity and create opportunities for expansion, ultimately leading to more employment opportunities. This perspective aligns with broader discussions among economists and theorists on the transformative role of AI in the workplace.

read more
Quote: David Solomon – Goldman Sachs CEO

Quote: David Solomon – Goldman Sachs CEO

“Markets run in cycles, and whenever we’ve historically had a significant acceleration in a new technology that creates a lot of capital formation and therefore lots of interesting new companies around it, you generally see the market run ahead of the potential. Are there going to be winners and losers? There are going to be winners and losers.” – David Solomon – Goldman Sachs CEO

The quote, “Markets run in cycles, and whenever we’ve historically had a significant acceleration in a new technology that creates a lot of capital formation and therefore lots of interesting new companies around it, you generally see the market run ahead of the potential. Are there going to be winners and losers? There are going to be winners and losers,” comes from a public discussion with David Solomon, CEO of Goldman Sachs, during Italian Tech Week in October 2025. This statement was made in the context of a wide-ranging interview that addressed the state of the US and global economy, the impact of fiscal stimulus and technology infrastructure spending, and, critically, the current investment climate surrounding artificial intelligence (AI) and other emergent technologies.

Solomon’s comments were prompted by questions around the record-breaking rallies in US and global equity markets and specifically the extraordinary market capitalisations reached by leading tech firms. He highlighted the familiar historical pattern: periods of market exuberance often occur when new technologies spur rapid capital formation, leading to the emergence of numerous new companies around a transformative theme. Solomon drew parallels with the Dot-com boom to underscore the cyclical nature of markets and to remind investors that dramatic phases of growth inevitably produce both outsized winners and significant casualties.

His insight reflects a seasoned banker’s view, grounded in empirical observation: while technological waves can drive periods of remarkable wealth creation and productivity gains, they also tend to attract speculative excesses. Market valuations in these periods often disconnect from underlying fundamentals, setting the stage for later corrections. The resulting market shake-outs separate enduring companies from those that fail to deliver sustainable value.

About David Solomon

David M. Solomon is one of the most prominent figures in global finance, serving as the CEO and Chairman of Goldman Sachs since 2018. Raised in New York and a graduate of Hamilton College, Solomon has built his reputation over four decades in banking—rising through leadership positions at Irving Trust, Drexel Burnham, and Bear Stearns before joining Goldman Sachs in 1999 as a partner. He subsequently became global head of the Financing Group, then co-head of the Investment Banking Division, playing a central role in shaping the firm’s capital markets strategy.

Solomon is known for his advocacy of organisational modernisation and culture change at Goldman Sachs—prioritising employee well-being, increasing agility, and investing heavily in technology. He combines traditional deal-making acumen with an openness to digital transformation. Beyond banking, Solomon has a notable side-career as a DJ under the name DJ D-Sol, performing electronic dance music at high-profile venues.

Solomon’s career reflects both the conservatism and innovative ambition associated with modern Wall Street leadership: an ability to see risk cycles clearly, and a willingness to pivot business models to suit shifts in technological and regulatory environments. His net worth in 2025 is estimated between $85 million and $200 million, owing to decades of compensation, equity, and investment performance.

Theoretical Foundations: Cycles, Disruptive Innovation, and Market Dynamics

Solomon’s perspective draws implicitly on a lineage of economic theory and market analysis concerning cycles of innovation, capital formation, and asset bubbles. Leading theorists and their contributions include:

  • Joseph Schumpeter: Schumpeter’s theory of creative destruction posited that economic progress is driven by cycles of innovation, where new technologies disrupt existing industries, create new market leaders, and ultimately cause the obsolescence or failure of firms unable to adapt. Schumpeter emphasised how innovation clusters drive periods of rapid growth, investment surges, and, frequently, speculative excess.

  • Carlota Perez: In Technological Revolutions and Financial Capital (2002), Perez advanced a model of techno-economic paradigms, proposing that every major technological revolution (e.g., steam, electricity, information technology) proceeds through phases: an initial installation period—characterised by exuberant capital inflows, speculation, and bubble formation—followed by a recessionary correction, and, eventually, a deployment period, where productive uses of the technology diffuse more broadly, generating deep-seated economic gains and societal transformation. Perez’s work helps contextualise Solomon’s caution about markets running ahead of potential.

  • Charles Kindleberger and Hyman Minsky: Both scholars examined the dynamics of financial bubbles. Kindleberger, in Manias, Panics, and Crashes, and Minsky, through his Financial Instability Hypothesis, described how debt-fuelled euphoria and positive feedback loops of speculation can drive financial markets to overshoot the intrinsic value created by innovation, inevitably resulting in busts.

  • Clayton Christensen: Christensen’s concept of disruptive innovation explains how emergent technologies, initially undervalued by incumbents, can rapidly upend entire industries—creating new winners while displacing former market leaders. His framework helps clarify Solomon’s points about the unpredictability of which companies will ultimately capture value in the current AI wave.

  • Benoit Mandelbrot: Applying his fractal and complexity theory to financial markets, Mandelbrot challenged the notion of equilibrium and randomness in price movement, demonstrating that markets are prone to extreme events—outlier outcomes that, while improbable under standard models, are a recurrent feature of cyclical booms and busts.

Practical Relevance in Today’s Environment

The patterns stressed by Solomon, and their theoretical antecedents, are especially resonant given the current environment: massive capital allocations into AI, cloud infrastructure, and adjacent technologies—a context reminiscent of previous eras where transformative innovations led markets both to moments of extraordinary wealth creation and subsequent corrections. These cycles remain a central lens for investors and business leaders navigating this era of technological acceleration.

By referencing both history and the future, Solomon encapsulates the balance between optimism over the potential of new technology and clear-eyed vigilance about the risks endemic to all periods of market exuberance.

read more
Quote: David Solomon – Goldman Sachs CEO

Quote: David Solomon – Goldman Sachs CEO

“AI really allows smart, talented, driven, sophisticated people to be more productive – to touch more people, have better information at their disposal, better analysis.” – David Solomon – Goldman Sachs CEO

David Solomon, CEO of Goldman Sachs, made the statement “AI really allows smart, talented, driven, sophisticated people to be more productive – to touch more people, have better information at their disposal, better analysis” during an interview at Italian Tech Week 2025, reflecting his conviction that artificial intelligence is redefining productivity and impact across professional services and finance.

David Solomon is one of the most influential figures in global finance, serving as Chairman and CEO of Goldman Sachs since 2018. Born in 1962 in Hartsdale, New York, Solomon’s early years were shaped by strong family values, a pursuit of education at Hamilton College, and a keen interest in sport and leadership. Solomon’s ascent in the industry began after stints at Irving Trust and Drexel Burnham, specialising early in commercial paper and junk bonds, then later at Bear Stearns where he played a central role in project financing. In 1999, he joined Goldman Sachs as a partner and quickly rose through the ranks—serving as Global Head of the Financing Group and later Co-Head of the Investment Banking Division for a decade.

His leadership is marked by an emphasis on modernisation, talent development, and integrating technology into the financial sector. Notably, Solomon has overseen increased investments in digital platforms and has reimagined work culture, including reducing working hours and implementing real-time performance review systems. Outside his professional life, Solomon is distinctively known for his passion for music, performing as “DJ D-Sol” at major electronic dance music venues, symbolising a leadership style that blends discipline with creative openness.

Solomon’s remarks on AI at Italian Tech Week are rooted in Goldman Sachs’ major investments in technology: with some 12,000 engineers and cutting-edge AI platforms, Solomon champions the view that technology not only streamlines operational efficiency but fundamentally redefines the reach and ability of talented professionals, providing richer data, deeper insights, and more effective analysis. He frames AI as part of a long continuum—from the days of microfiche and manual records to today’s instant, voice-powered analytics—positioning technology as both a productivity enabler and an engine for growth.

Leading Theorists and Context in AI Productivity

Solomon’s thinking sits at the crossroads of key theoretical advances in artificial intelligence and productivity economics. The transformation he describes draws extensively from foundational theorists and practitioners who have shaped our understanding of AI’s organisational impact:

  • Herbert Simon: A founder of artificial intelligence as a discipline, Simon’s concept of “bounded rationality” highlighted that real-world decision making could be fundamentally reshaped by computational power. Simon envisioned computers extending the limits of human cognition, a concept directly echoed in Solomon’s belief that AI produces leverage for talented professionals.

  • Erik Brynjolfsson: At MIT, Brynjolfsson has argued that AI is a “general purpose technology” like steam power or electricity, capable of diffusing productivity gains across every sector through automation, improved information processing, and new business models. His work clarifies that the impact of AI is not in replacing human value, but augmenting it, making people exponentially more productive.

  • Andrew Ng: As a pioneer in deep learning, Ng has emphasised the role of AI as a productivity tool: automating routine tasks, supporting complex analysis, and dramatically increasing the scale and speed at which decisions can be made. Ng’s teaching at Stanford and public writings focus on making AI accessible as a resource to boost human capability rather than a substitute.

  • Daron Acemoglu: The MIT economist challenges overly optimistic readings, arguing that the net benefits of AI depend on balanced deployment, policy, and organisational adaptation. Acemoglu frames the debate on whether AI will create or eliminate jobs, highlighting the strategic choices organisations must make—a theme Solomon directly addresses in his comments on headcount in banking.

  • Geoffrey Hinton: Widely known as “the godfather of deep learning,” Hinton’s research underpins the practical capabilities of AI systems—particularly in areas such as data analysis and decision support—that Solomon highlights as crucial to productive professional services.

 

Contemporary Application and Analysis

The productivity gains Solomon identifies are playing out across multiple sectors:

  • In financial services, AI-driven analytics enable deeper risk management, improved deal generation, and scalable client engagement.
  • In asset management and trading, platforms like Goldman Sachs’ own “Assistant” and generative coding tools (e.g., Cognition Labs’ Devin) allow faster, more nuanced analysis and automation.
  • The “power to touch more people” is realised through personalised client service, scalable advisory, and rapid market insight, bridging human expertise and computational capacity.

Solomon’s perspective resonates strongly with current debates on the future of work. While risks—such as AI investment bubbles, regulatory uncertainty, and workforce displacement—are acknowledged, Solomon positions AI as a strategic asset: not a threat to jobs, but a catalyst for organisational expansion and client impact, consistent with the lessons learned through previous technology cycles.

Theoretical Context Table

Theorist
Core Idea
Relevance to Solomon’s Statement
Herbert Simon
Bounded rationality, decision support
AI extending cognitive limits and enabling smarter analysis
Erik Brynjolfsson
AI as general purpose technology
Productivity gains and diffusion through diverse organisations
Andrew Ng
AI augments tasks, boosts human productivity
AI as a tool for scalable information and superior outcomes
Daron Acemoglu
Balance of job creation/destruction by technology
Strategic choices in deploying AI impact workforce and growth
Geoffrey Hinton
Deep learning, data analysis
Enabling advanced analytics and automation in financial services

Essential Insights

  • AI’s impact is cumulative and catalytic, empowering professionals to operate at far greater scale and depth than before, as illustrated by Solomon’s personal technological journey—from manual information gathering to instantaneous AI-driven analytics.
  • The quote’s context reflects the practical reality of AI at the world’s leading financial institutions, where technology spend rivals infrastructure, and human-machine synergy is central to strategy.
  • Leading theorists agree: real productivity gains depend on augmenting human capability, strategic deployment, and continual adaptation—principles explicitly recognised in Solomon’s operational philosophy and in global best practice.

read more
Quote: Jamie Dimon – JP Morgan Chase CEO

Quote: Jamie Dimon – JP Morgan Chase CEO

“Take the Internet bubble. Remember that blew up and I can name 100 companies that were worth $50 billion and disappeared…. So there will be some real big companies, real big success. [ AI ]will work in spite of the fact that not everyone invested is going to have a great investment return.” – Jamie Dimon, CEO JP Morgan Chase

Jamie Dimon’s observation about artificial intelligence investment echoes his experience witnessing the dot-com bubble’s collapse at the turn of the millennium—a period when he was navigating his own career transition from Citigroup to Bank One. Speaking to Bloomberg in London during October 2025, the JPMorgan Chase chairman drew upon decades of observing technological disruption to contextualise the extraordinary capital deployment currently reshaping the AI landscape. His commentary serves as a measured counterpoint to the euphoria surrounding generative artificial intelligence, reminding investors that transformative technologies invariably produce both spectacular winners and catastrophic losses.

The Speaker: Institutional Banking’s Preeminent Figure

Jamie Dimon has commanded JPMorgan Chase since 2006, transforming it into America’s largest bank by assets whilst establishing himself as Wall Street’s most influential voice. His journey to this position began in 1982 when he joined American Express as an assistant to Sandy Weill, embarking upon what would become one of the most consequential partnerships in American finance. For sixteen years, Dimon and Weill orchestrated a series of acquisitions that built Travelers Group into a financial services colossus, culminating in the 1998 merger with Citicorp to form Citigroup.

The relationship ended abruptly that same year when Weill asked Dimon to resign—a decision Weill later characterised as regrettable to The New York Times. The ouster proved fortuitous. In 2000, Dimon assumed leadership of Bank One, a struggling Chicago-based institution he successfully revitalised. When JPMorgan acquired Bank One in 2004, Dimon became president and chief operating officer before ascending to chief executive two years later. Under his stewardship, JPMorgan’s stock value has tripled, and in 2023 the bank recorded the largest annual profit in US banking history at nearly $50 billion.

Dimon’s leadership during the 2008 financial crisis distinguished him amongst his peers. Whilst competitors collapsed or required government rescue, JPMorgan emerged strengthened, acquiring Bear Stearns and Washington Mutual. He reprised this role during the 2023 regional banking crisis, coordinating an industry response that saw eleven major banks contribute $30 billion to stabilise First Republic Bank. This pattern of crisis management has positioned him as what analyst Mike Mayo termed “a senior statesperson” for the financial industry.

Beyond banking, Dimon maintains substantial political engagement. Having donated over $500,000 to Democratic candidates between 1989 and 2009, he has since adopted a more centrist posture, famously declaring to CNBC in 2019 that “my heart is Democratic, but my brain is kind of Republican”. He served briefly on President Trump’s business advisory council in 2017 and has repeatedly faced speculation about presidential ambitions, confirming in 2016 he would “love to be president” whilst acknowledging the practical obstacles. In 2024, he endorsed Nikki Haley in the Republican primary before speaking positively about Trump following Haley’s defeat.

The Technological Context: AI’s Investment Frenzy

Dimon’s October 2025 remarks addressed the extraordinary capital deployment underway in artificial intelligence infrastructure. His observation that approximately $1 trillion in AI-related spending was occurring “this year” encompasses investments by hyperscalers—the massive cloud computing providers—alongside venture capital flowing to companies like OpenAI, which despite substantial losses continues attracting vast sums. This investment boom has propelled equity markets into their third consecutive year of bull-market conditions, with asset prices reaching elevated levels and credit spreads compressing to historical lows.

At JPMorgan itself, Dimon revealed the bank has maintained systematic AI investment since 2012, allocating $2 billion annually and employing 2,000 specialists dedicated to the technology. The applications span risk management, fraud detection, marketing, customer service, and software development, with approximately 150,000 employees weekly utilising the bank’s internal generative AI tools. Crucially, Dimon reported achieving rough parity between the $2 billion expenditure and measurable benefits—a ratio he characterised as “the tip of the iceberg” given improvements in service quality that resist quantification.

His assessment that AI “will affect jobs” reflects the technology’s capacity to eliminate certain roles whilst enhancing others, though he expressed confidence that successful deployment would generate net employment growth at JPMorgan through retraining and redeployment programmes. This pragmatic stance—neither utopian nor dystopian—typifies Dimon’s approach to technological change: acknowledge disruption candidly whilst emphasising adaptive capacity.

The Dot-Com Parallel: Lessons from Previous Technological Euphoria

Dimon’s reference to the Internet bubble carries particular resonance given his vantage point during that era. In 1998, whilst serving as Citigroup’s president, he witnessed the NASDAQ’s ascent to unsustainable valuations before the March 2000 collapse obliterated trillions in market capitalisation. His claim that he could “name 100 companies that were worth $50 billion and disappeared” speaks to the comprehensive destruction of capital that accompanied the bubble’s deflation. Companies such as Pets.com, Webvan, and eToys became cautionary tales—businesses predicated upon sound concepts executed prematurely or inefficiently, consuming vast investor capital before failing entirely.

Yet from this wreckage emerged the digital economy’s defining enterprises. Google, incorporated in 1998, survived the downturn to become the internet’s primary gateway. Facebook, founded in 2004, built upon infrastructure and lessons from earlier social networking failures. YouTube, established in 2005, capitalised on broadband penetration that earlier video platforms lacked. Dimon’s point—that “there will be some real big companies, real big success” emerging from AI investment despite numerous failures—suggests that capital deployment exceeding economically optimal levels nonetheless catalyses innovation producing enduring value.

This perspective aligns with economic theories recognising that technological revolutions characteristically involve overshoot. The railway boom of the 1840s produced excessive track mileage and widespread bankruptcies, yet established transportation infrastructure enabling subsequent industrialisation. The telecommunications bubble of the late 1990s resulted in overbuilt fibre-optic networks, but this “dark fibre” later supported broadband internet at marginal cost. Dimon’s observation that technological transitions prove “productive” in aggregate “in spite of the fact that not everyone invested is going to have a great investment return” captures this dynamic: society benefits from infrastructure investment even when investors suffer losses.

Schumpeterian Creative Destruction and Technological Transition

Joseph Schumpeter’s concept of creative destruction provides theoretical foundation for understanding the pattern Dimon describes. Writing in Capitalism, Socialism and Democracy (1942), Schumpeter argued that capitalism’s essential characteristic involves “the process of industrial mutation that incessantly revolutionises the economic structure from within, incessantly destroying the old one, incessantly creating a new one.” This process necessarily produces winners and losers—incumbent firms clinging to obsolete business models face displacement by innovators exploiting new technological possibilities.

Schumpeter emphasised that monopolistic competition amongst innovators drives this process, with entrepreneurs pursuing temporary monopoly rents through novel products or processes. The expectation of extraordinary returns attracts excessive capital during technology booms, funding experiments that collectively advance knowledge even when individual ventures fail. This mechanism explains why bubbles, whilst financially destructive, accelerate technological diffusion: the availability of capital enables rapid parallel experimentation impossible under conservative financing regimes.

Clayton Christensen’s theory of disruptive innovation, elaborated in The Innovator’s Dilemma (1997), complements Schumpeter’s framework by explaining why established firms struggle during technological transitions. Christensen observed that incumbent organisations optimise for existing customer needs and established value networks, rendering them structurally incapable of pursuing initially inferior technologies serving different markets. Entrants unburdened by legacy systems and customer relationships therefore capture disruptive innovations’ benefits, whilst incumbents experience declining relevance.

Dimon’s acknowledgement that “there will be jobs that are eliminated” whilst predicting net employment growth at JPMorgan reflects these dynamics. Artificial intelligence constitutes precisely the type of general-purpose technology that Christensen’s framework suggests will restructure work organisation. Routine tasks amenable to codification face automation, requiring workforce adaptation through “retraining and redeployment”—the organisational response Dimon describes JPMorgan implementing.

Investment Cycles and Carlota Pérez’s Technological Surges

Carlota Pérez’s analysis in Technological Revolutions and Financial Capital (2002) offers sophisticated understanding of the boom-bust patterns characterising technological transitions. Pérez identifies a consistent sequence: technological revolutions begin with an “irruption” phase as entrepreneurs exploit new possibilities, followed by a “frenzy” phase when financial capital floods in, creating asset bubbles disconnected from productive capacity. Inevitable crash precipitates a “synergy” phase when surviving innovations diffuse broadly, enabling a “maturity” phase of stable growth until the next technological revolution emerges.

The dot-com bubble exemplified Pérez’s frenzy phase—capital allocated indiscriminately to internet ventures regardless of business fundamentals, producing the NASDAQ’s March 2000 peak before three years of decline. The subsequent synergy phase saw survivors like Amazon and Google achieve dominance whilst countless failures disappeared. Dimon’s reference to “100 companies that were worth $50 billion and disappeared” captures the frenzy phase’s characteristic excess, whilst his citation of “Facebook, YouTube, Google” represents the synergy phase’s enduring value creation.

Applying Pérez’s framework to artificial intelligence suggests current investment levels—the $1 trillion deployment Dimon referenced—may indicate the frenzy phase’s advanced stages. Elevated asset prices, compressed credit spreads, and widespread investor enthusiasm traditionally precede corrections enabling subsequent consolidation. Dimon’s observation that he remains “a long-term optimist” whilst cautioning that “asset prices are high” reflects precisely the ambivalence appropriate during technological transitions’ financial euphoria: confidence in transformative potential tempered by recognition of valuation excess.

Hyman Minsky’s Financial Instability Hypothesis

Hyman Minsky’s financial instability hypothesis, developed throughout the 1960s and 1970s, explains the endogenous generation of financial fragility during stable periods. Minsky identified three financing postures: hedge finance, where cash flows cover debt obligations; speculative finance, where near-term cash flows cover interest but not principal, requiring refinancing; and Ponzi finance, where cash flows prove insufficient even for interest, necessitating asset sales or further borrowing to service debt.

Economic stability encourages migration from hedge toward speculative and ultimately Ponzi finance as actors’ confidence increases. During technological booms, this migration accelerates—investors fund ventures lacking near-term profitability based upon anticipated future cash flows. The dot-com era witnessed classic Ponzi dynamics: companies burning capital quarterly whilst promising eventual dominance justified continued financing. When sentiment shifted, refinancing evaporated, triggering cascading failures.

Dimon’s comment that “not everyone invested is going to have a great investment return” implicitly acknowledges Minskian dynamics. The $1 trillion flowing into AI infrastructure includes substantial speculative and likely Ponzi finance—investments predicated upon anticipated rather than demonstrated cash flows. OpenAI’s losses despite massive valuation exemplify this pattern. Yet Minsky recognised that such dynamics, whilst generating financial instability, also fund innovation exceeding levels conservative finance would support. Society gains from experiments capital discipline would preclude.

Network Effects and Winner-Take-All Dynamics

The persistence of “real big companies, real big success” emerging from technological bubbles reflects network effects characteristic of digital platforms. Economist W. Brian Arthur’s work on increasing returns demonstrated that technologies exhibiting positive feedback—where adoption by some users increases value for others—tend toward monopolistic market structures. Each additional Facebook user enhances the platform’s value to existing users, creating barriers to competitor entry that solidify dominance.

Carl Shapiro and Hal Varian’s Information Rules (1998) systematically analysed information goods’ economics, emphasising that near-zero marginal costs combined with network effects produce natural monopolies in digital markets. This explains why Google commands search, Amazon dominates e-commerce, and Facebook controls social networking despite numerous well-funded competitors emerging during the dot-com boom. Superior execution combined with network effects enabled these firms to achieve sustainable competitive advantage.

Artificial intelligence exhibits similar dynamics. Training large language models requires enormous capital and computational resources, but deploying trained models incurs minimal marginal cost. Firms achieving superior performance attract users whose interactions generate data enabling further improvement—a virtuous cycle competitors struggle to match. Dimon’s prediction of “some real big companies, real big success” suggests he anticipates winner-take-all outcomes wherein a handful of AI leaders capture disproportionate value whilst numerous competitors fail.

Public Policy Implications: Industrial Policy and National Security

During the Bloomberg interview, Dimon addressed the Trump administration’s emerging industrial policy, particularly regarding strategic industries like rare earth minerals and semiconductor manufacturing. His endorsement of government support for MP Materials—a rare earth processor—reveals pragmatic acceptance that national security considerations sometimes warrant departure from pure market principles. This stance reflects growing recognition that adversarial competition with China necessitates maintaining domestic production capacity in strategically critical sectors.

Dani Rodrik’s work on industrial policy emphasises that whilst governments possess poor records selecting specific winners, they can effectively support broad technological capabilities through coordinated investment in infrastructure, research, and human capital. Mariana Mazzucato’s The Entrepreneurial State (2013) documents government’s crucial role funding high-risk innovation underlying commercial technologies—the internet, GPS, touchscreens, and voice recognition all emerged from public research before private commercialisation.

Dimon’s caution that industrial policy must “come with permitting” and avoid “virtue signalling” reflects legitimate concerns about implementation quality. Subsidising industries whilst maintaining regulatory barriers preventing their operation achieves nothing—a pattern frustrating American efforts to onshore manufacturing. His emphasis on “long-term purchase agreements” as perhaps “the most important thing” recognises that guaranteed demand reduces risk more effectively than capital subsidies, enabling private investment that government funding alone cannot catalyse.

Market Conditions and Forward-Looking Concerns

Dimon’s October 2025 assessment of macroeconomic conditions combined optimism about continued expansion with caution regarding inflation risks. His observation that “consumers are still okay” because of employment—”jobs, jobs, jobs”—identifies the crucial variable determining economic trajectory. Consumer spending constitutes approximately 70% of US GDP; sustained employment supports spending even as other indicators suggest vulnerability.

Yet his expression of being “a little more nervous about inflation not coming down like people expect” challenges consensus forecasts anticipating Federal Reserve interest rate cuts totalling 100 basis points over the subsequent twelve months. Government spending—which Dimon characterised as “inflationary”—combined with potential supply-side disruptions from tariffs could reverse disinflationary trends. Should inflation prove stickier than anticipated, the Fed would face constraints limiting monetary accommodation, potentially triggering the 2026 recession Dimon acknowledged “could happen.”

This assessment demonstrates Dimon’s characteristic refusal to offer false certainty. His acknowledgement that forecasts “have almost always been wrong, and the Fed’s been wrong too” reflects epistemic humility appropriate given macroeconomic forecasting’s poor track record. Rather than pretending precision, he emphasises preparedness: “I hope for the best, plan for the worst.” This philosophy explains JPMorgan’s consistent outperformance—maintaining sufficient capital and liquidity to withstand adverse scenarios whilst remaining positioned to exploit opportunities competitors’ distress creates.

Leadership Philosophy and Organisational Adaptation

The interview revealed Dimon’s approach to deploying artificial intelligence throughout JPMorgan’s operations. His emphasis that “every time we meet as a business, we ask, what are you doing that we could do to serve your people?” reflects systematic organisational learning rather than top-down technology imposition. This methodology—engaging managers to identify improvement opportunities rather than mandating specific implementations—enables bottom-up innovation whilst maintaining strategic coherence.

Dimon’s observation that “as managers learn how to do it, they’re asking more questions” captures the iterative process through which organisations absorb disruptive technologies. Initial deployments generate understanding enabling more sophisticated applications, creating momentum as possibilities become apparent. The statistic that 150,000 employees weekly utilise JPMorgan’s internal AI tools suggests successful cultural embedding—technology adoption driven by perceived utility rather than compliance.

This approach contrasts with common patterns wherein organisations acquire technology without changing work practices, yielding disappointing returns. Dimon’s insistence on quantifying benefits—”we have about $2 billion of benefit” matching the $2 billion expenditure—enforces accountability whilst acknowledging that some improvements resist measurement. The admission that quantifying “improved service” proves difficult “but we know” it occurs reflects sophisticated understanding that financial metrics capture only partial value.

Conclusion: Technological Optimism Tempered by Financial Realism

Jamie Dimon’s commentary on artificial intelligence investment synthesises his extensive experience navigating technological and financial disruption. His parallel between current AI enthusiasm and the dot-com bubble serves not as dismissal but as realistic framing—transformative technologies invariably attract excessive capital, generating both spectacular failures and enduring value creation. The challenge involves maintaining strategic commitment whilst avoiding financial overextension, deploying technology systematically whilst preserving adaptability, and pursuing innovation whilst managing risk.

His perspective carries weight because it emerges from demonstrated judgement. Having survived the dot-com collapse, steered JPMorgan through the 2008 crisis, and maintained the bank’s technological competitiveness across two decades, Dimon possesses credibility competitors lack. When he predicts “some real big companies, real big success” whilst cautioning that “not everyone invested is going to have a great investment return,” the statement reflects neither pessimism nor hype but rather accumulated wisdom about how technological revolutions actually unfold—messily, expensively, destructively, and ultimately productively.

read more
Quote: Jamie Dimon – JP Morgan Chase CEO

Quote: Jamie Dimon – JP Morgan Chase CEO

“People shouldn’t put their head in the sand. [AI] is going to affect jobs. Think of every application, every service you do; you’ll be using .. AI – some to enhance it. Some of it will be you doing the same job; you’re doing a better job at it. There will be jobs that are eliminated, but you’re better off being way ahead of the curve.” – Jamie Dimon, CEO JP Morgan Chase

Jamie Dimon delivered these observations on artificial intelligence during an interview with Bloomberg’s Tom Mackenzie in London on 7 October 2025, where he discussed JPMorgan Chase’s decade-long engagement with AI technology and its implications for the financial services sector. His comments reflect both the pragmatic assessment of a chief executive who has committed substantial resources to technological transformation and the broader perspective of someone who has navigated multiple economic cycles throughout his career.

The Context of Dimon’s Statement

JPMorgan Chase has been investing in AI since 2012, well before the recent generative AI explosion captured public attention. The bank now employs 2,000 people dedicated to AI initiatives and spends $2 billion annually on these efforts. This investment has already generated approximately $2 billion in quantifiable benefits, with Dimon characterising this as merely “the tip of the iceberg.” The technology permeates every aspect of the bank’s operations—from risk management and fraud detection to marketing, idea generation and customer service.

What makes Dimon’s warning particularly salient is his acknowledgement that approximately 150,000 JPMorgan employees use the bank’s suite of AI tools weekly. This isn’t theoretical speculation about future disruption; it’s an ongoing transformation within one of the world’s largest financial institutions, with assets of $4.0 trillion. The bank’s approach combines deployment across business functions with what Dimon describes as a cultural shift—managers and leaders are now expected to ask continuously: “What are you doing that we could do to serve your people? Why can’t you do better? What is somebody else doing?”

Dimon’s perspective on job displacement is notably unsentimental whilst remaining constructive. He rejects the notion of ignoring AI’s impact, arguing that every application and service will incorporate the technology. Some roles will be enhanced, allowing employees to perform better; others will be eliminated entirely. His solution centres on anticipatory adaptation rather than reactive crisis management—JPMorgan has established programmes for retraining and redeploying staff. For the bank itself, Dimon envisions more jobs overall if the institution succeeds, though certain functions will inevitably contract.

His historical framing of technological disruption provides important context. Drawing parallels to the internet bubble, Dimon noted that whilst hundreds of companies worth billions collapsed, the period ultimately produced Facebook, YouTube and Google. He applies similar logic to current AI infrastructure spending, which is approaching $1 trillion annually across the sector. There will be “a lot of losers, a lot of winners,” but the aggregate effect will prove productive for the economy.

Jamie Dimon: A Biography

Jamie Dimon has served as Chairman and Chief Executive Officer of JPMorgan Chase since 2006, presiding over its emergence as the leading US bank by domestic assets under management, market capitalisation and publicly traded stock value. Born on 13 March 1956, Dimon’s ascent through American finance has been marked by both remarkable achievements and notable setbacks, culminating in a position where he is widely regarded as the dominant banking executive of his generation.

Dimon earned his bachelor’s degree from Tufts University in 1978 before completing an MBA at Harvard Business School in 1982. His career began with a brief stint as a management consultant at Boston Consulting Group, followed by his entry into American Express, where he worked under the mentorship of Sandy Weill—a relationship that would prove formative. At the age of 30, Dimon was appointed chief financial officer of Commercial Credit, later becoming the firm’s president. This role placed him at the centre of an aggressive acquisition strategy that included purchasing Primerica Corporation in 1987 and The Travelers Corporation in 1993.

From 1990 to 1998, Dimon served as Chief Operating Officer of both Travelers and Smith Barney, eventually becoming Co-Chairman and Co-CEO of the combined brokerage following the 1997 merger of Smith Barney and Salomon Brothers. When Travelers Group merged with Citicorp in 1998 to form Citigroup, Dimon was named president of the newly created financial services giant. However, his tenure proved short-lived; he departed later that year following a conflict with Weill over leadership succession.

This professional setback led to what would become one of the defining chapters of Dimon’s career. In 2000, he was appointed CEO of Bank One, a struggling institution that required substantial turnaround efforts. When JPMorgan Chase merged with Bank One in July 2004, Dimon became president and chief operating officer of the combined entity. He assumed the role of CEO on 1 January 2006, and one year later was named Chairman of the Board.

Under Dimon’s leadership, JPMorgan Chase navigated the 2008 financial crisis with relative success, earning him recognition as one of the few banking chiefs to emerge from the period with an enhanced reputation. As Duff McDonald wrote in his 2009 book “Last Man Standing: The Ascent of Jamie Dimon and JPMorgan Chase,” whilst much of the crisis stemmed from “plain old avarice and bad judgment,” Dimon and JPMorgan Chase “stood apart,” embodying “the values of clarity, consistency, integrity, and courage”.

Not all has been smooth sailing. In May 2012, JPMorgan Chase reported losses of at least $2 billion from trades that Dimon characterised as “flawed, complex, poorly reviewed, poorly executed and poorly monitored”—an episode that became known as the “London Whale” incident and attracted investigations from the Federal Reserve, SEC and FBI. In May 2023, Dimon testified under oath in lawsuits accusing the bank of serving Jeffrey Epstein, the late sex offender who was a client between 1998 and 2013.

Dimon’s political evolution reflects a pragmatic centrism. Having donated more than $500,000 to Democratic candidates between 1989 and 2009 and maintained close ties to the Obama administration, he later distanced himself from strict partisan identification. “My heart is Democratic,” he told CNBC in 2019, “but my brain is kind of Republican.” He primarily identifies as a “capitalist” and a “patriot,” and served on President Donald Trump’s short-lived business advisory council before Trump disbanded it in 2017. Though he confirmed in 2016 that he would “love to be president,” he deemed a campaign “too hard and too late” and ultimately decided against serious consideration of a 2020 run. In 2024, he endorsed Nikki Haley in the Republican primary before speaking more positively about Trump following Haley’s defeat.

As of May 2025, Forbes estimated Dimon’s net worth at $2.5 billion. He serves on the boards of numerous organisations, including the Business Roundtable, Bank Policy Institute and Harvard Business School, whilst also sitting on the executive committee of the Business Council and the Partnership for New York City.

Leading Theorists on AI and Labour Displacement

The question of how artificial intelligence will reshape employment has occupied economists, technologists and social theorists for decades, producing a rich body of work that frames Dimon’s observations within broader academic and policy debates.

John Maynard Keynes introduced the concept of “technological unemployment” in his 1930 essay “Economic Possibilities for our Grandchildren,” arguing that society was “being afflicted with a new disease” caused by “our discovery of means of economising the use of labour outrunning the pace at which we can find new uses for labour.” Keynes predicted this would be a temporary phase, ultimately leading to widespread prosperity and reduced working hours. His framing established the foundation for understanding technological displacement as a transitional phenomenon requiring societal adaptation rather than permanent catastrophe.

Joseph Schumpeter developed the theory of “creative destruction” in his 1942 work “Capitalism, Socialism and Democracy,” arguing that innovation inherently involves the destruction of old economic structures alongside the creation of new ones. Schumpeter viewed this process as the essential fact about capitalism—not merely a side effect but the fundamental engine of economic progress. His work provides the theoretical justification for Dimon’s observation about the internet bubble: widespread failure and waste can coexist with transformative innovation and aggregate productivity gains.

Wassily Leontief, winner of the 1973 Nobel Prize in Economics, warned in 1983 that workers might follow the path of horses, which were displaced en masse by automobable and tractor technology in the early twentieth century. His input-output economic models attempted to trace how automation would ripple through interconnected sectors, suggesting that technological displacement might be more comprehensive than previous episodes. Leontief’s scepticism about labour’s ability to maintain bargaining power against capital in an automated economy presaged contemporary concerns about inequality and the distribution of AI’s benefits.

Erik Brynjolfsson and Andrew McAfee at MIT have produced influential work on digital transformation and employment. Their 2014 book “The Second Machine Age” argued that we are in the early stages of a transformation as profound as the Industrial Revolution, with digital technologies now able to perform cognitive tasks previously reserved for humans. They coined the term “skill-biased technological change” to describe how modern technologies favour workers with higher levels of education and adaptability, potentially exacerbating income inequality. Their subsequent work on “machine learning” and “AI and the modern productivity paradox” has explored why measured productivity gains have lagged behind apparent technological advances—a puzzle relevant to Dimon’s observation that some AI benefits are difficult to quantify precisely.

Daron Acemoglu at MIT has challenged technological determinism, arguing that the impact of AI on employment depends crucially on how the technology is designed and deployed. In his 2019 paper “Automation and New Tasks: How Technology Displaces and Reinstates Labor” (co-authored with Pascual Restrepo), Acemoglu distinguished between automation that merely replaces human labour and technologies that create new tasks and roles. He has advocated for “human-centric AI” that augments rather than replaces workers, and has warned that current tax structures and institutional frameworks may be biasing technological development towards excessive automation. His work directly addresses Dimon’s categorisation of AI applications: some will enhance existing jobs, others will eliminate them, and the balance between these outcomes is not predetermined.

Carl Benedikt Frey and Michael Osborne at Oxford produced a widely cited 2013 study estimating that 47 per cent of US jobs were at “high risk” of automation within two decades. Their methodology involved assessing the susceptibility of 702 occupations to computerisation based on nine key bottlenecks, including creative intelligence, social intelligence and perception and manipulation. Whilst their headline figure attracted criticism for potentially overstating the threat—since many jobs contain a mix of automatable and non-automatable tasks—their framework remains influential in assessing which roles face displacement pressure.

Richard Freeman at Harvard has explored the institutional and policy responses required to manage technological transitions, arguing that the distribution of AI’s benefits depends heavily on labour market institutions, educational systems and social policy choices. His work emphasises that historical episodes of technological transformation involved substantial political conflict and institutional adaptation, suggesting that managing AI’s impact will require deliberate policy interventions rather than passive acceptance of market outcomes.

Shoshana Zuboff at Harvard Business School has examined how digital technologies reshape not merely what work is done but how it is monitored, measured and controlled. Her concept of “surveillance capitalism” highlights how data extraction and algorithmic management may fundamentally alter the employment relationship, potentially creating new forms of workplace monitoring and performance pressure even for workers whose jobs are augmented rather than eliminated by AI.

Klaus Schwab, founder of the World Economic Forum, has framed current technological change as the “Fourth Industrial Revolution,” characterised by the fusion of technologies blurring lines between physical, digital and biological spheres. His 2016 book of the same name argues that the speed, scope and systems impact of this transformation distinguish it from previous industrial revolutions, requiring unprecedented coordination between governments, businesses and civil society.

The academic consensus, insofar as one exists, suggests that AI will indeed transform employment substantially, but that the nature and distributional consequences of this transformation remain contested and dependent on institutional choices. Dimon’s advice to avoid “putting your head in the sand” and to stay “way ahead of the curve” aligns with this literature’s emphasis on anticipatory adaptation. His commitment to retraining and redeployment echoes the policy prescriptions of economists who argue that managing technological transitions requires active human capital investment rather than passive acceptance of labour market disruption.

What distinguishes Dimon’s perspective is his position as a practitioner implementing these technologies at scale within a major institution. Whilst theorists debate aggregate employment effects and optimal policy responses, Dimon confronts the granular realities of deployment: which specific functions can be augmented versus automated, how managers adapt their decision-making processes, what training programmes prove effective, and how to balance efficiency gains against workforce morale and capability retention. His assertion that JPMorgan has achieved approximately $2 billion in quantifiable benefits from $2 billion in annual AI spending—whilst acknowledging additional unquantifiable improvements—provides an empirical data point for theories about AI’s productivity impact.

The ten-year timeframe of JPMorgan’s AI journey also matters. Dimon’s observation that “people think it’s a new thing” but that the bank has been pursuing AI since 2012 challenges narratives of sudden disruption, instead suggesting a more gradual but accelerating transformation. This accords with Brynjolfsson and McAfee’s argument about the “productivity J-curve”—that the full economic benefits of transformative technologies often arrive with substantial lag as organisations learn to reconfigure processes and business models around new capabilities.

Ultimately, Dimon’s warning about job displacement, combined with his emphasis on staying ahead of the curve through retraining and redeployment, reflects a synthesis of Schumpeterian creative destruction, human capital theory, and practical experience managing technological change within a complex organisation. His perspective acknowledges both the inevitability of disruption and the possibility of managing transitions to benefit both institutions and workers—provided leadership acts proactively rather than reactively. For financial services professionals and business leaders more broadly, Dimon’s message is clear: AI’s impact on employment is neither hypothetical nor distant, but rather an ongoing transformation requiring immediate and sustained attention.

read more
Quote: Yann LeCun – Chief AI Scientist at Meta

Quote: Yann LeCun – Chief AI Scientist at Meta

“Before we reach human-level AI, we will have to reach cat-level AI and dog-level AI.” – Yann LeCun – Chief AI Scientist at Meta

Yann LeCun, a pioneering figure in artificial intelligence, is globally recognized for his foundational contributions to deep learning and neural networks. As the Chief AI Scientist at Meta (formerly Facebook) and a Silver Professor at New York University’s Courant Institute, LeCun has been instrumental in advancing technologies that underlie today’s AI systems, including convolutional neural networks (CNNs), which are now fundamental to image and pattern recognition in both industry and research.

LeCun’s journey in AI began in the late 1980s, when much of the scientific community considered neural networks to be a dead end. Undeterred, LeCun, alongside peers such as Geoffrey Hinton and Yoshua Bengio, continued to develop these models, ultimately proving their immense value. His early successes included developing neural networks capable of recognizing handwritten characters—a technology that became widely used by banks for automated check reading by the late 1990s.This unwavering commitment to neural networks earned LeCun, Hinton, and Bengio the 2018 Turing Award, often dubbed the “Nobel Prize of Computing,” and solidified their standing as the “Godfathers of AI”.

The quote, “Before we reach human-level AI, we will have to reach cat-level AI and dog-level AI,” encapsulates LeCun’s pragmatic approach to artificial intelligence. He emphasizes that replicating the full suite of human cognitive abilities is a long-term goal—one that cannot be achieved without first creating machines that can perceive, interpret, and interact with the world with the flexibility, intuition, and sensory-motor integration seen in animals like cats and dogs. Unlike current AI, which excels in narrow, well-defined tasks, a cat or a dog can navigate complex, uncertain environments, learn from limited experience, and adapt fluidly—capabilities that still elude artificial agents. LeCun’s perspective highlights the importance of incremental progress in AI: only by mastering the subtleties of animal intelligence can we aspire to build machines that match or surpass human cognition.

LeCun’s work continues to shape how researchers and industry leaders think about the future of AI—not as an overnight leap to artificial general intelligence, but as a gradual journey through, and beyond, the marvels of natural intelligence found throughout the animal kingdom.

read more
Term: AI Inference

Term: AI Inference

AI inference refers to the process in which a trained artificial intelligence (AI) or machine learning model analyzes new, unseen data to make predictions or decisions. After a model undergoes training—learning patterns, relationships, or rules from labeled datasets—it enters the inference phase, where it applies that learned knowledge to real-world situations or fresh inputs.

This process typically involves the following steps:

  • Training phase: The model is exposed to large, labeled datasets (for example, images with known categories), learning to recognize key patterns and features.
  • Inference phase: The trained model receives new data (such as an unlabeled image) and applies its knowledge to generate a prediction or decision (like identifying objects within the image).

AI inference is fundamental because it operationalizes AI, enabling it to be embedded into real-time applications such as voice assistants, autonomous vehicles, medical diagnosis tools, and fraud detection systems. Unlike the resource-intensive training phase, inference is generally optimized for speed and efficiency—especially important for tasks on edge devices or in situations requiring immediate results.

As generative and agent-based AI applications mature, the demand for faster and more scalable inference is rapidly increasing, driving innovation in both software and hardware to support these real-time or high-volume use cases.

A major shift in AI inference is occurring as new elements—such as test time compute (TTC), chain-of-thought reasoning, and adaptive inference—reshape how and where computational resources are allocated in AI systems.

Expanded Elements in AI Inference

  • Test-Time Compute (TTC): This refers to the computational effort expended during inference rather than during initial model training. Traditionally, inference consisted of a single, fast forward pass through the model, regardless of the complexity of the question. Recent advances, particularly in generative AI and large language models, involve dynamically increasing compute at inference time for more challenging problems. This allows the model to “think harder” by performing additional passes, iterative refinement, or evaluating multiple candidate responses before selecting the best answer

  • Chain-of-Thought Reasoning: Modern inference can include step-by-step reasoning, where models break complex problems into sub-tasks and generate intermediate steps before arriving at a final answer. This process may require significantly more computation during inference, as the model deliberates and evaluates alternative solutions—mimicking human-like problem solving rather than instant pattern recognition.

  • Adaptive Compute Allocation: With TTC, AI systems can allocate more resources dynamically based on the difficulty or novelty of the input. Simple questions might still get an immediate, low-latency response, while complex or ambiguous tasks prompt the model to use additional compute cycles for deeper reasoning and improved accuracy.

Impact: Shift in Compute from Training to Inference

  • From Heavy Training to Intelligent Inference: The traditional paradigm put most of the computational burden and cost on the training phase, after which inference was light and static. With TTC and chain-of-thought reasoning, more computation shifts into the inference phase. This makes inference more powerful and flexible, allowing for real-time adaptation and better performance on complex, real-world tasks without the need for ever-larger model sizes.

  • Strategic and Operational Implications: This shift enables organizations to optimize resources by focusing on smarter, context-aware inference rather than continually scaling up training infrastructure. It also allows for more responsive AI systems that can improve decision-making and user experiences in dynamic environments.

  • Industry Adoption: Modern models from leading labs (such as OpenAI and Google’s Gemini) now support iterative, compute-intensified inference modes, yielding substantial gains on benchmarks and real-world applications, especially where deep reasoning or nuanced analysis is required.

These advancements in test time compute and reasoned inference mark a pivotal transformation in AI, moving from static, single-pass prediction to dynamic, adaptive, and resource-efficient problem-solving at the moment of inference.

Related strategy theorist: Yann LeCun

Yann LeCun is widely recognized as a pioneering theorist in neural networks and deep learning—the foundational technologies underlying modern AI inference. His contributions to convolutional neural networks and strategies for scalable, robust AI learning have shaped the current landscape of AI deployment and inference capabilities.

“AI inference is the core mechanism by which machine learning models transform training into actionable intelligence, supporting everything from real-time analysis to agent-based automation.”

Yann LeCun is a French-American computer scientist and a foundational figure in artificial intelligence, especially in the areas of deep learning, computer vision, and neural networks. Born on July 8, 1960, in Soisy-sous-Montmorency, France, he received his Diplôme d’Ingénieur from ESIEE Paris in 1983 and earned his PhD in Computer Science from Sorbonne University (then Université Pierre et Marie Curie) in 1987. His doctoral research introduced early methods for back-propagation in neural networks, foreshadowing the architectures that would later revolutionize AI.

LeCun began his research career at the Centre National de la Recherche Scientifique (CNRS) in France, focusing on computer vision and image recognition. His expertise led him to postdoctoral work at the University of Toronto, where he collaborated with other leading minds in neural networks. In 1988, he joined AT&T Bell Laboratories in New Jersey, eventually becoming head of the Image Processing Research Department. There, LeCun led the development of convolutional neural networks (CNNs), which became the backbone for modern image and speech recognition systems. His technology for handwriting and character recognition was widely adopted in banking, reading a significant share of checks in the U.S. in the early 2000s.

LeCun also contributed to the creation of DjVu, a high-efficiency image compression technology, and the Lush programming language. In 2003, he became a professor at New York University (NYU), where he founded the NYU Center for Data Science, advancing interdisciplinary AI research.

In 2013, LeCun became Director of AI Research at Facebook (now Meta), where he leads the Facebook AI Research (FAIR) division, focusing on both theoretical and applied AI at scale. His leadership at Meta has pushed forward advancements in self-supervised learning, agent-based systems, and the practical deployment of deep learning technologies.

LeCun, along with Yoshua Bengio and Geoffrey Hinton, received the 2018 Turing Award—the highest honor in computer science—for his pioneering work in deep learning. The trio is often referred to as the “Godfathers of AI” for their collective influence on the field.

 

Yann LeCun’s Thinking and Approach

LeCun’s intellectual focus is on building intelligent systems that can learn from data efficiently and with minimal human supervision. He strongly advocates for self-supervised and unsupervised learning as the future of AI, arguing that these approaches best mimic how humans and animals learn. He believes that for AI to reach higher forms of reasoning and perception, systems must be able to learn from raw, unlabeled data and develop internal models of the world.

LeCun is also known for his practical orientation—developing architectures (like CNNs) that move beyond theory to solve real-world problems efficiently. His thinking consistently emphasizes the importance of scaling AI not just through bigger models, but through more robust, data-efficient, and energy-efficient algorithms.

He has expressed skepticism about narrow, brittle AI systems that rely heavily on supervised learning and excessive human labeling. Instead, he envisions a future where AI agents can learn, reason, and plan with broader autonomy, similar to biological intelligence. This vision guides his research and strategic leadership in both academia and industry.

LeCun remains a prolific scientist, educator, and spokesperson for responsible and open AI research, championing collaboration and the broad dissemination of AI knowledge.

read more
Quote: Andrew Ng – AI Guru

Quote: Andrew Ng – AI Guru

“For the majority of businesses, focus on building applications using agentic workflows rather than solely scaling traditional AI. That’s where the greatest opportunity lies.” – Andrew Ng – AI Guru

Andrew Ng is widely recognized as a pioneering figure in artificial intelligence, renowned for his roles as co-founder of Google Brain, former chief scientist at Baidu, and founder of DeepLearning.AI and Landing AI. His work has shaped the trajectory of modern AI, influencing its academic, industrial, and entrepreneurial development on a global scale.

The quote “For the majority of businesses, focus on building applications using agentic workflows rather than solely scaling traditional AI. That’s where the greatest opportunity lies.” captures a key transformation underway in how organizations approach AI adoption. Ng delivered this insight during a Luminary Talk at the Snowflake Summit in June 2024, in a discussion centered on the rise of agentic workflows within AI applications.

Historically, businesses have harnessed AI by leveraging static, rule-based automation or applying large language models to single-step tasks—prompting a system to generate a document or answer a question in one go. Ng argues this paradigm is now giving way to a new era driven by AI agents capable of multi-step reasoning, planning, tool use, and collaboration—what he terms “agentic workflows”.

Agentic workflows differ from traditional approaches by allowing autonomous AI agents to adapt, break down complex projects, and iterate in real time, much as a human team might tackle a multifaceted problem. For example, instead of a single prompt generating a sales report, an AI agent in an agentic workflow could gather the relevant data, perform analysis, adjust its approach based on interim findings, and refine the output after successive rounds of review and self-critique. Ng has highlighted design patterns such as reflection, planning, multi-agent collaboration, and dynamic tool use as central to these workflows.

Ng’s perspective is that businesses stand to gain the most not merely from increasing the size or data intake of AI models, but from designing systems where AI agents can independently coordinate and accomplish sophisticated goals. He likens this shift to the leap from single-threaded to multi-threaded computing, opening up exponential gains in capability and value creation.

For business leaders, Andrew Ng’s vision offers a roadmap: the frontier of competitive advantage lies in reimagining how AI-powered agents are integrated into business processes, unlocking new possibilities for efficiency, innovation, and scalability that go beyond what traditional, “one-shot” AI can deliver.

Ng continues to lead at the intersection of AI innovation and practical business strategy, championing agentic AI as the next great leap for organizations seeking to realize the full promise of artificial intelligence.

read more
Term: AI Agents

Term: AI Agents

AI Agents are autonomous software systems that interact with their environment, perceive data, and independently make decisions and take actions to achieve specific, user-defined goals. Unlike traditional software, which follows static, explicit instructions, AI agents are guided by objective functions and have the ability to reason, learn, plan, adapt, and optimize responses based on real-time feedback and changing circumstances.

Key characteristics of AI agents include:

  • Autonomy: They can initiate and execute actions without constant human direction, adapting as new data or situations arise.
  • Rational decision-making: AI agents use data and perceptions of their environment to select actions that maximize predefined goals or rewards (their “objective function”), much like rational agents in economics.
  • Learning and Adaptation: Through techniques like machine learning, agents improve their performance over time by learning from experience.
  • Multimodal abilities: Advanced agents process various types of input/output—text, audio, video, code, and more—and often collaborate with humans or other agents to complete complex workflows or transactions.
  • Versatility: They range from simple (like thermostats) to highly complex systems (like conversational AI assistants or autonomous vehicles).

Examples include virtual assistants that manage calendars or customer support, code-review bots in software development, self-driving cars navigating traffic, and collaborative agents that orchestrate business processes.

Related Strategy Theorist – Stuart Russell

As a renowned AI researcher and co-author of the seminal textbook “Artificial Intelligence: A Modern Approach,” Russell has shaped foundational thinking on agent-based systems and rational decision-making. He has also been at the forefront of advocating for the alignment of agent objectives with human values, providing strategic frameworks for deploying autonomous agents safely and effectively across industries.

read more
Quote: Ilya Sutskever – Safe Superintelligence

Quote: Ilya Sutskever – Safe Superintelligence

“AI will do all the things that we can do. Not just some of them, but all of them. The big question is what happens then: Those are dramatic questions… the rate of progress will become really extremely fast for some time at least, resulting in unimaginable things. And in some sense, whether you like it or not, your life is going to be affected by AI to a great extent.” –  Ilya Sutskever – Safe Superintelligence

Ilya Sutskever stands among the most influential figures shaping the modern landscape of artificial intelligence. Born in Russia and raised in Israel and Canada, Sutskever’s early fascination with mathematics and computer programming led him to the University of Toronto, where he studied under the legendary Geoffrey Hinton. His doctoral work broke new ground in deep learning, particularly in developing recurrent neural networks and sequence modeling—technologies that underpin much of today’s AI-driven language and translation systems.

Sutskever’s career is marked by a series of transformative achievements. He co-invented AlexNet, a neural network that revolutionized computer vision and triggered the deep learning renaissance. At Google Brain, he advanced sequence-to-sequence models, laying the foundation for breakthroughs in machine translation. As a co-founder and chief scientist at OpenAI, Sutskever played a pivotal role in developing the GPT series of language models, which have redefined what machines can achieve in natural language understanding and generation.

Beyond his technical contributions, Sutskever is recognized for his thought leadership on the societal implications of AI. He has consistently emphasized the unpredictable nature of advanced AI systems, particularly as they acquire reasoning capabilities that may outstrip human understanding. His recent work focuses on AI safety and alignment, co-founding Safe Superintelligence Inc. to ensure that future superintelligent systems act in ways beneficial to humanity.

The quote featured today encapsulates Sutskever’s vision: a world where AI’s capabilities will extend to all domains of human endeavor, bringing about rapid and profound change. For business leaders and strategists, his words are both a warning and a call to action—highlighting the necessity of anticipating technological disruption and embracing innovation at a pace that matches AI’s accelerating trajectory.

read more
Term: Artificial General Intelligence (AGI)

Term: Artificial General Intelligence (AGI)

Artificial General Intelligence (AGI) is defined as a form of artificial intelligence that can understand, learn, and apply knowledge across the full spectrum of human cognitive tasks—matching or even exceeding human capabilities in any intellectual endeavor. Unlike current artificial intelligence systems, which are typically specialized (known as narrow AI) and excel only in specific domains such as language translation or image recognition, AGI would possess the versatility and adaptability of the human mind.

AGI enables machines to perform essentially all human cognitive tasks at or above top human expert level, acquire new skills, and transfer its capabilities to entirely new domains, embodying a level of intelligence no single human possesses—rather, it would represent the combined expertise of top minds across all fields.

Alternative Name – Superintelligence:
The term superintelligence or Artificial Superintelligence (ASI) refers to an intelligence that not only matches but vastly surpasses human abilities in virtually every aspect. While AGI is about equaling human-level intelligence, superintelligence describes systems that can independently solve problems, create knowledge, and innovate far beyond even the best collective human intellect.

 
Level
Description
Narrow AI
Specialized systems that perform limited tasks (e.g., playing chess, image recognition)
AGI
Systems with human-level cognitive abilities across all domains, adaptable and versatile
Superintelligence
Intelligence that exceeds human capabilities in all domains, potentially by wide margins

Key contrasts between AGI and (narrow) AI:

  • Scope: AGI can generalize across different tasks and domains; narrow AI is limited to narrowly defined problems.
  • Learning and Adaptation: AGI learns and adapts to new situations much as humans do, while narrow AI cannot easily transfer skills to new, unfamiliar domains.
  • Cognitive Sophistication: AGI mimics the full range of human intelligence; narrow AI does not.
 

Strategy Theorist — Ilya Sutskever:
Ilya Sutskever is a leading figure in the pursuit of AGI, known for his foundational contributions to deep learning and as a co-founder of OpenAI. Sutskever’s work focuses on developing models that move beyond narrow applications toward truly general intelligence, shaping both the technical roadmap and ethical debate around AGI’s future.

Ilya Sutskever’s views on the impact of superintelligence are characterized by a blend of optimism for its transformative potential and deep caution regarding its unpredictability and risks. Sutskever believes superintelligence could revolutionize industries, particularly healthcare, and deliver unprecedented economic, social, and scientific breakthroughs within the next decade. He foresees AI as a force that can solve complex problems and dramatically extend human capabilities. For business, this implies radical shifts: automating sophisticated tasks, generating new industries, and redefining competitive advantages as organizations adapt to a new intelligence landscape.

However, Sutskever consistently stresses that the rise of superintelligent AI is “extremely unpredictable and unimaginable,” warning that its self-improving nature could quickly move beyond human comprehension and control. He argues that while the rewards are immense, the risks—including loss of human oversight and the potential for misuse or harm—demand proactive, ethical, and strategic guidance. Sutskever champions the need for holistic thinking and interdisciplinary engagement, urging leaders and society to prepare for AI’s integration not with fear, but with ethical foresight, adaptation, and resilience.

He has prioritized AI safety and “superalignment” as central to his strategies, both at OpenAI and through his new Safe Superintelligence venture, actively seeking mechanisms to ensure that the economic and societal gains from superintelligence do not come at unacceptable risks. Sutskever’s message for corporate leaders and policymakers is to engage deeply with AI’s trajectory, innovate responsibly, and remain vigilant about both its promise and its perils.

In summary, AGI is the milestone where machines achieve general, human-equivalent intelligence, while superintelligence describes a level of machine intelligence that greatly surpasses human performance. The pursuit of AGI, championed by theorists like Ilya Sutskever, represents a profound shift in both the potential and challenges of AI in society.

read more
Quote:  Tom Davenport — Academic, consultant, author

Quote: Tom Davenport — Academic, consultant, author

“AI doesn’t replace strategic thinking—it accelerates it.” — Tom Davenport — Academic, consultant, author

Tom Davenport’s quote captures the essence of the relationship between human judgment and advances in artificial intelligence. Davenport, a leading authority on analytics and business process innovation, has spent decades studying how organizations make decisions and adopt new technologies.

As AI systems have rapidly evolved—from early rule-based approaches to today’s powerful generative models—their promise is often misunderstood. Some fear AI might make human thinking obsolete, especially in complex arenas like strategy. Davenport has consistently challenged this notion. He argues that AI’s true value lies in amplifying, not eliminating, the need for rigorous, creative, and forward-looking thought. AI is a tool that enables strategists to test more ideas, analyze larger datasets, and see farther into future possibilities—but it is strategic thinking, shaped by human experience and ambition, that guides AI toward meaningful goals.

Davenport’s perspective is grounded in his extensive work with businesses and his scholarship at leading universities. In his conversations and writings, he notes that while AI democratizes access to information and automates routine analysis, a competitive edge still hinges on asking the right questions and crafting distinctive strategies. The leaders who thrive in the AI era are those who learn to harness its speed and breadth, using it to accelerate the cycles of planning, validation, and innovation rather than replace the uniquely human qualities of insight and judgment.

About Tom Davenport

Tom Davenport, born in 1954, is an influential American academic, business consultant, and author. He specializes in analytics, business process innovation, and knowledge management. Davenport is well-known for his pioneering books such as Competing on Analytics and his widely-cited research on how organizations create value from data. Affiliated with prestigious institutions, he has helped shape how leaders think about information, technology, and business transformation.

Davenport’s views on AI are informed by years of advising Fortune 500 companies, conducting academic research, and contributing to thought leadership at the intersection of technology and management. His insights have been instrumental in helping organizations adapt to the changing landscape of digital innovation, emphasizing that technology serves best when paired with human creativity, analytical rigor, and strategic vision

read more
Quote:  Ginni Rometty, Former IBM CEO

Quote: Ginni Rometty, Former IBM CEO

“Artificial intelligence is not a strategy, but a means to rethink your strategy.” — Ginni Rometty, Former IBM CEO

Ginni Rometty’s statement, “Artificial intelligence is not a strategy, but a means to rethink your strategy,” emerged from her front-row vantage point in one of the era’s most significant technological transformations. As the first woman to serve as chairman, president, and CEO of IBM, Rometty’s nearly four-decade career at the company offers a compelling backdrop to her insight.

Her leadership at IBM began in 2012, at a time when the company confronted industry-wide disruption driven by the rise of cloud computing, big data, and artificial intelligence. Rometty recognized early on that AI—while transformative—was not a plug-and-play solution, but a set of tools that could empower organizations to fundamentally reshape their approaches to competition, operations, and growth. This realization guided IBM’s pivot toward cognitive computing, analytics, and cloud-based solutions during her tenure.

A defining episode during Rometty’s leadership was IBM’s acquisition of the open-source powerhouse Red Hat for $34 billion—a strategic move to anchor IBM’s transition into the cloud era and enable clients to rethink how they deliver value in increasingly digital markets. Throughout these changes, Rometty was adamant: adopting technologies like AI is not an end in itself but a catalyst for critically reexamining and reinventing business strategies.

The quote distills her conviction that simply acquiring cutting-edge technology is not sufficient. Instead, success depends on leaders’ willingness to challenge old assumptions and design new strategies that fully leverage the potential of AI. Rometty’s perspective, forged by navigating IBM through turbulent shifts, underscores the necessity of using innovation to reimagine, not merely digitize, the future of enterprise.

About Ginni Rometty

Ginni Rometty, born in 1957, joined IBM as a systems engineer in 1981 and steadily advanced through key leadership roles—culminating in her appointment as CEO from 2012 to 2020. During her tenure, she spearheaded bold decisions: negotiating the purchase of PricewaterhouseCoopers’ IT consulting business in 2002, prioritizing investments in cloud, analytics, and cognitive computing, and repositioning IBM for the demands and opportunities of the modern digital landscape.

Her leadership style and vision earned her recognition among Bloomberg’s 50 Most Influential People in the World, Fortune’s “50 Most Powerful Women in Business,” and Forbes’ Top 50 Women in Tech. While her tenure included periods of financial challenge and criticism over IBM’s performance, Rometty’s overarching legacy is her focus on transformation—seeing technology as a lever for reinventing strategy, not merely executing it.

This context enriches the meaning of her quote, highlighting its origins in both lived experience and hard-won leadership insight.

read more
Quote:  Andrew Ng, AI guru

Quote: Andrew Ng, AI guru

“In the age of AI, strategy is no longer just about where to play; it’s about how to adapt.” — Andrew Ng, AI guru

This quote from Andrew Ng captures a profound shift in how organizations and leaders must approach strategy in the era of artificial intelligence. Traditionally, strategic planning has focused on identifying the right markets, customers, or products—the “where to play” aspect. However, as AI rapidly transforms industries, Ng argues that the ability to adapt to ongoing technological changes has become just as crucial, if not more so.

The background for this perspective stems from Ng’s deep involvement in the practical deployment of AI at scale. With advances in machine learning and automation, the competitive landscape is continuously evolving. It is no longer enough to set a single strategic direction; leaders need to develop organizational agility to embrace new technologies and iterate their models, processes, and offerings in response to rapid change. Ng’s message emphasizes that AI is not a static tool, but a disruptive force that requires companies to rethink how they respond to uncertainty and opportunity. This shift from fixed planning to adaptive learning mirrors the very nature of AI systems themselves, which are designed to learn, update, and improve over time.

Ng’s insight also reflects his broader view that AI should be used to automate routine tasks, freeing up human talent to focus on creative, strategic, and adaptive functions. As such, the modern strategic imperative is about continually repositioning and reinventing—not just staking out a position and defending it.

About Andrew Ng

Andrew Ng is one of the world’s most influential figures in artificial intelligence and machine learning. Born in 1976, he is a British-American computer scientist and technology entrepreneur. Ng co-founded Google Brain, where he played a pivotal role in advancing deep learning research, and later served as Chief Scientist at Baidu, leading a large AI group. He is also a prominent educator, co-founding Coursera and creating widely popular online courses that have democratized access to AI knowledge for millions worldwide.

Ng has consistently advocated for practical, human-centered adoption of AI. He introduced the widely referenced idea that “AI is the new electricity,” underscoring its foundational and transformative impact across industries. He has influenced both startups and established enterprises through initiatives such as Landing AI and the AI Fund, which focus on applying AI to real-world problems and fostering AI entrepreneurship.

Andrew Ng is known for his clear communication and balanced perspective on the opportunities and challenges of AI. Recognized globally for his contributions, he has been named among Time magazine’s 100 Most Influential People and continues to shape the trajectory of AI through his research, teaching, and thought leadership. His work encourages businesses and individuals alike to not only adopt AI technologies, but to cultivate the adaptability and critical thinking needed to thrive in an age of constant change.

read more
Quote: Daniel Kahneman, Nobel Laureate

Quote: Daniel Kahneman, Nobel Laureate

“AI is great at multitasking: it can misunderstand five tasks at once.” — Daniel Kahneman, Nobel Laureate

This wry observation from Daniel Kahneman highlights the persistent gap between expectation and reality in the deployment of artificial intelligence. As AI systems increasingly promise to perform multiple complex tasks—ranging from analyzing data and interpreting language to making recommendations—there remains a tendency to overestimate their capacity for genuine understanding. Kahneman’s quote playfully underscores how, far from being infallible, AI can compound misunderstandings when juggling several challenges simultaneously.

The context for this insight is rooted in Kahneman’s lifelong exploration of the limits of decision-making—first in humans, and, by extension, in the systems designed to emulate or augment human judgment. AI’s appeal often stems from its speed and apparent ability to handle many tasks at once. However, as with human cognition, multitasking can amplify errors if the underlying comprehension is lacking or the input data is ambiguous. Kahneman’s expertise in uncovering the predictable errors and cognitive biases that affect human reasoning makes his skepticism toward AI’s supposed multitasking prowess particularly telling. The remark serves as a reminder to remain critical and measured in evaluating AI’s true capabilities, especially in contexts where precision and nuance are essential.

About Daniel Kahneman

Daniel Kahneman (1934–2024) was an Israeli-American psychologist whose groundbreaking work revolutionized the understanding of human judgment, decision-making, and the psychology of risk. Awarded the 2002 Nobel Memorial Prize in Economic Sciences, he was recognized “for having integrated insights from psychological research into economic science, especially concerning human judgment and decision-making under uncertainty”.

Together with collaborator Amos Tversky, Kahneman identified a series of cognitive heuristics and biases—systematic errors in thinking that affect the way people judge probabilities and make decisions. Their work led to the development of prospect theory, which challenged the traditional economic view that humans are rational actors, and established the foundation of behavioral economics.

Kahneman’s research illuminated how individuals routinely overgeneralize from small samples, fall prey to stereotyping, and exhibit overconfidence—even when handling simple probabilities. His influential book, Thinking, Fast and Slow, distilled decades of research into a compelling narrative about how the mind works, the pitfalls of intuition, and the enduring role of error in human reasoning.

In his later years, Kahneman continued to comment on the limitations of decision-making processes, increasingly turning his attention to how these limits inform the development and evaluation of artificial intelligence. His characteristic blend of humor and rigor, as exemplified in the quoted observation about AI multitasking, continues to inspire thoughtful scrutiny of technology and its role in society.

read more
Quote:  Andrew Ng, AI guru

Quote: Andrew Ng, AI guru

“AI is like teenage sex—everyone talks about it, nobody really knows how to do it.” — Andrew Ng, AI guru

Andrew Ng, captures the sense of hype, confusion, and uncertainty that has often surrounded artificial intelligence (AI) in recent years. Delivered with humor, it reflects the atmosphere in which AI has become a buzzword: widely discussed in boardrooms, newsrooms, and tech circles, yet rarely understood in its real-world applications or complexities.

The backdrop to this quote is the rapid growth in public and corporate interest in AI. From the early days of AI research in the mid-20th century, the field has experienced cycles of intense excitement (“AI springs”) and subsequent setbacks (“AI winters”), often fueled by unrealistic expectations and misunderstanding of the technology’s actual capabilities. In the last decade, as machine learning and deep learning began to make headlines with breakthroughs in image recognition, natural language processing, and game-playing, many organizations felt pressure to claim they were leveraging AI—regardless of whether they truly understood how to implement it or what it could achieve.

Ng’s remark wittily punctures the inflated discourse by suggesting that, like teenage sex, the reality of AI is far less straightforward than the bravado implies. It serves as both a caution and an invitation: to move beyond surface-level conversations and focus instead on genuine understanding and effective implementation.

About Andrew Ng

Andrew Ng is one of the most influential figures in artificial intelligence and machine learning. He is known for his clear-eyed optimism and his ability to communicate complex technical ideas in accessible language. Ng co-founded Google Brain, led Baidu’s AI Group, and launched the pioneering online machine learning course on Coursera, which has introduced AI to millions worldwide.

Ng frequently emphasizes AI’s transformative potential, famously stating that “AI is the new electricity”—suggesting that, much like electricity revolutionized industries in the past, AI will fundamentally change every sector in the coming decades. Beyond technical achievement, he advocates for practical and responsible adoption of AI, striving to bridge the gap between hype and meaningful progress.

His humorous comparison of AI discourse to teenage sex has become a memorable and oft-cited line at technology conferences and in articles. It encapsulates not only the social dynamics at play in emerging technological fields, but also Ng’s approachable style and his drive to demystify artificial intelligence for a broader audience

read more
Quote: Sholto Douglas, Anthropic researcher

Quote: Sholto Douglas, Anthropic researcher

“We believe coding is extremely important because coding is that first step in which you will see AI research itself being accelerated… We think it is the most important leading indicator of model capabilities.”

Sholto Douglas, Anthropic researcher

Sholto Douglas is regarded as one of the most promising new minds in artificial intelligence research. Having graduated from the University of Sydney with a degree in Mechatronic (Space) Engineering under the guidance of Ian Manchester and Stefan Williams, Douglas entered the field of AI less than two years ago, quickly earning respect for his innovative contributions. At Anthropic, one of the leading AI research labs, he specializes in scaling reinforcement learning (RL) techniques within advanced language models, focusing on pushing the boundaries of what large language models can learn and execute autonomously.

Context of the Quote

The quote, delivered by Douglas in an interview with Redpoint—a venture capital firm known for its focus on disruptive startups and technology—underscores the central thesis driving Anthropic’s recent research efforts:

“We believe coding is extremely important because coding is that first step in which you will see AI research itself being accelerated… We think [coding is] the most important leading indicator of model capabilities.”

This statement reflects both the technical philosophy and the strategic direction of Anthropic’s latest research. Douglas views coding not only as a pragmatic benchmark but as a foundational skill that unlocks model self-improvement and, by extension, accelerates progress toward artificial general intelligence (AGI).

Claude 4 Launch: Announcements and Impact

Douglas’ remarks came just ahead of the public unveiling of Anthropic’s Claude 4, the company’s most sophisticated model to date. The event highlighted several technical milestones:

  • Reinforcement Learning Breakthroughs: Douglas described how, over the past year, RL techniques in language models had evolved from experimental to demonstrably successful, especially in complex domains like competitive programming and advanced mathematics. For the first time, they achieved “proof of an algorithm that can give us expert human reliability and performance, given the right feedback loop”.
  • Long-Term Vision: The launch positioned coding proficiency as the “leading indicator” for broader model capabilities, setting the stage for future models that can meaningfully contribute to their own research and improvement.
  • Societal Implications: Alongside the technical announcements, the event and subsequent interviews addressed how rapid advances in AI—exemplified by Claude 4—will impact industries, labor markets, and global policy, urging stakeholders to prepare for a world where AI agents are not just tools but collaborative problem-solvers.
 

Why This Moment Matters

Douglas’ focus on coding as a metric is rooted in the idea that tasks requiring deep logic and creative problem-solving, such as programming, provide a “canary in the coal mine” for model sophistication. Success in these domains demonstrates a leap not only in computational power or data processing, but in the ability of AI models to autonomously reason, plan, and build tools that further accelerate their own learning cycles.

The Claude 4 launch, and Douglas’ role within it, marks a critical inflection point in AI research. The ability of language models to code at—or beyond—expert human levels signals the arrival of AI systems capable of iteratively improving themselves, raising both hopes for extraordinary breakthroughs and urgent questions around safety, alignment, and governance.

Sholto Douglas’ Influence

Though relatively new to the field, Douglas has emerged as a thought leader shaping Anthropic’s approach to scalable, interpretable, and safe AI. His insights bridge technical expertise and strategic foresight, providing a clear-eyed perspective on the trajectory of rapidly advancing language models and their potential to fundamentally reshape the future of research and innovation.

read more

Download brochure

Introduction brochure

What we do, case studies and profiles of some of our amazing team.

Download

Our latest podcasts on Spotify

Sign up for our newsletters - free

Global Advisors | Quantified Strategy Consulting