Select Page

Global Advisors | Quantified Strategy Consulting

training
Term: Diffusion models

Term: Diffusion models

“Diffusion models are a class of generative artificial intelligence (AI) models that create new data instances by learning to reverse a gradual, step-by-step process of adding noise to training data.” – Diffusion models

Diffusion models are a class of generative artificial intelligence models that create new data instances by learning to reverse a gradual, step-by-step process of adding noise to training data. They represent one of the most significant advances in machine learning, emerging as the dominant generative approach since the introduction of Generative Adversarial Networks in 2014.

Core Mechanism

Diffusion models operate through a dual-phase process inspired by non-equilibrium thermodynamics in physics. The mechanism mirrors the natural diffusion phenomenon, where molecules move from areas of high concentration to low concentration. In machine learning, this principle is inverted to generate high-quality synthetic data.

The process consists of two complementary components:

  • Forward diffusion process: Training data is progressively corrupted by adding Gaussian noise through a series of small, incremental steps. Each step introduces controlled complexity via a Markov chain, gradually transforming structured data into pure noise.
  • Reverse diffusion process: The model learns to reverse this noise-addition procedure, starting from random noise and iteratively removing it to reconstruct data that matches the original training distribution.

During training, the model learns to predict the noise added at each step of the forward process by minimising a loss function that measures the difference between predicted and actual noise. Once trained, the model can generate entirely new data by passing randomly sampled noise through the learned denoising process.

Key Components and Architecture

Three essential elements enable diffusion models to function effectively:

  • Forward diffusion process: Adds noise to data in successive small steps, with each iteration increasing randomness until the data resembles pure noise.
  • Reverse diffusion process: The neural network learns to iteratively remove noise, generating data that closely resembles training examples.
  • Score function: Estimates the gradient of the data distribution with respect to noise, guiding the reverse diffusion process to produce realistic samples.

A notable architectural advancement is the Latent Diffusion Model (LDM), which runs the diffusion process in latent space rather than pixel space. This approach significantly reduces training costs and accelerates inference speed by first compressing data with an autoencoder, then performing the diffusion process on learned semantic representations.

Advantages Over Alternative Approaches

Diffusion models offer several compelling advantages compared to competing generative models such as GANs and Variational Autoencoders (VAEs):

  • Superior image quality: They generate highly realistic images that closely match the distribution of real data, outperforming GANs through their distinct mechanisms for precise replication of real-world imagery.
  • Stable training: Unlike GANs, diffusion models avoid mode collapse and unstable training dynamics, providing a more reliable learning process.
  • Flexibility: They can model complex data distributions without requiring explicit likelihood estimation.
  • Theoretical foundations: Based on well-understood principles from stochastic processes and statistical mechanics, providing strong mathematical grounding.
  • Simple loss functions: Training employs straightforward and efficient loss functions that are easier to optimise.

Applications and Impact

Diffusion models have revolutionised digital content creation across multiple domains. Notable applications include:

  • Text-to-image generation (Stable Diffusion, Google Imagen)
  • Text-to-video synthesis (OpenAI SORA)
  • Medical imaging and diagnostic applications
  • Autonomous vehicle development
  • Audio and sound generation
  • Personalised AI assistants

Mathematical Foundation

Diffusion models are formally classified as latent variable generative models that map to latent space using a fixed Markov chain. The forward process gradually adds noise to obtain the approximate posterior:

q(x_{1:T}|x_0)

where x_1, \ldots, x_T are latent variables with the same dimensionality as the original data x_0. The reverse process learns to invert this transformation, generating new samples from pure noise through iterative denoising steps.

Theoretical Lineage: Yoshua Bengio and Deep Learning Foundations

Whilst diffusion models represent a relatively recent innovation, their theoretical foundations are deeply rooted in the work of Yoshua Bengio, a pioneering figure in deep learning and artificial intelligence. Bengio’s contributions to understanding neural networks, representation learning, and generative models have profoundly influenced the development of modern AI systems, including diffusion models.

Bengio, born in 1964 in Paris and now based in Canada, is widely recognised as one of the three “godfathers of AI” alongside Yann LeCun and Geoffrey Hinton. His career has been marked by fundamental contributions to machine learning theory and practice. In the 1990s and 2000s, Bengio conducted groundbreaking research on neural networks, including work on the vanishing gradient problem and the development of techniques for training deep architectures. His research on representation learning established that neural networks learn hierarchical representations of data, a principle central to understanding how diffusion models capture complex patterns.

Bengio’s work on energy-based models and probabilistic approaches to learning directly informed the theoretical framework underlying diffusion models. His emphasis on understanding the statistical principles governing generative processes provided crucial insights into how models can learn to reverse noising processes. Furthermore, Bengio’s advocacy for interpretability and theoretical understanding in deep learning has influenced the rigorous mathematical treatment of diffusion models, distinguishing them from more empirically-driven approaches.

In recent years, Bengio has become increasingly focused on AI safety and the societal implications of advanced AI systems. His recognition of diffusion models’ potential-both for beneficial applications and potential risks-reflects his broader commitment to ensuring that powerful generative technologies are developed responsibly. Bengio’s continued influence on the field ensures that diffusion models are developed with attention to both theoretical rigour and ethical considerations.

The connection between Bengio’s foundational work on deep learning and the emergence of diffusion models exemplifies how theoretical advances in understanding neural networks eventually enable practical breakthroughs in generative modelling. Diffusion models represent a maturation of principles Bengio helped establish: the power of hierarchical representations, the importance of probabilistic frameworks, and the value of learning from data through carefully designed loss functions.

References

1. https://www.superannotate.com/blog/diffusion-models

2. https://www.geeksforgeeks.org/artificial-intelligence/what-are-diffusion-models/

3. https://en.wikipedia.org/wiki/Diffusion_model

4. https://www.coursera.org/articles/diffusion-models

5. https://www.assemblyai.com/blog/diffusion-models-for-machine-learning-introduction

6. https://www.splunk.com/en_us/blog/learn/diffusion-models.html

7. https://lilianweng.github.io/posts/2021-07-11-diffusion-models/

"Diffusion models are a class of generative artificial intelligence (AI) models that create new data instances by learning to reverse a gradual, step-by-step process of adding noise to training data." - Term: Diffusion models

read more
Term: Model weights

Term: Model weights

“Model weights are the crucial numerical parameters learned during training that define a model’s internal knowledge, dictating how input data is transformed into outputs and enabling it to recognise patterns and make predictions.” – Model weights

Model weights represent the learnable numerical parameters within a neural network that determine how input data is processed to generate predictions, functioning similarly to synaptic strengths in a biological brain.1,2,4 These values control the influence of specific features on the output, such as edges in images or tokens in language models, through operations like matrix multiplications, convolutions, or weighted sums across layers.1,2,3 Initially randomised, weights are optimised during training via algorithms like gradient descent, which iteratively adjust them to minimise a loss function measuring the difference between predictions and actual targets.1,2,5

In practice, for a simple linear regression model expressed as y = wx + b, the weight w scales the input x to predict y, while b is the bias term.2 In complex architectures like convolutional neural networks (CNNs) or large language models (LLMs), weights include filters detecting textures and fully connected layers combining features, often numbering in billions.1,2,5 This enables tasks from image classification to real-time translation, with pre-trained weights facilitating transfer learning on custom datasets.1

Weights are distinct from biases, which add normalisation and extra characteristics to the weighted sum before activation functions, aiding forward and backward propagation.3,6 Protecting these parameters is vital, as they encode the model’s performance, robustness, and decision logic; unauthorised changes can lead to malfunction.5 In LLMs, weights boost emphasis on words or associations, shaping generative outputs.3

Key Theorist: Geoffrey Hinton

The preeminent theorist linked to model weights is **Geoffrey Hinton**, often called the ‘Godfather of Deep Learning’ for pioneering backpropagation and neural network training techniques that optimise these parameters.1,2 Hinton’s seminal 1986 paper with David Rumelhart and Ronald Williams popularised backpropagation, the cornerstone algorithm for adjusting weights layer-by-layer based on error gradients, revolutionising machine learning.2,4

Born in 1947 in Wimbledon, London, Hinton descends from a lineage of scientists: his great-great-grandfather George Boole invented Boolean logic, his grandfather Charles Howard Hinton coined ‘hyperspace’, and his great-uncle was logician Bertrand Russell. Initially studying experimental psychology at Cambridge (BA 1970), Hinton earned a PhD in AI from Edinburgh in 1978, focusing on Boltzmann machines-early stochastic neural networks with learnable weights. Disillusioned with symbolic AI, he championed connectionism, simulating brain-like learning via weights.

In the 1980s, amid the first AI winter, Hinton persisted at Carnegie Mellon and Toronto, developing restricted Boltzmann machines for unsupervised pre-training of weights, addressing vanishing gradients. His 2006 breakthrough with Alex Krizhevsky and Ilya Sutskever-training deep belief networks on ImageNet-proved deep nets with billions of weights could excel, sparking the deep learning revolution.1 At Google Brain (2013-2023), he advanced capsule networks and transformers indirectly influencing LLMs. Hinton quit Google in 2023, warning of AI risks, and won the 2018 Turing Award with Yann LeCun and Yoshua Bengio. His work directly underpins how modern models, including LLMs, learn weights to recognise patterns and predict outcomes.3,5

References

1. https://www.ultralytics.com/glossary/model-weights

2. https://www.tencentcloud.com/techpedia/132448

3. https://blog.metaphysic.ai/weights-in-machine-learning/

4. https://tedai-sanfrancisco.ted.com/glossary/weights/

5. https://alliancefortrustinai.org/how-model-weights-can-be-used-to-fine-tune-ai-models/

6. https://h2o.ai/wiki/weights-and-biases/

"Model weights are the crucial numerical parameters learned during training that define a model's internal knowledge, dictating how input data is transformed into outputs and enabling it to recognise patterns and make predictions." - Term: Model weights

read more
Term: Loss function

Term: Loss function

“A loss function, also known as a cost function, is a mathematical function that quantifies the difference between a model’s predicted output and the actual ‘ground truth’ value for a given input.” – Loss function

A loss function is a mathematical function that quantifies the discrepancy between a model’s predicted output and the actual ground truth value for a given input. Also referred to as an error function or cost function, it serves as the objective function that machine learning and artificial intelligence algorithms seek to optimize during training efforts.

Core Purpose and Function

The loss function operates as a feedback mechanism within machine learning systems. When a model makes a prediction, the loss function calculates a numerical value representing the prediction error-the gap between what the model predicted and what actually occurred. This error quantification is fundamental to the learning process. During training, algorithms such as backpropagation use the gradient of the loss function with respect to the model’s parameters to iteratively adjust weights and biases, progressively reducing the loss and improving predictive accuracy.

The relationship between loss function and cost function warrants clarification: whilst these terms are often used interchangeably, a loss function technically applies to a single training example, whereas a cost function typically represents the average loss across an entire dataset or batch. Both, however, serve the same essential purpose of guiding model optimization.

Key Roles in Machine Learning

Loss functions fulfil several critical functions within machine learning systems:

  • Performance measurement: Loss functions provide a quantitative metric to evaluate how well a model’s predictions align with actual results, enabling objective assessment of model effectiveness.
  • Optimization guidance: By calculating prediction error, loss functions direct the learning algorithm to adjust parameters iteratively, creating a clear path toward improved predictions.
  • Bias-variance balance: Effective loss functions help balance model bias (oversimplification) and variance (overfitting), essential for generalisation to new, unseen data.
  • Training signal: The gradient of the loss function provides the signal by which learning algorithms update model weights during backpropagation.

Common Loss Function Types

Different machine learning tasks require different loss functions. For regression problems involving continuous numerical predictions, Mean Squared Error (MSE) and Mean Absolute Error (MAE) are widely employed. The MAE formula is:

\text{MAE} = \frac{1}{n} \sum_{i=1}^{n} \left| y_i - \hat{y}_i \right|

For classification tasks dealing with categorical data, Binary Cross-Entropy (also called Log Loss) is commonly used for binary classification problems. The formula is:

L(y, f(x)) = -[y \cdot \log(f(x)) + (1 - y) \cdot \log(1 - f(x))]

where y represents the true binary label (0 or 1) and f(x) is the predicted probability of the positive class.

For multi-class classification, Categorical Cross-Entropy extends this concept. Additionally, Hinge Loss is particularly useful in binary classification where clear separation between classes is desired:

L(y, f(x)) = \max(0, 1 - y \cdot f(x))

The Huber Loss function provides robustness to outliers by combining quadratic and linear components, switching between them based on a threshold parameter delta (?).

Related Strategy Theorist: Vladimir Vapnik

Vladimir Naumovich Vapnik (born 1935) stands as a foundational figure in the theoretical underpinnings of loss functions and machine learning optimisation. A Soviet and later American computer scientist, Vapnik’s work on Statistical Learning Theory and Support Vector Machines (SVMs) fundamentally shaped how the machine learning community understands loss functions and their role in model generalisation.

Vapnik’s most significant contribution to loss function theory came through his development of Support Vector Machines in the 1990s, where he introduced the concept of the hinge loss function-a loss function specifically designed to maximise the margin between classification boundaries. This represented a paradigm shift in thinking about loss functions: rather than simply minimising prediction error, Vapnik’s approach emphasised confidence and margin, ensuring models were not merely correct but confidently correct by a specified distance.

Born in the Soviet Union, Vapnik studied mathematics at the University of Uzbekistan before joining the Institute of Control Sciences in Moscow, where he conducted groundbreaking research on learning theory. His theoretical framework, Vapnik-Chervonenkis (VC) theory, provided mathematical foundations for understanding how models generalise from training data to unseen examples-a concept intimately connected to loss function design and selection.

Vapnik’s insight that different loss functions encode different assumptions about what constitutes “good” model behaviour proved revolutionary. His work demonstrated that the choice of loss function directly influences not just training efficiency but the model’s ability to generalise. This principle remains central to modern machine learning: data scientists select loss functions strategically to encode domain knowledge and desired model properties, whether robustness to outliers, confidence in predictions, or balanced handling of imbalanced datasets.

Vapnik’s career spanned decades of innovation, including his later work on transductive learning and learning using privileged information. His theoretical contributions earned him numerous accolades and established him as one of the most influential figures in machine learning science. His emphasis on understanding the mathematical foundations of learning-particularly through the lens of loss functions and generalisation bounds-continues to guide contemporary research in deep learning and artificial intelligence.

Practical Significance

The selection of an appropriate loss function significantly impacts model performance and training efficiency. Data scientists carefully consider different loss functions to achieve specific objectives: reducing sensitivity to outliers, better handling noisy data, minimising overfitting, or improving performance on imbalanced datasets. The loss function thus represents not merely a technical component but a strategic choice that encodes domain expertise and learning objectives into the machine learning system itself.

References

1. https://www.datacamp.com/tutorial/loss-function-in-machine-learning

2. https://h2o.ai/wiki/loss-function/

3. https://c3.ai/introduction-what-is-machine-learning/loss-functions/

4. https://www.geeksforgeeks.org/machine-learning/ml-common-loss-functions/

5. https://arxiv.org/html/2504.04242v1

6. https://www.youtube.com/watch?v=v_ueBW_5dLg

7. https://www.ibm.com/think/topics/loss-function

8. https://en.wikipedia.org/wiki/Loss_function

9. https://www.datarobot.com/blog/introduction-to-loss-functions/

"A loss function, also known as a cost function, is a mathematical function that quantifies the difference between a model's predicted output and the actual 'ground truth' value for a given input." - Term: Loss function

read more
Term: Synthetic data

Term: Synthetic data

“Synthetic data is artificially generated information that computationally or algorithmically mimics the statistical properties, patterns, and structure of real-world data without containing any actual observations or sensitive personal details.” – Synthetic data

What is Synthetic Data?

Synthetic data is artificially generated information that computationally or algorithmically mimics the statistical properties, patterns, and structure of real-world data without containing any actual observations or sensitive personal details. It is created using advanced generative AI models or statistical methods trained on real datasets, producing new records that are statistically identical to the originals but free from personally identifiable information (PII).

This approach enables privacy-preserving data use for analytics, AI training, software testing, and research, addressing challenges like data scarcity, high costs, and compliance with regulations such as GDPR.

Key Characteristics and Generation Methods

  • Privacy Protection: No one-to-one relationships exist between synthetic records and real individuals, eliminating re-identification risks.1,3
  • Utility Preservation: Retains correlations, distributions, and insights from source data, serving as a perfect proxy for real datasets.1,2
  • Flexibility: Easily modifiable for bias correction, scaling, or scenario testing without compliance issues.1

Synthetic data is generated through methods including:

  • Statistical Distribution: Analysing real data to identify distributions (e.g., normal or exponential) and sampling new data from them.4
  • Model-Based: Training machine learning models, such as generative adversarial networks (GANs), to replicate data characteristics.1,4
  • Simulation: Using computer models for domains like physical simulations or AI environments.7

Types of Synthetic Data

Type Description
Fully Synthetic Entirely new data with no real-world elements, matching statistical properties.4,5
Partially Synthetic Sensitive parts of real data replaced, rest unchanged.5
Hybrid Real data augmented with synthetic records.5

Applications and Benefits

  • AI and Machine Learning: Trains models efficiently when real data is scarce or sensitive, accelerating development in fields like autonomous systems and medical imaging.2,7
  • Software Testing: Simulates user behaviour and edge cases without real data risks.2
  • Data Sharing: Enables collaboration while complying with privacy laws; Gartner predicts most AI data will be synthetic by 2030.1

Best Related Strategy Theorist: Kalyan Veeramachaneni

Kalyan Veeramachaneni, a principal research scientist at MIT’s Schwarzman College of Computing, is a leading figure in synthetic data strategies, particularly for scalable, privacy-focused data generation in AI.

Born in India, Veeramachaneni earned his PhD in computer science from the University of Mainz, Germany, focusing on machine learning and data privacy. He joined MIT in 2011 after postdoctoral work at the University of Illinois. His research bridges AI, data science, and privacy engineering, pioneering automated machine learning (AutoML) and synthetic data techniques.

Veeramachaneni’s relationship to synthetic data stems from his development of generative models that create datasets with identical mathematical properties to real ones, adding ‘noise’ to mask originals. This innovation, detailed in MIT Sloan publications, supports competitive advantages through secure data sharing and algorithm development. His work has influenced enterprise AI strategies, emphasising synthetic data’s role in overcoming real-data limitations while preserving utility.

References

1. https://mostly.ai/synthetic-data-basics

2. https://accelario.com/glossary/synthetic-data/

3. https://mitsloan.mit.edu/ideas-made-to-matter/what-synthetic-data-and-how-can-it-help-you-competitively

4. https://aws.amazon.com/what-is/synthetic-data/

5. https://www.salesforce.com/data/synthetic-data/

6. https://tdwi.org/pages/glossary/synthetic-data.aspx

7. https://en.wikipedia.org/wiki/Synthetic_data

8. https://www.ibm.com/think/topics/synthetic-data

9. https://www.urban.org/sites/default/files/2023-01/Understanding%20Synthetic%20Data.pdf

"Synthetic data is artificially generated information that computationally or algorithmically mimics the statistical properties, patterns, and structure of real-world data without containing any actual observations or sensitive personal details." - Term: Synthetic data

read more
Term: Scaling hypothesis

Term: Scaling hypothesis

“The scaling hypothesis in artificial intelligence is the theory that the cognitive ability and performance of general learning algorithms will reliably improve, or even unlock new, more complex capabilities, as computational resources, model size, and the amount of training data are increased.” – Scaling hypothesis

The **scaling hypothesis** in artificial intelligence posits that the cognitive ability and performance of general learning algorithms, particularly deep neural networks, will reliably improve-or even unlock entirely new, more complex capabilities-as computational resources, model size (number of parameters), and training data volume are increased.1,5

This principle suggests predictable, power-law improvements in model performance, often manifesting as emergent behaviours such as enhanced reasoning, general problem-solving, and meta-learning without architectural changes.2,3,5 For instance, larger models like GPT-3 demonstrated abilities in arithmetic and novel tasks not explicitly trained, supporting the idea that intelligence arises from simple units applied at vast scale.2,4

Key Components

  • Model Size: Increasing parameters and layers in neural networks, such as transformers.3
  • Training Data: Exposing models to exponentially larger, diverse datasets to capture complex patterns.1,4
  • Compute: Greater computational power and longer training durations, akin to extended study time.3,4

Empirical evidence from models like GPT-3, BERT, and Vision Transformers shows consistent gains across language, vision, and reinforcement learning tasks, challenging the need for specialised architectures.1,4,5

Historical Context and Evidence

Rooted in early connectionism, the hypothesis gained prominence in the late 2010s with large-scale models like GPT-3 (2020), where scaling alone outperformed complex alternatives.1,5 Proponents argue it charts a path to artificial general intelligence (AGI), potentially requiring millions of times current compute for human-level performance.2

Best Related Strategy Theorist: Gwern Branwen

Gwern Branwen stands as the foremost theorist formalising the **scaling hypothesis**, authoring the seminal 2020 essay The Scaling Hypothesis that synthesised empirical trends into a radical paradigm for AGI.5 His work posits that neural networks, when scaled massively, generalise better, become more Bayesian, and exhibit emergent sophistication as the optimal solution to diverse tasks-echoing brain-like universal learning.5

Biography: Gwern Branwen (born c. 1984) is an independent researcher, writer, and programmer based in the USA, known for his prolific contributions to AI, psychology, statistics, and effective altruism under the pseudonym ‘Gwern’. A self-taught polymath, he dropped out of university to pursue independent scholarship, funding his work through Patreon and commissions. Branwen maintains gwern.net, a vast archive of over 1,000 essays blending rigorous analysis with original experiments, such as modafinil self-trials and AI scaling forecasts.

His relationship to the scaling hypothesis stems from deep dives into deep learning papers, predicting in 2019-2020 that ‘blessings of scale’-predictable performance gains-would dominate AI progress. Influencing OpenAI’s strategy, Branwen’s calculations extrapolated GPT-3 results, estimating 2.2 million times more compute for human parity, reinforcing bets on transformers and massive scaling.2,5 A critic of architectural over-engineering, he advocates simple algorithms at unreachable scales as the AGI secret, impacting labs like OpenAI and Anthropic.

Implications and Critiques

While driving breakthroughs, concerns include resource concentration enabling unchecked AGI development, diminishing interpretability, and potential misalignment without safety innovations.4 Interpretations range from weak (error reduction as power law) to strong (novel abilities emerge).6

References

1. https://www.envisioning.com/vocab/scaling-hypothesis

2. https://johanneshage.substack.com/p/scaling-hypothesis-the-path-to-artificial

3. https://drnealaggarwal.info/what-is-scaling-in-relation-to-ai/

4. https://www.species.gg/blog/the-scaling-hypothesis-made-simple

5. https://gwern.net/scaling-hypothesis

6. https://philsci-archive.pitt.edu/23622/1/psa_scaling_hypothesis_manuscript.pdf

7. https://lastweekin.ai/p/the-ai-scaling-hypothesis

"The scaling hypothesis in artificial intelligence is the theory that the cognitive ability and performance of general learning algorithms will reliably improve, or even unlock new, more complex capabilities, as computational resources, model size, and the amount of training data are increased." - Term: Scaling hypothesis

read more
Quote: Jack Clark – Import AI

Quote: Jack Clark – Import AI

“Since 2020, we have seen a 600 000x increase in the computational scale of decentralized training projects, for an implied growth rate of about 20x/year.” – Jack Clark – Import AI

Jack Clark on Exponential Growth in Decentralized AI Training

The Quote and Its Context

Jack Clark’s statement about the 600,000x increase in computational scale for decentralized training projects over approximately five years (2020-2025) represents a striking observation about the democratization of frontier AI development.1,2,3,4 This 20x annual growth rate reflects one of the most significant shifts in the technological and political economy of artificial intelligence: the transition from centralized, proprietary training architectures controlled by a handful of well-capitalized labs toward distributed, federated approaches that enable loosely coordinated collectives to pool computational resources globally.

Jack Clark: Architect of AI Governance Thinking

Jack Clark is the Head of Policy at Anthropic and one of the most influential voices shaping how we think about AI development, governance, and the distribution of technological power.1 His trajectory uniquely positions him to observe this transformation. Clark co-authored the original GPT-2 paper at OpenAI in 2019, a moment he now reflects on as pivotal—not merely for the model’s capabilities, but for what it revealed about scaling laws: the discovery that larger models trained on more data would exhibit predictably superior performance across diverse tasks, even without task-specific optimization.1

This insight proved prophetic. Clark recognized that GPT-2 was “a sketch of the future”—a partial glimpse of what would emerge through scaling. The paper’s modest performance advances on seven of eight tested benchmarks, achieved without narrow task optimization, suggested something fundamental about how neural networks could be made more generally capable.1 What followed validated his foresight: GPT-3, instruction-tuned variants, ChatGPT, Claude, and the subsequent explosion of large language models all emerged from the scaling principles Clark and colleagues had identified.

However, Clark’s thinking has evolved substantially since those early days. Reflecting in 2024, five years after GPT-2’s release, he acknowledged that while his team had anticipated many malicious uses of advanced language models, they failed to predict the most disruptive actual impact: the generation of low-grade synthetic content driven by economic incentives rather than malicious intent.1 This humility about the limits of foresight informs his current policy positions.

The Political Economy of Decentralized Training

Clark’s observation about the 600,000x scaling in decentralized training projects is not merely a technical metric—it is a statement about power distribution. Currently, the frontier of AI capability depends on the ability to concentrate vast amounts of computational resources in physically centralized clusters. Companies like Anthropic, OpenAI, and hyperscalers like Google and Meta control this concentrated compute, which has enabled governments and policymakers to theoretically monitor and regulate AI development through chokepoints: controlling access to advanced semiconductors, tracking large training clusters, and licensing centralized development entities.3,4

Decentralized training disrupts this assumption entirely. If computational resources can be pooled across hundreds of loosely federated organizations and individuals globally—each contributing smaller clusters of GPUs or other accelerators—then the frontier of AI capability becomes distributed across many actors rather than concentrated in a few.3,4 This changes everything about AI policy, which has largely been built on the premise of controllable centralization.

Recent proof-of-concepts underscore this trajectory:

  • Prime Intellect’s INTELLECT-1 (10 billion parameters) demonstrated that decentralized training at scale was technically feasible, a threshold achievement because it showed loosely coordinated collectives could match capabilities that previously required single-company efforts.3,9

  • INTELLECT-2 (32 billion parameters) followed, designed to compete with modern reasoning models through distributed training, suggesting that decentralized approaches were not merely proof-of-concept but could produce competitive frontier-grade systems.4

  • DiLoCoX, an advancement on DeepMind’s DiLoCo technology, demonstrated a 357x speedup in distributed training while achieving model convergence across decentralized clusters with minimal network bandwidth (1Gbps)—a crucial breakthrough because communication overhead had previously been the limiting factor in distributed training.2

The implied growth rate of 20x annually suggests an acceleration curve where technical barriers to decentralized training are falling faster than regulatory frameworks or policy interventions can adapt.

Leading Theorists and Intellectual Lineages

Scaling Laws and the Foundations

The intellectual foundation for understanding exponential growth in AI capabilities rests on the work of researchers who formalized scaling laws. While Clark and colleagues at OpenAI contributed to this work through GPT-2 and subsequent research, the broader field—including contributions from Jared Kaplan, Dario Amodei, and others at Anthropic—established that model performance scales predictably with increases in parameters, data, and compute.1 These scaling laws create the mathematical logic that enables decentralized systems to be competitive: a 32-billion-parameter model trained via distributed methods can approach the capabilities of centralized training at similar scales.

Political Economy and Technological Governance

Clark’s thinking is situated within broader intellectual traditions examining how technology distributes power. His emphasis on the “political economy” of AI reflects influence from scholars and policymakers concerned with how technological architectures embed power relationships. The notion that decentralized training redistributes who can develop frontier AI systems draws on longstanding traditions in technology policy examining how architectural choices (centralized vs. distributed systems) have political consequences.

His advocacy for polycentric governance—distributing decision-making about AI behavior across multiple scales from individuals to platforms to regulatory bodies—reflects engagement with governance theory emphasizing that monocentric control is often less resilient and responsive than systems with distributed decision-making authority.5

The “Regulatory Markets” Framework

Clark has articulated the need for governments to systematically monitor the societal impact and diffusion of AI technologies, a position he advanced through the concept of “Regulatory Markets”—market-driven mechanisms for monitoring AI systems. This framework acknowledges that traditional command-and-control regulation may be poorly suited to rapidly evolving technological domains and that measurement and transparency might be more foundational than licensing or restriction.1 This connects to broader work in regulatory innovation and adaptive governance.

The Implications of Exponential Decentralization

The 600,000x growth over five years, if sustained or accelerated, implies several transformative consequences:

On AI Policy: Traditional approaches to AI governance that assume centralized training clusters and a small number of frontier labs become obsolete. Export controls on advanced semiconductors, for instance, become less effective if 100 organizations in 50 countries can collectively train competitive models using previous-generation chips.3,4

On Open-Source Development: The growth depends crucially on the availability of open-weight models (like Meta’s LLaMA or DeepSeek) and accessible software stacks (like Prime.cpp) that enable distributed inference and fine-tuning.4 The democratization of capability is inseparable from the proliferation of open-source infrastructure.

On Sovereignty and Concentration: Clark frames this as essential for “sovereign AI”—the ability for nations, organizations, and individuals to develop and deploy capable AI systems without dependence on centralized providers. However, this same decentralization could enable the rapid proliferation of systems with limited safety testing or alignment work.4

On Clark’s Own Policy Evolution: Notably, Clark has found himself increasingly at odds with AI safety and policy positions he previously held or was associated with. He expresses skepticism toward licensing regimes for AI development, restrictions on open-source model deployment, and calls for worldwide development pauses—positions that, he argues, would create concentrated power in the present to prevent speculative future risks.1 Instead, he remains confident in the value of systematic societal impact monitoring and measurement, which he has championed through his work at Anthropic and in policy forums like the Bletchley and Seoul AI safety summits.1

The Unresolved Tension

The exponential growth in decentralized training capacity creates a central tension in AI governance: it democratizes access to frontier capabilities but potentially distributes both beneficial and harmful applications more widely. Clark’s quote and his broader work reflect an intellectual reckoning with this tension—recognizing that attempts to maintain centralized control through policy and export restrictions may be both technically infeasible and politically counterproductive, yet that some form of measurement and transparency remains essential for democratic societies to understand and respond to AI’s societal impacts.

References

1. https://jack-clark.net/2024/06/03/import-ai-375-gpt-2-five-years-later-decentralized-training-new-ways-of-thinking-about-consciousness-and-ai/

2. https://jack-clark.net/2025/06/30/import-ai-418-100b-distributed-training-run-decentralized-robots-ai-myths/

3. https://jack-clark.net/2024/10/14/import-ai-387-overfitting-vs-reasoning-distributed-training-runs-and-facebooks-new-video-models/

4. https://jack-clark.net/2025/04/21/import-ai-409-huawei-trains-a-model-on-8000-ascend-chips-32b-decentralized-training-run-and-the-era-of-experience-and-superintelligence/

5. https://importai.substack.com/p/import-ai-413-40b-distributed-training

6. https://www.youtube.com/watch?v=uRXrP_nfTSI

7. https://importai.substack.com/p/import-ai-375-gpt-2-five-years-later/comments

8. https://jack-clark.net

9. https://jack-clark.net/2024/12/03/import-ai-393-10b-distributed-training-run-china-vs-the-chip-embargo-and-moral-hazards-of-ai-development/

10. https://www.lesswrong.com/posts/iFrefmWAct3wYG7vQ/ai-labs-statements-on-governance

"Since 2020, we have seen a 600 000x increase in the computational scale of decentralized training projects, for an implied growth rate of about 20x/year." - Quote: Jack Clark

read more
Quote: Yann Lecun

Quote: Yann Lecun

“Most of the infrastructure cost for AI is for inference: serving AI assistants to billions of people.”
— Yann LeCun, VP & Chief AI Scientist at Meta

Yann LeCun made this comment in response to the sharp drop in Nvidia’s share price on January 27, 2024, following the launch of Deepseek R1, a new AI model developed by Deepseek AI. This model was reportedly trained at a fraction of the cost incurred by Hyperscalers like OpenAI, Anthropic, and Google DeepMind, raising questions about whether Nvidia’s dominance in AI compute was at risk.

The market reaction stemmed from speculation that the training costs of cutting-edge AI models—previously seen as a key driver of Nvidia’s GPU demand—could decrease significantly with more efficient methods. However, LeCun pointed out that most AI infrastructure costs come not from training but from inference, the process of running AI models at scale to serve billions of users. This suggests that Nvidia’s long-term demand may remain strong, as inference still relies heavily on high-performance GPUs.

LeCun’s view aligned with analyses from key AI investors and industry leaders. He supported the argument made by Antoine Blondeau, co-founder of Alpha Intelligence Capital, who described Nvidia’s stock drop as “vastly overblown” and “NOT a ‘Sputnik moment’”, referencing the concern that Nvidia’s market position was insecure. Additionally, Jonathan Ross, founder of Groq, shared a video titled “Why $500B isn’t enough for AI,” explaining why AI compute demand remains insatiable despite efficiency gains.

This discussion underscores a critical aspect of AI economics: while training costs may drop with better algorithms and hardware, the sheer scale of inference workloads—powering AI assistants, chatbots, and generative models for billions of users—remains a dominant and growing expense. This supports the case for sustained investment in AI infrastructure, particularly in Nvidia’s GPUs, which continue to be the gold standard for inference at scale.

read more
Infographic: Four critical DeepSeek enablers

Infographic: Four critical DeepSeek enablers

The DeepSeek team has introduced several high-impact changes to Large Language Model (LLM) architecture to enhance performance and efficiency:

  1. Multi-Head Latent Attention (MLA): This mechanism enables the model to process multiple facets of input data simultaneously, improving both efficiency and performance. MLA reduces the memory required to compute a transformer’s attention by a factor of 7.5x to 20x, a breakthrough that makes large-scale AI applications more feasible. Unlike Flash Attention, which improves data organization in memory, MLA compresses the KV cache into a lower-dimensional space, significantly reducing memory usage—down to 5% to 13% of traditional attention mechanisms—while maintaining performance.
  2. Mixture-of-Experts (MoE) Architecture: DeepSeek employs an MoE system that activates only a subset of its total parameters during any given task. For instance, in DeepSeek-V3, only 37 billion out of 671 billion parameters are active at a time, significantly reducing computational costs. This approach enhances efficiency and aligns with the trend of making AI models more compute-light, allowing freed-up GPU resources to be allocated to multi-modal processing, spatial intelligence, or genomic analysis. MoE models, as also leveraged by Mistral and other leading AI labs, allow for scalability while keeping inference costs manageable.
  3. FP8 Floating Point Precision: To enhance computational efficiency, DeepSeek-V3 utilizes FP8 floating point precision during training, which helps in reducing memory usage and accelerating computation. This follows a broader trend in AI to optimize training methodologies, potentially influencing the approach taken by U.S.-based LLM providers. Given China’s restricted access to high-end GPUs due to U.S. export controls, optimizations like FP8 and MLA are critical in overcoming hardware limitations.
  4. DeepSeek-R1 and Test-Time Compute Capabilities: DeepSeek-R1 is a model that leverages reinforcement learning (RL) to enable test-time compute, significantly improving reasoning capabilities. The model was trained using an innovative RL strategy, incorporating fine-tuned Chain of Thought (CoT) data and supervised fine-tuning (SFT) data across multiple domains. Notably, DeepSeek demonstrated that any sufficiently powerful LLM can be transformed into a high-performance reasoning model using only 800k curated training samples. This technique allows for rapid adaptation of smaller models, such as Qwen and LLaMa-70b, into competitive reasoners.
  5. Distillation to Smaller Models: The team has developed distilled versions of their models, such as DeepSeek-R1-Distill, which are fine-tuned on synthetic data generated by larger models. These distilled models contain fewer parameters, making them more efficient while retaining significant capabilities. DeepSeek’s ability to achieve comparable reasoning performance at a fraction of the cost of OpenAI’s models (5% of the cost, according to Pelliccione) has disrupted the AI landscape.

The Impact of Open-Source Models:

DeepSeek’s success highlights a fundamental shift in AI development. Traditionally, leading-edge models have been closed-source and controlled by Western AI firms like OpenAI, Google, and Anthropic. However, DeepSeek’s approach, leveraging open-source components while innovating on training efficiency, has disrupted this dynamic. Pelliccione notes that DeepSeek now offers similar performance to OpenAI at just 5% of the cost, making high-quality AI more accessible. This shift pressures proprietary AI companies to rethink their business models and embrace greater openness.

Challenges and Innovations in the Chinese AI Ecosystem:

China’s AI sector faces major constraints, particularly in access to high-performance GPUs due to U.S. export restrictions. Yet, Chinese companies like DeepSeek have turned these challenges into strengths through aggressive efficiency improvements. MLA and FP8 precision optimizations exemplify how innovation can offset hardware limitations. Furthermore, Chinese AI firms, historically focused on scaling existing tech, are now contributing to fundamental advancements in AI research, signaling a shift towards deeper innovation.

The Future of AI Control and Adaptation:

DeepSeek-R1’s approach to training AI reasoners poses a challenge to traditional AI control mechanisms. Since reasoning capabilities can now be transferred to any capable model with fewer than a million curated samples, AI governance must extend beyond compute resources and focus on securing datasets, training methodologies, and deployment platforms. OpenAI has previously obscured Chain of Thought traces to prevent leakage, but DeepSeek’s open-weight release and published RL techniques have made such restrictions ineffective.

Broader Industry Context:

  • DeepSeek benefits from Western open-source AI developments, particularly Meta’s LLama model disclosures, which provided a foundation for its advancements. However, DeepSeek’s success also demonstrates that China is shifting from scaling existing technology to innovating at the frontier.
  • Open-source models like DeepSeek will see widespread adoption for enterprise and research applications, though Western businesses are unlikely to build their consumer apps on a Chinese API.
  • The AI innovation cycle is exceptionally fast, with breakthroughs assessed daily or weekly. DeepSeek’s advances are part of a rapidly evolving competitive landscape dominated by U.S. big tech players like OpenAI, Google, Microsoft, and Meta, who continue to push for productization and revenue generation. Meanwhile, Chinese AI firms, despite hardware and data limitations, are innovating at an accelerated pace and have proven capable of challenging OpenAI’s dominance.

These innovations collectively contribute to more efficient and effective LLMs, balancing performance with resource utilization while shaping the future of AI model development.

Sources: Global Advisors, Jack Clark – Anthropic, Antoine Blondeau, Alberto Pelliccione, infoq.com, medium.com, en.wikipedia.org, arxiv.org

read more
Quote: Jack Clark

Quote: Jack Clark

“The most surprising part of DeepSeek-R1 is that it only takes ~800k samples of ‘good’ RL reasoning to convert other models into RL-reasoners. Now that DeepSeek-R1 is available people will be able to refine samples out of it to convert any other model into an RL reasoner.” – Jack Clark, Anthropic

Jack Clark, Co-founder of Anthropic, co-chair of the AI Index at Stanford University, co-chair of OECD working group on AI & Compute, shed light on the significance of DeepSeek-R1, a revolutionary AI reasoning model developed by China’s DeepSeek team. In an article posted in his newsletter on the 27th January 2025, Clark highlighted that it only takes approximately 800k samples of “good” RL (Reinforcement Learning) reasoning to convert other models into RL-reasoners.

The Power of Fine-Tuning

DeepSeek-R1 is not just a powerful AI model; it also provides a framework for fine-tuning existing models to enhance their reasoning capabilities. By leveraging the 800k samples curated with DeepSeek-R1, researchers can refine any other model into an RL reasoner. This approach has been demonstrated by fine-tuning open-source models like Qwen and Llama using the same dataset.

Implications for AI Policy

The release of DeepSeek-R1 has significant implications for AI policy and control. As Clark notes, if you need fewer than a million samples to convert any model into a “thinker,” it becomes much harder to control AI systems. This is because the valuable data, including chains of thought from reasoning models, can be leaked or shared openly.

A New Era in AI Development

The availability of DeepSeek-R1 and its associated techniques has created a new era in AI development. With an open weight model floating around the internet, researchers can now bootstrap any other sufficiently powerful base model into being an AI reasoner. This has the potential to accelerate AI progress worldwide.

Key Takeaways:

  • Fine-tuning is key : DeepSeek-R1 demonstrates that fine-tuning existing models with a small amount of data (800k samples) can significantly enhance their reasoning capabilities.
  • Open-source and accessible : The model and its techniques are now available for anyone to use, making it easier for researchers to develop powerful AI reasoners.
  • Implications for control : The release of DeepSeek-R1 highlights the challenges of controlling AI systems, as valuable data can be leaked or shared openly.

Conclusion

DeepSeek-R1 has marked a significant milestone in AI development, showcasing the power of fine-tuning and open-source collaboration. As researchers continue to build upon this work, we can expect to see even more advanced AI models emerge, with far-reaching implications for various industries and applications.

read more
Quote: Marc Andreessen

Quote: Marc Andreessen

“DeepSeek-R1 is AI’s Sputnik moment.” – Marc Andreessen, Andreesen Horowitz

In a 27th January 2025 X statement that sent shockwaves through the tech community, venture capitalist Marc Andreessen declared that DeepSeek’s R1 AI reasoning model is “AI’s Sputnik moment.” This analogy draws parallels between China’s breakthrough in artificial intelligence and the Soviet Union’s historic achievement of launching the first satellite into orbit in 1957.

The Rise of DeepSeek-R1

DeepSeek, a Chinese AI lab, has made headlines with its open-source release of R1, a revolutionary AI reasoning model that is not only more cost-efficient but also poses a significant threat to the dominance of Western tech giants. The model’s ability to reduce compute requirements by half without sacrificing accuracy has sent shockwaves through the industry.

A New Era in AI

The release of DeepSeek-R1 marks a turning point in the AI arms race, as it challenges the long-held assumption that only a select few companies can compete in this space. By making its research open-source, DeepSeek is empowering anyone to build their own version of R1 and tailor it to their needs.

Implications for Megacap Stocks

The success of DeepSeek-R1 has significant implications for megacap stocks like Microsoft, Alphabet, and Amazon, which have long relied on proprietary AI models to maintain their technological advantage. The pen-source nature of R1 threatens to wipe out this advantage, potentially disrupting the business models of these tech giants.

Nvidia’s Nightmare

The news comes as a blow to Nvidia CEO Jensen Huang, who is ramping up production of his Blackwell microchip, a more advanced version of his industry-leading Hopper series H100s. The chip controls 90% of the AI semiconductor market, but R1’s ability to reduce compute requirements may render these chips less essential.

A New Era of Innovation

Perplexity AI founder Aravind Srinivas praised DeepSeek’s team for catching up to the West by employing clever solutions, including switching from binary encoding to floating point 8. This innovation not only reduces costs but also demonstrates that China is no longer just a copycat, but a leader in AI innovation.

read more
Quote: Jeffrey Emanuel

Quote: Jeffrey Emanuel

“With R1, DeepSeek essentially cracked one of the holy grails of AI: getting models to reason step-by-step without relying on massive supervised datasets.” – Jeffrey Emanuel

Jeffrey Emanuel’s statement (“The Short Case for Nvidia Stock” – 25th January 2025) highlights a groundbreaking achievement in AI with DeepSeek’s R1 model, which has made significant strides in enabling step-by-step reasoning without the traditional reliance on vast supervised datasets:

  1. Innovation Through Reinforcement Learning (RL):
    • The R1 model employs reinforcement learning, a method where models learn through trial and error with feedback. This approach reduces the dependency on large labeled datasets typically required for training, making it more efficient and accessible.
  2. Advanced Reasoning Capabilities:
    • R1 excels in tasks requiring logical inference and mathematical problem-solving. Its ability to demonstrate step-by-step reasoning is crucial for complex decision-making processes, applicable across various industries from autonomous systems to intricate problem-solving tasks.
  3. Efficiency and Accessibility:
    • By utilizing RL and knowledge distillation techniques, R1 efficiently transfers learning to smaller models. This democratizes AI technology, allowing global researchers and developers to innovate without proprietary barriers, thus expanding the reach of advanced AI solutions.
  4. Impact on Data-Scarce Industries:
    • The model’s capability to function with limited data is particularly beneficial in sectors like medicine and finance, where labeled data is scarce due to privacy concerns or high costs. This opens doors for more ethical and feasible AI applications in these fields.
  5. Competitive Landscape and Innovation:
    • R1 positions itself as a competitor to models like OpenAI’s o1, signaling a shift towards accessible AI technology. This fosters competition and encourages other companies to innovate similarly, driving advancements across the AI landscape.

In essence, DeepSeek’s R1 model represents a significant leap in AI efficiency and accessibility, offering profound implications for various industries by reducing data dependency and enhancing reasoning capabilities.

read more

Download brochure

Introduction brochure

What we do, case studies and profiles of some of our amazing team.

Download

Our latest podcasts on Spotify

Sign up for our newsletters - free

Global Advisors | Quantified Strategy Consulting