| |
|
A daily bite-size selection of top business content.
PM edition. Issue number 1226
Latest 10 stories. Click the button for more.
|
| |
"Artificial intelligence (AI) is now an integral part of new chemistry development and is set to supercharge the future of material engineering and reduce the time to discover, test, and deploy new materials and designs." - Council on Foreign Relations - Leapfrogging China's Critical Minerals Dominance
This statement from the influential report Leapfrogging China's Critical Minerals Dominance: How Innovation Can Secure U.S. Supply Chains, published by the Council on Foreign Relations (CFR) and Silverado Policy Accelerator, underscores a pivotal shift in global resource strategy.1,3,4 Released on 5 February 2026, the report argues that the United States cannot compete with China through conventional mining and processing alone, given Beijing's decades-long entrenchment across the critical minerals ecosystem-from extraction to magnet manufacturing.1,2 Instead, it advocates 'leapfrogging' via disruptive technologies, with artificial intelligence (AI) positioned as a transformative force in accelerating materials discovery and engineering.1,4
Context of the Quote and Geopolitical Stakes
Critical minerals-such as rare-earth elements (REEs), lithium, cobalt, and nickel-are indispensable for advanced technologies, including electric vehicles, renewable energy systems, defence equipment, and semiconductors.1,5 China dominates this sector, controlling over 90% of heavy REE processing and nearly all permanent magnet production, creating strategic chokepoints that it has weaponised through export controls since 2023.1 In October 2025, Beijing expanded restrictions on REEs and related technologies, nearly halting global supply chains and exposing U.S. vulnerabilities.1
The report emerges amid escalating U.S.-China tensions under the second Trump administration, where retaliatory tariffs and bans on semiconductor inputs like gallium and germanium have intensified.1 Traditional responses, such as expanding domestic mining, face insurmountable hurdles: multi-year permitting, billions in upfront costs, environmental concerns, and China's unmatched scale.1,2 The quote highlights AI's potential to bypass these by supercharging chemistry and materials engineering, slashing discovery-to-deployment timelines from decades to years.1
Authors and Their Expertise
The quote originates from a report co-authored by two leading experts in geoeconomics and supply chain policy.
- Heidi Crebo-Rediker, Senior Fellow for Geoeconomics at CFR and a member of Silverado's Strategic Council, brings deep experience from her time as U.S. State Department Chief Economist (2014-2017) and roles at Goldman Sachs and the National Economic Council. Her work focuses on financial sanctions, economic statecraft, and resilient supply chains.3,4
- Mahnaz Khan, Vice President of Policy for Critical Supply Chains at Silverado Policy Accelerator, specialises in frontier technologies and mineral security. Silverado, a non-partisan think tank, drives innovation in national security challenges, and Khan's contributions emphasise pragmatic financing and allied cooperation to scale breakthroughs.3,4
Endorsed by CFR's Shannon O'Neil, Senior Vice President of Studies, the report calls for embedding innovation-including AI-driven materials engineering-into U.S. policy, alongside waste recovery, substitute materials, and international frameworks like the Forum on Resource Geostrategic Engagement (FORGE).2,4
Leading Theorists in AI-Driven Materials Science and Critical Minerals
The report's vision aligns with pioneering work at the intersection of AI, chemistry, and materials engineering, where theorists and researchers are revolutionising discovery processes.
- Alán Aspuru-Guzik (University of Toronto) is a trailblazer in AI for molecular discovery. His Molecular Space Exploration Engine (MOSE) and A-Lab-a fully autonomous lab-use reinforcement learning and generative models to design and synthesise novel materials, such as battery electrolytes, in weeks rather than years. Aspuru-Guzik's 'materials genome' approach treats chemical space as a vast data landscape for AI navigation, directly supporting faster REE substitutes and magnet alternatives.1
- Roald Hoffmann (Nobel Laureate in Chemistry, 1981), though not an AI specialist, laid foundational theories in extended Hückel molecular orbital methods, enabling computational simulations that AI now accelerates. His work on chemical bonding informs AI models predicting material properties under extreme conditions, vital for critical minerals applications.
- Andrea Goldsmith (Stanford) and collaborators in AI-optimised catalysis advance sustainable extraction from tailings and waste-key report recommendations. Their models integrate machine learning with quantum chemistry to design enzymes and photocatalysts for REE recovery, reducing environmental impact.1
- Jeremy Keith (EPFL) leads in generative AI for inorganic materials, developing models like M3GNet that predict properties across millions of crystal structures. This underpins high-throughput screening for rare-earth-free magnets, addressing China's heavy REE monopoly.1
These theorists converge on a paradigm where AI acts as an 'oracle' for inverse design: specifying desired properties (e.g., magnet strength without dysprosium) and generating viable compounds. Combined with robotic labs and quantum computing, this could cut development times by 90%, aligning precisely with the report's leapfrogging imperative.1,4
Implications for Materials Engineering
AI's integration promises not just speed but resilience: engineering alloys resilient to supply shocks, recycling magnets from e-waste at scale, and bioleaching minerals from industrial byproducts.1 U.S. investments, like the $1.4 billion in rare-earth magnet recycling (November 2025), exemplify this shift, targeting firms like MP Materials and ReElement Technologies.1 By prioritising innovation over replication, the West can forge secure supply chains, diminishing China's leverage and powering the next industrial era.
References
1. https://www.cfr.org/reports/leapfrogging-chinas-critical-minerals-dominance
2. https://www.cfr.org/articles/u-s-allies-aim-to-break-chinas-critical-minerals-dominance
3. https://www.silverado.org/publications/silverado-and-the-council-on-foreign-relations-release-new-report/
4. https://www.cfr.org/articles/new-cfr-report-outlines-how-the-u-s-can-leapfrog-chinas-critical-minerals-dominance
5. https://www.cfr.org
6. https://www.cfr.org/report/enter-dragon-and-elephant
7. https://podcasts.apple.com/us/podcast/this-is-how-the-us-can-become-a-player-in-rare-earth-metals/id1056200096?i=1000748342100

|
| |
| |
"To 'lean into the moment' means to engage fully with the present experience, situation, or task, rather than avoiding it or being distracted. It implies a willingness to be present, observant and responsive, especially when the situation might be uncomfortable or challenging." - Lean in to the moment
To lean into the moment means to engage fully with the present experience, situation, or task, rather than avoiding it or being distracted. It implies a willingness to be present, observant, and responsive, especially when the situation might be uncomfortable or challenging. This phrase draws from the broader idiom 'lean into', which signifies embracing or committing to something with determination, often in the face of uncertainty or difficulty.
The expression encourages owning the current reality, casting off concerns, and moving forward with confidence. For instance, it can involve pursuing a task with great effort and perseverance, accepting potentially negative traits to turn them positive, or persevering despite risk. In creative or professional contexts, it means embracing uncertainty to foster growth, as seen in teaching scenarios where one confronts fear head-on.
Origins and Evolution of the Phrase
The phrasal verb 'lean into' emerged in the mid-20th century in the US, meaning to embrace or commit fully. Early examples include a 1941 citation from Princeton Alumni Weekly: 'Kent Cooper is leaning into it at Columbia Business.' By the 21st century, 'lean in' (a related form) gained prominence, defined as persevering amid difficulty, and was popularised by Sheryl Sandberg's 2013 book Lean In, urging women to pursue leadership.
In mindfulness contexts, 'lean into the moment' aligns with practices of full presence, transforming challenges into opportunities for empowerment and clarity.
Key Theorist: Jon Kabat-Zinn and Mindfulness-Based Stress Reduction
The most relevant strategy theorist linked to 'leaning into the moment' is **Jon Kabat-Zinn**, a pioneer of mindfulness in modern psychology and stress management. His work embodies the concept through teachings on non-judgmental awareness of the present, even in discomfort.
Biography: Born in 1944 in New York City to a mathematician father (Elia Markenson) and a scientific illustrator mother (Sally Kabat-Dorfman), Kabat-Zinn earned a PhD in molecular biology from MIT in 1971. Initially focused on scientific research, a profound meditation experience shifted his path. In 1979, he founded the Mindfulness-Based Stress Reduction (MBSR) programme at the University of Massachusetts Medical Center, adapting ancient Buddhist practices into secular, evidence-based interventions for chronic pain and stress.
Relationship to the Term: Kabat-Zinn's philosophy directly mirrors 'leaning into the moment'. In MBSR, he teaches 'leaning into' sensations of pain or anxiety without resistance, using phrases like 'being with' or 'allowing' the experience fully. His seminal book Full Catastrophe Living (1990) instructs participants to 'lean into the sharp point' of discomfort, fostering presence and responsiveness. This approach has influenced corporate strategy, leadership training, and resilience-building, where executives 'lean into' uncertainty much like Kabat-Zinn's patients embrace challenging moments. His work underpins global mindfulness initiatives, with over 700 MBSR clinics worldwide by the 2020s.
Kabat-Zinn's integration of mindfulness into strategy emphasises observable benefits: reduced reactivity, enhanced focus, and adaptive decision-making in volatile environments.
References
1. https://www.webclique.net/lean-into-it/
2. https://idioms.thefreedictionary.com/lean+into+(someone+or+something)
3. https://www.merriam-webster.com/dictionary/lean%20in
4. https://grammarphobia.com/blog/2024/08/lean-into.html

|
| |
| |
"A thought experiment (also known by the German term Gedankenexperiment) is a hypothetical scenario imagined to explore the consequences of a theory, principle, or idea when a real-world physical experiment is impossible, unethical, or impractical." - Thought experiment
A **thought experiment**, known in German as Gedankenexperiment, is a hypothetical scenario imagined to explore the consequences of a theory, principle, or idea when conducting a real-world physical experiment is impossible, unethical, or impractical1,7. It involves using hypotheticals to logically reason out solutions to difficult questions, often simulating experimental processes through imagination alone1. These mental exercises are employed across disciplines, particularly philosophy and theoretical sciences, for purposes such as education, conceptual analysis, exploration, hypothesising, theory selection, and implementation2,7.
Thought experiments challenge beliefs, offer fresh perspectives, and examine abstract concepts imaginatively without real-world repercussions3. They construct extreme situations to reveal insights unavailable through formal logic or abstract reasoning, by generating mental models of scenarios and manipulating them via simulation2. Though sometimes circular or rhetorical to emphasise a point, they provide epistemic access to features of representations beyond propositional logic1,2.
Famous Examples
- Mary's Room (Frank Jackson, 1982): A scientist, Mary, knows everything about colour physically from a black-and-white room but learns something new upon seeing red, questioning qualia and physicalism2,3,5.
- Chinese Room (John Searle, 1980s): A person follows rules to manipulate Chinese symbols without understanding them, arguing computers simulate but do not comprehend meaning2,4.
- Drowning Child (Peter Singer, 2009): Would you save a drowning child if it ruined your shoes? This highlights obligations to aid distant strangers2,3.
- Trolley Problem: Divert a trolley to kill one instead of five? Variations probe ethics of action vs. inaction6.
- Brain in a Vat: Your brain in a vat fed simulated experiences questions reality and knowledge4.
Best Related Strategy Theorist: Erwin Schrödinger
Among theorists linked to thought experiments, **Erwin Schrödinger** stands out for his iconic contribution in quantum mechanics, with a profound backstory tying his work to strategic scientific reasoning.
Born in 1887 in Vienna, Austria, Schrödinger was a physicist whose diverse interests spanned philosophy, biology, and Eastern mysticism. He studied at the University of Vienna, served in World War I, and held professorships in Zurich, Berlin (succeeding Planck), Oxford, Graz, and Dublin. Awarded the 1933 Nobel Prize in Physics (shared with Paul Dirac) for wave mechanics, he fled Nazi Germany in 1933 due to his opposition to antisemitism, despite his own complex personal life7. Schrödinger's polymath nature influenced his interdisciplinary approach, later extending to genetics via his 1944 book What is Life?, inspiring DNA discoverers Watson and Crick.
His relationship to the thought experiment is epitomised by **Schrödinger's Cat** (1935), devised to critique the Copenhagen interpretation of quantum mechanics. Imagine a cat in a sealed box with a radioactive atom: if it decays (50% chance), poison releases, killing the cat. Quantum superposition implies the cat is simultaneously alive and dead until observed-a paradoxical Gedankenexperiment highlighting measurement problems and the absurdity of applying quantum rules macroscopically1,7. This strategic tool exposed flaws in prevailing theories, spurring debates on wave function collapse, many-worlds interpretation, and quantum reality. Schrödinger used it not to endorse but to provoke clearer strategies for quantum theory, cementing thought experiments' role in scientific strategy7.
References
1. https://thedecisionlab.com/reference-guide/neuroscience/thought-experiments
2. https://www.missiontolearn.com/thought-experiments/
3. https://bigthink.com/personal-growth/seven-thought-experiments-thatll-make-you-question-everything/
4. https://www.toptenz.net/top-10-most-famous-thought-experiments.php
5. https://adarshbadri.me/philosophy/philosophical-thought-experiments/
6. https://guides.gccaz.edu/philosophy-guide/experiments
7. https://plato.stanford.edu/entries/thought-experiment/
8. https://miamioh.edu/howe-center/hwac/disciplinary-writing-guides/philosophy/thought-experiments.html

|
| |
| |
"AI is leverage because it can scale cognition. It can scale certain kinds of thinking and writing and analysis. And that means individuals can do more. Small teams can do more. It changes the power dynamics." - Bill Gurley - GP at Benchmark
Bill Gurley: The Visionary Venture Capitalist
Bill Gurley serves as a General Partner at Benchmark, one of Silicon Valley's most prestigious venture capital firms. Renowned for his prescient investments in transformative companies such as Uber, Airbnb, and Zillow, Gurley has a track record of identifying technologies that reshape industries and power structures1,4,7. His perspective on artificial intelligence (AI) stems from deep engagement with the sector, including discussions on scaling laws, model sizes, and inference costs in podcasts like BG2 with Brad Gerstner1,2. In the quoted interview with Tim Ferriss, Gurley articulates how AI acts as a force multiplier, enabling individuals and small teams to achieve outsized impact by scaling cognitive tasks traditionally limited by human capacity7.
Context of the Quote
The quote originates from a conversation hosted by Tim Ferriss, where Gurley explores AI's role in the modern economy. He emphasises that AI scales cognition - encompassing thinking, writing, and analysis - thereby democratising high-level intellectual work. This shift empowers solo entrepreneurs and lean teams, disrupting traditional power dynamics dominated by large organisations with vast resources7. Gurley's views align with his broader commentary on AI's rapid evolution, including the implications of massive compute clusters by leaders like Elon Musk, OpenAI, and Meta, and the surprising efficiency of smaller models trained beyond conventional limits1. He highlights real-world applications, such as inference costs outweighing training in products like Amazon's Alexa, underscoring AI's scalability for practical deployment1.
Backstory on Leading Theorists in AI Scaling and Leverage
Gurley's idea of AI as leverage builds on foundational theories in AI scaling laws and cognitive amplification. Key figures include:
- Sam Altman (OpenAI CEO): Altman has championed scaling massive models, predicting that AI will handle every cognitive task humans perform within 3-4 years, unlocking trillions in value from replaced human labour2. Discussions with Gurley reference OpenAI's ongoing training of 405 billion parameter models1.
- Elon Musk: Musk forecasts AI surpassing human cognition across all tasks imminently, driving investments in enormous compute clusters for training and inference scaling by factors of a million or billion1,2.
- Mark Zuckerberg (Meta): Zuckerberg revealed Meta's Llama models, including an 8 billion and 70 billion parameter version, trained past the 'Chinchilla point' - a theoretical diminishing returns threshold from a Google paper - to pack superior intelligence into smaller sizes with fixed datasets1. This supports Gurley's thesis on efficient scaling for broader access.
- Chinchilla Scaling Law Authors (Google DeepMind): Their seminal paper defined optimal data-to-model size ratios for pre-training, challenging earlier assumptions and influencing debates on whether bigger always means better1. Meta's breakthroughs by exceeding this point validate continued gains from extended training.
- Satya Nadella and Jensen Huang: Microsoft and Nvidia leaders emphasise inference scaling, with Nadella noting compute demands exploding as models handle complex reasoning chains, aligning with Gurley's power shift to agile users2.
These theorists collectively underpin Gurley's observation: AI's ability to scale cognition via compute, data, and innovative training redefines leverage, favouring nimble players over bureaucratic giants1,2,3. Gurley's real-world examples, like a 28-year-old entrepreneur superpowered by AI for site selection, illustrate this in action across regions including China3.
Implications for Power Dynamics
Gurley's quote signals a paradigm shift akin to an 'Industrial Revolution for intelligence production', where inference compute scales exponentially, enabling small entities to rival incumbents1,2. Venture trends, such as mega-funds writing huge cheques to AI startups, reflect this frenzy, blurring early and late-stage investing5. Yet Gurley cautions staying 'far from the edge', advocating focus on core innovations amid hype4.
References
1. https://www.youtube.com/watch?v=iTwZzUApGkA
2. https://www.youtube.com/watch?v=yPD1qEbeyac
3. https://www.podchemy.com/notes/840-bill-gurley-investing-in-the-ai-era-10-days-in-china-and-important-life-lessons-from-bob-dylan-jerry-seinfeld-mrbeast-and-more-06a5cd0f-d113-5200-bbc0-e9f57705fc2c
4. https://www.youtube.com/watch?v=D0230eZsRFw
5. https://orbanalytics.substack.com/p/the-new-normal-bill-gurley-breaks
6. https://podcasts.apple.com/ca/podcast/ep20-ai-scaling-laws-doge-fsd-13-trump-markets-bg2/id1727278168?i=1000677811828
7. https://tim.blog/2025/12/17/bill-gurley-running-down-a-dream/

|
| |
| |
"AI is no longer a future concept for BHP. It is increasingly part of how we run our operations. Our focus is on applying it in practical, governed ways that support our teams in achieving safer, more productive and more reliable outcomes." - Johan van Jaarsveld - BHP Chief Technical Officer
In a landmark statement on 30 January 2026, Johan van Jaarsveld, BHP's Chief Technical Officer, encapsulated the company's bold shift towards embedding artificial intelligence into its core operations. This perspective, drawn from BHP's article 'AI is improving performance across global mining operations', underscores a strategic pivot where AI transitions from experimental tool to operational mainstay, driving safer, more productive, and reliable outcomes in one of the world's largest mining enterprises.1,5
Who is Johan van Jaarsveld?
Johan van Jaarsveld assumed the role of Chief Technical Officer at BHP effective 1 March 2024, bringing over 25 years of expertise spanning resources, finance, and technology across continents including Asia, Canada, Australia, and South Africa.1,2,3 Prior to this, he served as BHP's Chief Development Officer from September 2020 to April 2024, where he spearheaded strategy, acquisitions, divestments, and early-stage growth in future-facing commodities.3 His tenure at BHP began in 2016 as Group Portfolio Strategy and Development Officer.
Before joining BHP, van Jaarsveld held senior executive positions at global giants: Senior Vice President of Business Development at Barrick Gold Corporation in Toronto (2015-2016), Managing Director at Goldman Sachs in Hong Kong (2011-2014), Managing Director at The Blackstone Group in Hong Kong (2008-2011), and Vice President at Lehman Brothers (2007).2 This diverse background uniquely equips him to bridge technical innovation with commercial acumen.
Academically, van Jaarsveld holds a PhD in Engineering (Extractive Metallurgy) from the University of Melbourne (2001), a Master of Commerce in Applied Finance from Melbourne Business School (2002), and a Bachelor of Engineering (Chemical) from Stellenbosch University, South Africa.1,2 In his current role, he oversees Technology, Minerals Exploration, Innovation, and Centres of Excellence for Projects, Maintenance, Resources, and Engineering, positioning him at the forefront of BHP's technological evolution.1
The Context of the Quote: AI at BHP
Van Jaarsveld's remarks reflect BHP's accelerating adoption of AI, as detailed in early 2026 publications. AI is enabling BHP to 'understand operations in new ways and act earlier', enhancing performance across global mining sites.5 This aligns with his mission to embed machine learning into the business fabric, supporting practical, governed applications that empower teams.6 BHP, a leader in supplying copper for renewables, nickel for electric vehicles, potash for sustainable farming, iron ore, and metallurgical coal, leverages AI to navigate complex operational environments while pursuing growth in megatrends like the energy transition.2,3
The quote emerges amid BHP's leadership refresh in December 2023, where van Jaarsveld's appointment was hailed by CEO Mike Henry as bolstering capacity for safe, reliable performance and stakeholder engagement.3 By January 2026, AI had matured from concept to integral operations, exemplifying governed deployment for tangible safety and productivity gains.1,5
Leading Theorists and Evolution of AI in Mining
The integration of AI in mining draws from foundational theories in artificial intelligence, machine learning, and operational optimisation, pioneered by key figures whose work underpins industrial applications.
- John McCarthy (1927-2011): Coined 'artificial intelligence' in 1956 and developed LISP, laying groundwork for AI systems adaptable to mining data analysis.[No specific search result; general knowledge of AI history.]
- Geoffrey Hinton, Yann LeCun, and Yoshua Bengio: The 'Godfathers of AI' advanced deep learning neural networks, enabling predictive maintenance and ore grade estimation in mining-core to BHP's AI strategies.[No specific search result; general knowledge.]
- Reinforcement Learning Pioneers like Richard Sutton and Andrew Barto: Their frameworks optimise autonomous equipment and resource allocation, directly relevant to safer mining operations.[No specific search result; general knowledge.]
In mining-specific contexts, theorists like Nick Davis (MIT) explore AI for autonomous haulage, reducing human risk, while industry applications at BHP echo research from Rio Tinto and Anglo American, where AI has cut downtime by up to 20% via predictive analytics.[Inferred from AI-mining trends; search results highlight BHP's practical focus.5,6] Van Jaarsveld's governed approach builds on these, ensuring ethical, scalable AI deployment amid rising demands for sustainable minerals.
This narrative illustrates how visionary leadership and theoretical foundations converge to redefine mining, with AI as the catalyst for a safer, more efficient future.
References
1. https://www.bhp.com/about/board-and-management/johan-van-jaarsveld
2. https://cio-sa.co.za/profiles/johan-van-jaarsveld/
3. https://www.bhp.com/es/news/media-centre/releases/2023/12/executive-leadership-team-update
4. https://www.marketscreener.com/insider/JOHAN-VAN-JAARSVELD-A1Y5XA/
5. https://im-mining.com/2026/01/30/ai-helping-bhp-understand-operations-in-new-ways-and-act-earlier-van-jaarsveld-says/
6. https://www.miningmagazine.com/technology/news-analysis/4414802/bhp-faith-ai
7. https://www.bhp.com/about/board-and-management

|
| |
| |
"Abundance is defined as a state where essential resources - such as housing, energy, healthcare, and transportation - are made flourishing, affordable, and universally accessible through an intentional focus on increasing supply." - Abundance
Abundance is defined as a state where essential resources - such as housing, energy, healthcare, and transportation - are made flourishing, affordable, and universally accessible through an intentional focus on increasing supply.1,2
Comprehensive Definition and Context
The concept of abundance represents a paradigm shift in political and economic thinking, advocating a 'politics of plenty' that prioritises building and innovation over scarcity-driven approaches. Coined prominently in the 2025 book Abundance by Ezra Klein and Derek Thompson, it critiques how past regulations - intended to solve 1970s problems - now hinder progress in the 2020s by blocking urban density, green energy, and infrastructure projects.2,4
At its core, abundance calls for liberalism that not only protects but actively builds. It argues that modern crises stem from insufficient supply rather than mere distribution failures. Solutions involve streamlining regulations, boosting innovation in areas like clean energy, housing, and biotechnology, and fostering high-density economic hubs to enhance idea generation and mobility.1,2 This contrasts with traditional scarcity mindsets, where progressives fear growth and conservatives resist government intervention, trapping societies in unaffordability.4
Key pillars include:
- Housing: Permitting high-rise developments in vital cities without undue barriers to increase supply and affordability.1
- Energy and Infrastructure: Accelerating clean energy and transport projects to meet demands sustainably.2
- Healthcare and Innovation: Expanding medical residencies, drug approvals, and R&D while balancing equity with supply growth - a 'floor without a ceiling' model, as seen in France.1
- Governance Reform: Reducing legalistic processes that prioritise procedure over outcomes.7
Critics note it de-emphasises redistribution in favour of supply-side innovation, potentially overlooking power dynamics, though proponents see it as a path beyond socialist left and populist right extremes.3,4,5
Key Theorist: Ezra Klein
Ezra Klein is the pre-eminent theorist behind the abundance agenda, co-authoring the seminal book Abundance with Derek Thompson. A leading liberal thinker, Klein shifted focus from political polarisation to economic abundance, arguing it offers a unifying path forward.1,2
Born in 1984 in Irvine, California, Klein rose through blogging on Wonkblog at The Washington Post, analysing policy with data-driven rigour. He co-founded Vox in 2014 as editor-in-chief, building it into a platform for explanatory journalism. In 2021, he launched The Ezra Klein Show podcast and joined The New York Times as a columnist, influencing discourse on liberalism's failures.1,2
Klein's relationship to abundance stems from observing how liberal governance stagnated: over-regulation stifles building, exacerbating shortages in housing and energy. In conversations, like with Tyler Cowen, he defends scaling elite institutions (e.g., doubling Harvard's size) and critiques demand-side fixes without supply increases.1 His classically liberal view of power - checking arbitrary domination - underpins abundance as a corrective to equity-obsessed policies that neglect production.3 Klein positions it as reclaiming progressivism's building ethos, countering both left-wing caution and right-wing anti-statism.2,4
Through Abundance, Klein provides intellectual firepower for a 'liberalism that builds', impacting policymakers and coalitions seeking tangible solutions.6,7
References
1. https://conversationswithtyler.com/episodes/ezra-klein-3/
2. https://www.simonandschuster.com/books/Abundance/Ezra-Klein/9781668023488
3. https://www.peoplespolicyproject.org/2025/06/09/abundance-has-a-theory-of-power/
4. https://en.wikipedia.org/wiki/Abundance_(Klein_and_Thompson_book)
5. https://www.bostonreview.net/articles/the-real-path-to-abundance/
6. https://www.inclusiveabundance.org/abundance-in-action/published-work/abundance-a-primer
7. https://www.eesi.org/articles/view/abundance-and-its-insights-for-policymakers

|
| |
| |
"I regard consciousness as fundamental. I regard matter as derivative from consciousness. We cannot get behind consciousness. Everything that we talk about, everything that we regard as existing, postulates consciousness." - Max Planck - Nobel laureate
This striking statement, made by Max Planck in a 1931 interview with The Observer, encapsulates a radical departure from the materialist worldview dominant in physics at the time. Planck, the father of quantum theory, challenges the notion that matter is the foundation of existence, proposing instead that consciousness underpins all reality. Spoken amid the revolutionary upheavals of early quantum mechanics, the quote reflects his lifelong reconciliation of empirical science with metaphysical inquiry.1,2,3
Max Planck: Life, Legacy, and Philosophical Evolution
Born in 1858 in Kiel, Germany, Max Karl Ernst Ludwig Planck rose from a family of scholars to become one of the 20th century's most influential physicists. He studied at the universities of Munich and Berlin, earning his doctorate in 1879. Initially drawn to thermodynamics, Planck's pivotal moment came in 1900 when he introduced the concept of energy quanta to resolve the 'ultraviolet catastrophe' in black-body radiation-a breakthrough that birthed quantum theory. For this, he received the Nobel Prize in Physics in 1918.3
Planck's career spanned turbulent times: he served as president of the Kaiser Wilhelm Society (later the Max Planck Society) and navigated the intellectual and political storms of two world wars. A devout Lutheran, he grappled with the implications of his discoveries, often emphasising the limits of scientific materialism. In works like Where Is Science Going? (1932), he argued that science presupposes an external world known only through consciousness, echoing themes in his famous quote.3,5
By 1931, at age 72, Planck was reflecting on quantum mechanics' philosophical ramifications. The interview in The Observer captured his mature view: matter derives from consciousness, not vice versa. This idealist stance contrasted with contemporaries like Einstein, who favoured a deterministic universe, yet aligned with Planck's belief in a 'conscious and intelligent Mind' as the force binding atomic particles.3,5
The Context of the Quote: Quantum Revolution and Metaphysical Stirrings
The quote emerged during a period of crisis in physics. Quantum mechanics, propelled by Planck's quanta, Heisenberg's uncertainty principle, and Schrödinger's wave equation, shattered classical determinism. Reality at the subatomic level appeared probabilistic, observer-dependent-raising profound questions about observation's role. Planck, who reluctantly accepted these implications, saw consciousness not as a quantum byproduct but as fundamental.4,5
In the interview, Planck addressed the 'reality crisis': if physical laws are mental constructs, what grounds existence? His response prioritised consciousness as the irreducible starting point, influencing later debates in quantum interpretation, such as the Copenhagen interpretation where measurement (tied to observation) collapses the wave function.3
Leading Theorists on Consciousness and Matter
Planck's views resonate with a lineage of thinkers bridging physics, philosophy, and metaphysics. Here are key figures whose ideas shaped or paralleled his:
- Immanuel Kant (1724-1804): The German philosopher posited that space, time, and causality are a priori structures of the mind, not properties of things-in-themselves. Planck echoed this by insisting we cannot 'get behind consciousness' to access unmediated reality.3
- Ernst Mach (1838-1916): Planck's early influence, Mach advocated 'economical descriptions' of phenomena, rejecting absolute space and atoms as metaphysical. His positivism nudged Planck towards quantum ideas but clashed with Planck's later spiritual realism.5
- Arthur Eddington (1882-1944): The British astrophysicist, like Planck, argued in The Nature of the Physical World (1928) that the mind constructs physical laws. He quipped, 'We have found a strange footprint on the shores of the unknown,' mirroring Planck's consciousness primacy.5
- Werner Heisenberg (1901-1976): Planck's successor, Heisenberg's uncertainty principle highlighted the observer's role. Though more agnostic, he noted in Physics and Philosophy (1958) that quantum theory demands a 'sharper formulation of the concept of reality,' aligning with Planck's critique.3
- David Bohm (1917-1992): Later, Bohm developed implicate order theory, positing a holistic reality where consciousness and matter interpenetrate-directly inspired by Planck's 'matrix of all matter' as a conscious mind.5
These theorists, from Kantian idealism to quantum pioneers, form the intellectual backdrop. Planck stands out for wedding rigorous physics with unapologetic metaphysics, suggesting science's foundations rest on conscious postulate.1,3,5
Enduring Relevance
Planck's declaration prefigures modern discussions in philosophy of mind, panpsychism, and quantum consciousness theories (e.g., by Roger Penrose and Stuart Hameroff). It invites reflection: if consciousness is fundamental, how does this reshape our understanding of the universe, free will, and even artificial intelligence? As Planck implied, all inquiry begins-and ends-with the mind.4,5
References
1. https://libquotes.com/max-planck/quote/lbm8d8r
2. https://www.quotescosmos.com/quotes/Max-Planck-quote-1.html
3. https://en.wikiquote.org/wiki/Max_Planck
4. https://bigthink.com/words-of-wisdom/max-planck-i-regard-consciousness-as-fundamental/
5. https://www.informationphilosopher.com/solutions/scientists/planck/
6. https://todayinsci.com/P/Planck_Max/PlanckMax-Quotations.htm

|
| |
| |
"Tokenisation is the process of converting sensitive data or real-world assets into non-sensitive, unique digital identifiers (tokens) for secure use, commonly seen in data security (replacing credit card numbers with tokens) or blockchain (representing assets like real estate as digital tokens)." - Tokenisation
Tokenisation is the process of replacing sensitive data or real-world assets with non-sensitive, unique digital identifiers called tokens. These tokens have no intrinsic value or meaning outside their specific context, ensuring security in data handling or asset representation on blockchain networks.
In data security, tokenisation substitutes sensitive information like credit card numbers with tokens stored in secure vaults, allowing safe processing without exposing originals. This meets standards such as PCI DSS, GDPR, and HIPAA, reducing breach risks as stolen tokens are useless without vault access.
In blockchain and crypto, it converts assets like real estate, artwork, or shares into digital tokens on a blockchain, enabling fractional ownership, trading, and custody while linking to the physical asset in secure facilities.
How Tokenisation Works
Typically involves three parties: the data/asset owner, an intermediary (e.g., merchant), and a secure vault provider. Sensitive data is sent to the vault, replaced by a unique token, and the original is discarded or stored securely. Tokens preserve data format and length for system compatibility, unlike encryption which alters them.
- Vaulted Tokenisation: Original data stays in a central vault; tokens are de-tokenised only when needed within the vault.
- Format-Preserving: Tokens match original data structure for seamless integration.
- Blockchain Tokenisation: Assets are represented by tokens on networks like Ethereum, with compliance and custody mechanisms.
Benefits of Tokenisation
- Enhanced security against breaches and insider threats.
- Regulatory compliance with reduced audit scope.
- Improved performance via smaller token sizes.
- Data anonymisation for analytics and AI/ML.
- Flexibility across cloud, on-premises, and hybrid setups.
Key Theorist: Don Tapscott
Don Tapscott, a pioneering strategist in digital economics and blockchain, is closely linked to asset tokenisation through his co-authorship of Blockchain Revolution (2016). With Alex Tapscott, he popularised the concept of tokenising real-world assets, arguing it democratises finance by enabling fractional ownership and liquidity for illiquid assets like property.
Born in 1947 in Canada, Tapscott began as a management consultant, authoring bestsellers like The Digital Economy (1995), which foresaw internet-driven business shifts. He founded the Tapscott Group and New Paradigm, advising firms and governments. His blockchain work critiques centralised finance, promoting decentralised ledgers for transparency. As Chair of the Blockchain Research Institute, he influences policy, with tokenisation central to his vision of a 'token economy' transforming global markets.
References
1. https://brave.com/glossary/tokenization/
2. https://entro.security/glossary/tokenization/
3. https://www.fortra.com/blog/what-data-tokenization-key-concepts-and-benefits
4. https://www.fortanix.com/faq/tokenization/data-tokenization
5. https://www.mckinsey.com/featured-insights/mckinsey-explainers/what-is-tokenization
6. https://www.ibm.com/think/topics/tokenization
7. https://www.keyivr.com/us/knowledge/guides/guide-what-is-tokenization/
8. https://chain.link/education-hub/tokenization

|
| |
| |
"The pleasant surprise is how much you can accomplish when you properly harness your agents, and how big companies are leaning in and able to actually get volume done on that basis." - Nate B Jones - AI News & Strategy Daily
Context of the Quote
This quote from Nate B Jones captures a pivotal moment in the evolution of AI agents within enterprise settings. Delivered in his AI News & Strategy Daily series, it highlights the unexpected productivity gains when organisations implement AI agents correctly. Jones emphasises that major firms like JP Morgan and Walmart are already deploying these systems at scale, achieving high-volume outputs that traditional software cycles could not match1,2. The core insight is that proper orchestration-combining AI with human oversight-unlocks disproportionate value, countering the hype-driven delays many companies face.
Backstory on Nate B Jones
Nate B Jones is a leading voice in enterprise AI strategy, known for his pragmatic frameworks that guide businesses from AI hype to production deployment. Through his platform natebjones.com and Substack newsletter Nate's Newsletter, he distils complex AI developments into actionable insights for executives1,2,7. Jones produces daily video briefings like AI News & Strategy Daily, where he analyses real-world use cases, warns against common pitfalls such as over-reliance on unproven models, and provides custom prompts for rapid agent prototyping2,4.
His work focuses on bridging the gap between AI potential and enterprise reality. For instance, he critiques the 'human throttle'-where hesitation and risk aversion limit agent autonomy-and advocates for decision infrastructure like audit logs and reversible processes to build trust3. Jones has documented production AI agents at scale, urging leaders to act swiftly as competitors gain 'durable advantage' through accumulated institutional intelligence2. His library of use cases spans finance (e.g., JP Morgan's choreographed workflows) to operations, emphasising that agents excel in 'level four' tasks: AI drafts, humans review, then AI proceeds1. By October 2025, his briefings were already forecasting 2026 as a year of job-by-job AI transformation5.
Leading Theorists and the Subject of AI Agents
AI agents-autonomous systems that perceive, reason, act, and learn to achieve goals-represent a shift from passive tools to proactive workflows. Nate B Jones builds on foundational work by key theorists:
- Stuart Russell and Peter Norvig: Pioneers of modern AI, their textbook Artificial Intelligence: A Modern Approach defines rational agents as entities maximising expected utility in dynamic environments. This underpins Jones's emphasis on structured autonomy over raw intelligence1,3.
- Andrew Ng: Dubbed the 'Godfather of AI,' Ng popularised agentic workflows at Stanford and through Landing AI. He advocates 'agentic reasoning,' where AI chains tools and decisions, aligning with Jones's production playbooks for enterprises like Walmart2.
- Yohei Nakajima: Creator of BabyAGI (2023), an early open-source agent framework that demonstrated recursive task decomposition. This inspired Jones's warnings against hype, stressing expert-designed workflows for complex problems1,4.
- Anthropic Researchers: Their work on Constitutional AI and agent patterns (e.g., long-running memory) informs Jones's analyses of scalable agents, as seen in his breakdowns of reliable architectures6.
Jones synthesises these ideas into enterprise strategy, arguing that agents are not future tech but 'production infrastructure now.' He counters delays by outlining six principles for quick builds (days or weeks), including context-aware prompts and risk-mitigated deployment2. This positions him as a practitioner-theorist, translating academic foundations into C-suite playbooks amid the 2025-2026 agent revolution.
Broader Implications for Workflows
Jones's quote underscores a paradigm shift: AI agents amplify top human talent, making them 'more fingertippy' rather than replacing them1. Big companies succeed by 'leaning in'-auditing processes, building observability, and iterating fast-yielding volume at scale. For leaders, the message is clear: harness agents properly, or risk irreversible competitive lag2,3.
References
1. https://www.youtube.com/watch?v=obqjIoKaqdM
2. https://natesnewsletter.substack.com/p/executive-briefing-your-2025-ai-agent
3. https://www.youtube.com/watch?v=7NjtPH8VMAU
4. https://www.youtube.com/watch?v=1FKxyPAJ2Ok
5. https://natesnewsletter.substack.com/p/2026-sneak-peek-the-first-job-by-9ac
6. https://www.youtube.com/watch?v=xNcEgqzlPqs
7. https://www.natebjones.com

|
| |
| |
"A stablecoin is a type of cryptocurrency designed to maintain a stable value, unlike volatile assets like Bitcoin, by pegging its price to a stable reserve asset, usually a fiat currency (like the USD) or a commodity (like gold)." - Stablecoin
What is a Stablecoin?
A **stablecoin** is a type of cryptocurrency engineered to preserve a consistent value relative to a specified asset, such as a fiat currency (e.g., the US dollar), a commodity (e.g., gold), or a basket of assets, in stark contrast to the high volatility of assets like Bitcoin.
Unlike traditional cryptocurrencies, stablecoins employ stabilisation mechanisms including reserve assets held by custodians or algorithmic protocols that adjust supply and demand to sustain the peg. Fiat-backed stablecoins, the most common variant, mirror money market funds by holding reserves in short-term assets like treasury bonds, commercial paper, or bank deposits. Commodity-backed stablecoins peg to physical assets like gold, while cryptocurrency-backed ones, such as DAI or Wrapped Bitcoin (WBTC), use overcollateralised crypto reserves managed via smart contracts on decentralised networks.
Types of Stablecoins
- Fiat-backed: Centralised issuers hold equivalent fiat reserves (e.g., USD) to support 1:1 redeemability.
- Commodity-backed: Pegged to commodities, with issuers maintaining physical reserves.
- Cryptocurrency-backed: Collateralised by other cryptocurrencies, often overcollateralised to buffer volatility.
- Algorithmic: Rely on smart contracts to dynamically adjust supply without full reserves, though prone to failure.
Despite the name, stablecoins are not immune to depegging, as evidenced by historical failures amid market stress or redemption pressures, potentially triggering systemic risks akin to fire-sale contagions in traditional finance. They facilitate rapid, low-cost blockchain transactions, serving as a bridge between fiat and crypto ecosystems for payments, settlements, and trading.
Regulatory Landscape
Governments worldwide are intensifying oversight due to stablecoins' growing role in transactions. For instance, Nebraska's Financial Innovation Act (2021, updated 2024) permits digital asset depositories to issue stablecoins backed by reserves in FDIC-insured institutions.
Key Theorist: Robert Shiller and the Conceptual Foundations
The most relevant strategy theorist linked to stablecoins is **Robert Shiller**, a Nobel Prize-winning economist whose pioneering work on financial stability, behavioural finance, and asset pricing underpins the economic rationale for pegged digital assets. Shiller's theories address the volatility that stablecoins explicitly counter, positioning them as practical applications of stabilising speculative markets.
Born in 1946 in Detroit, Michigan, Shiller earned his PhD in economics from MIT in 1972 under advisor Robert Solow. He joined Yale University in 1982, where he remains the Sterling Professor of Economics. Shiller gained prominence for developing the Case-Shiller Home Price Index, a leading US housing market benchmark. His seminal book, Irrational Exuberance (2000), presciently warned of the dot-com bubble and later the 2008 financial crisis, critiquing how narratives drive asset bubbles.
Shiller's relationship to stablecoins stems from his advocacy for financial innovations that mitigate volatility. In works like Finance and the Good Society (2012), he explores stabilising mechanisms such as index funds and derivatives, which parallel stablecoin pegs by tethering values to underlying assets. He has discussed cryptocurrencies in interviews and writings, noting their potential to enhance financial inclusion if stabilised-echoing stablecoins' design to combine crypto's efficiency with fiat-like reliability. Shiller's CAPE (Cyclically Adjusted Price-to-Earnings) ratio exemplifies pegging metrics to long-term fundamentals, a concept mirrored in stablecoin reserves. While not a crypto native, his behavioural insights explain depegging risks from herd mentality, making him the foremost theorist for stablecoin strategy in volatile markets.
References
1. https://en.wikipedia.org/wiki/Stablecoin
2. https://csrc.nist.gov/glossary/term/stablecoin
3. https://www.fidelity.com/learning-center/trading-investing/what-is-a-stablecoin
4. https://www.imf.org/en/publications/fandd/issues/2022/09/basics-crypto-conservative-coins-bains-singh
5. https://klrd.gov/2024/11/15/stablecoin-overview/
6. https://am.jpmorgan.com/us/en/asset-management/adv/insights/market-insights/market-updates/on-the-minds-of-investors/what-is-a-stablecoin/
7. https://www.bankofengland.co.uk/explainers/what-are-stablecoins-and-how-do-they-work
8. https://bvnk.com/blog/stablecoins-vs-bitcoin
9. https://business.cornell.edu/article/2025/08/stablecoins/

|
| |
|