| |
|
A daily bite-size selection of top business content.
PM edition. Issue number 1289
Latest 10 stories. Click the button for more.
|
| |
|
"There is always a flight to quality when there are things going on in the world, and we are quality." - Jane Fraser - Citi CEO
Citigroup's Services division has emerged as a cornerstone of stability, delivering net income of 2,2 billion dollars in the first quarter of 2026 with a return on tangible common equity of 27 percent, underscoring its role in attracting deposits and flows during uncertain times. This performance reflects deeper structural shifts within the bank, where cross-border transactions grew 12 percent and deposits expanded 16 percent, drawing institutional clients seeking reliable custody and administration amid global disruptions. The mechanism hinges on Citi's vast network spanning 180 countries, enabling it to capture operating deposits that fuel low-cost funding while rivals grapple with volatile liabilities. In practice, this translates to assets under custody and administration surging over 20 percent, as treasurers prioritise custodians with proven resilience in crises.
Geopolitical tensions and macroeconomic headwinds have consistently triggered capital reallocations towards established players, a pattern evident in prior episodes like the 2022 energy shocks and 2024 supply chain fractures. During such flights, quality manifests in operational reliability: Citi's mandate wins jumped 40 percent, signalling trust in its execution amid fragmented trade flows. Deposits, often overlooked as a defensive asset, become prized when short-term rates spike and liquidity dries up elsewhere; Citi's average deposits rose 4 percent in recent periods, bolstered by relationship transfers and higher client balances up 8 percent. This inflow supports a cost of credit at 2,8 billion dollars firm-wide, with U.S. card losses guided at 4,0 to 4,5 percent, demonstrating prudent risk management.
Jane Fraser's leadership since 2021 has intensified this positioning through a sweeping transformation, completing over 80 percent of a multiyear overhaul that simplifies processes and embeds AI for efficiency. Headcount reductions, including nearly 500 million dollars in severance in Q1 2026, accompany automation that eliminates redundant roles while preserving client-facing expertise. Fraser's internal directives demand a commercial mindset, urging staff to secure the full wallet rather than secondary positions, directly enhancing deposit and flow capture. This cultural pivot addresses longstanding critiques of Citi's inefficiency, where returns lagged peers; now, with an efficiency ratio of 58 percent and ROTCE at 13,1 percent, the bank edges towards its 10 to 11 percent full-year 2026 target.
Historical Context and Strategic Evolution
Citigroup's pedigree as a global powerhouse traces to its merger origins, but pre-Fraser eras suffered from sprawl: sprawling consumer banking, regulatory fines exceeding 10 billion dollars post-2008, and returns mired below 5 percent. Fraser's 2021 ascent marked a pivot to five core businesses-Services, Markets, Banking, U.S. Personal Banking, Wealth-exiting non-core personal banking in 14 markets to focus on institutional strengths. This refocus amplified Services as the crown jewel, generating 17 percent revenue growth in Q1 2026 on 40 percent mandate expansion, as clients consolidate with fewer, trusted providers. Markets complemented with 7 billion dollars revenue up 19 percent and 2,6 billion dollars net income, thriving on volatility that funnels trades to liquid platforms.
The Q1 2026 earnings, reported April 14 with net income of 5,8 billion dollars, EPS of 3,06 dollars, and 24,6 billion dollars revenue up 14 percent, validated this trajectory. Four of five cores posted double-digit revenue gains, with positive operating leverage across most units, despite 14,3 billion dollars expenses up 7 percent. Capital strength at 12,7 percent CET1-110 basis points above requirements-affords flexibility for buybacks and dividends, reinforcing quality perceptions. Yet, seasonality tempers optimism; Fraser cautioned that macro uncertainty and investment needs persist, with credit reserves near 22 billion dollars.
Technological and Operational Underpinnings
AI and automation underpin Citi's quality claim, re-engineering workflows to sustain services amid flux. As transformation nears completion, roles evolve: some vanish, others emerge in high-value areas like investment banking. This mirrors industry trends where banks deploy gen AI for compliance and tokenization, enhancing cross-border efficiency-Citi's 12 percent transaction growth exemplifies this. Deposits benefit indirectly; streamlined onboarding and custody draw operating balances, which grew robustly as clients shift from higher-yield alternatives.
In mathematical terms, the value of these flows ties to funding cost dynamics. Consider deposit beta, the sensitivity of deposit rates to policy changes: lower betas preserve net interest margins during hikes. Citi's operating deposits, sticky due to services integration, exhibit betas below peers, formalised as where is deposit rate and policy rate. Empirical evidence from Q1 shows resilience, with balances up despite rate uncertainty. Services' high ROTCE-27 percent-derives from scalable revenues: fee income scales with transaction volumes , with transaction fee, volume, custody rate, assets.
Debates and Investor Scrutiny
Sceptics question sustainability: Citi's stock dipped 0,05 percent post-earnings to 126,22 dollars, reflecting doubts on full-year delivery amid severance costs and macro risks. Critics highlight historical underperformance; Euromoney notes Fraser's challenge in fixing woeful returns, with structure preceding profitability. Job cuts-potentially 20 000 roles-risk morale erosion, countering Fraser's human-centered ethos. Rivals like JPMorgan boast superior ROTCE above 20 percent consistently, pressuring Citi to close the gap. Objections centre on execution: will AI deliver without regulatory hurdles, and can Services maintain 29,9 percent ROTCE amid competition from fintech custodians?
Fraser counters with results: Wealth's 21 percent pretax margin and 10,1 percent ROTCE, alongside Retail Services' 7 percent revenue rise on 3 percent balance growth. Management holds 2026 guidance unchanged, targeting 60 percent efficiency via headcount trims. Debates pivot to macro: conflicting data complicates Fed decisions, yet Citi's 110 basis points buffer insulates against downturns.
Strategic Tensions and Competitive Landscape
Tension arises between simplification and global ambition. Exiting legacy units freed 1 billion dollars in efficiencies, but retaining 180-country footprint demands scale rivals lack. Services thrives on network effects: larger custody basins attract mandates, creating a virtuous cycle formalised as where mandates, network size, services quality. Markets' volatility capture-equities and fixed income up amid flows-positions Citi for flight scenarios, where quality means liquidity and prime brokerage.
Versus peers, Citi lags in consumer scale but leads in cross-border: 16 percent deposit growth outpaces JPMorgan's domestic focus. Fraser's memo slams old habits, grading on results not effort, aligning incentives with flow capture. Wealth integration and leadership changes in capital markets bolster this.
Implications and Enduring Relevance
This positioning matters as tail risks mount-elections, trade wars, AI-driven disruptions. Flights to quality historically boost top-tier banks' deposits 5 to 10 percent, per past cycles; Citi's Q1 gains presage this. For investors, ROTCE trajectory signals value unlocking: from sub-10 percent to 13,1 percent, with 56 percent EPS growth. Clients benefit from resilient infrastructure, tokenization pilots enhancing settlement.
Fraser's vision-a disciplined, winning Citi-hinges on execution in 2026, proving transformation yields consistent 10 to 11 percent returns. Amid uncertainty, quality endures: deep relationships, tech-enabled services, and balance sheet strength draw flows when others falter. This not only sustains funding but amplifies franchise value, cementing Citi's role in global finance.
|
| |
| |
"I kind of disagree with Yann [LeCun] on a few things.. I think there might be a 50/50 chance there's some things.. missing that we still need to make breakthroughs in, perhaps world models... But my betting is pretty strongly that we've seen how successful these foundation models have been. They can do incredibly impressive things." - Demis Hassabis - Google DeepMind CEO
The disagreement between Demis Hassabis and Yann LeCun represents one of the most consequential technical debates in AI development: whether the current trajectory of large language models and foundation models will suffice to reach artificial general intelligence, or whether fundamentally different architectures-specifically world models-are necessary.1,2 Hassabis's statement reflects genuine uncertainty about this question while expressing confidence in the demonstrated capabilities of existing approaches, yet this framing obscures a more complex strategic reality in which both positions may be partially correct.
The LeCun Critique and Its Foundations
Yann LeCun, Chief AI Scientist at Meta, has articulated a systematic critique of large language models as a path to AGI. His argument centers on fundamental architectural limitations: LLMs excel at pattern matching and text prediction but lack the capacity for causal reasoning, physical intuition, and hypothesis testing through mental simulation.5 LeCun contends that these capabilities are not merely enhancements but essential prerequisites for systems that can reason about novel scenarios, plan across extended time horizons, and generate genuinely original insights rather than recombining training data in sophisticated ways.
This critique gains force from observable limitations in current systems:
- LLMs struggle with long-horizon causality and cannot reliably simulate how interventions propagate through complex systems over time
- They lack grounding in physical reality and cannot develop intuitive physics from first principles
- They cannot perform hypothesis testing through mental simulation-the capacity to imagine counterfactuals and evaluate their plausibility
- They generate novel combinations of existing concepts but rarely produce genuinely new scientific theories or technological breakthroughs
Hassabis's Measured Disagreement
Hassabis does not dismiss LeCun's concerns but rather assigns them a probabilistic weight: a 50/50 chance that breakthroughs in world models remain necessary.1 This formulation is revealing. It acknowledges that the case for architectural innovation is substantial enough to warrant serious consideration, yet expresses greater confidence in the trajectory of foundation models. His "strong betting" on foundation models reflects both their demonstrated capabilities and the practical reality that scaling these systems continues to yield improvements.5
The distinction matters because Hassabis is not claiming that foundation models are sufficient in principle, only that they have proven more capable than skeptics anticipated and that their development path remains productive. This is a claim about empirical trajectory rather than theoretical sufficiency.
World Models: The Missing Ingredient or Complementary Layer?
World models represent a distinct architectural approach: systems that learn latent representations of physical reality by ingesting video, sensor data, or simulation environments and developing internal models of causality, object permanence, dynamics, and spatial reasoning.5 Rather than predicting text tokens, world models predict future states of the physical world given current observations and proposed actions.
The strategic question is whether world models should replace foundation models or augment them. Hassabis has increasingly emphasized that the future likely involves convergence rather than replacement:5
- Foundation models (like Gemini) handle multimodal data across text, images, video, and audio but lack true understanding of physics and causality
- World models capture spatial dynamics, intuitive physics, and mechanical understanding-the embodied knowledge that cannot be fully conveyed through language alone
- Integrated systems combining both capabilities could enable robotics, autonomous driving, and scientific simulation at scales currently impossible
This convergence thesis sidesteps the binary framing of the Hassabis-LeCun disagreement. It suggests that both architectures address genuine gaps in the other and that AGI may require their synthesis rather than the victory of one approach.
The Empirical Case for Foundation Models
Hassabis's confidence in foundation models rests on concrete achievements. These systems have demonstrated:
- Multimodal reasoning across text, images, video, and audio in ways that were not possible five years ago
- Transfer learning across domains-capabilities developed in one context generalizing to novel problems
- Emergent abilities that appear at scale without explicit programming for those capabilities
- Practical utility in scientific domains, from protein structure prediction (AlphaFold) to materials discovery
The scaling laws that govern foundation models have not yet plateaued, and each increase in compute, data, and model size has continued to yield measurable improvements.5 This empirical success creates a rational basis for continued investment in this direction, even if theoretical arguments suggest limitations.
The Timing and Resource Allocation Problem
Beneath the technical disagreement lies a practical question about resource allocation. If world models are necessary but foundation models are not yet exhausted, the optimal strategy involves parallel development rather than pivot. Yet resources are finite, and the competitive dynamics of AI development create pressure to commit heavily to whichever approach appears most promising in the near term.
Hassabis's 50/50 framing may reflect this tension. By acknowledging substantial probability that world models are necessary while betting more heavily on foundation models, he preserves optionality while maintaining focus on the approach with demonstrated momentum. DeepMind has invested in world model research (including projects like Genie and VEO), but this remains secondary to foundation model scaling.2
The AGI Definition Problem
The disagreement also hinges on how AGI is defined. If AGI requires only superhuman performance on a broad range of tasks, foundation models may suffice. If AGI requires causal reasoning, hypothesis testing, and the capacity to generate genuinely novel scientific insights, world models become more essential.5 Hassabis has defined AGI as a system exhibiting all human cognitive capabilities-true innovation and creativity, planning, reasoning, consistent performance across domains, continual learning, and the ability to understand and explain the world through simulation and hypothesis testing.5 By this definition, current foundation models fall short, yet Hassabis still expresses confidence that scaling them will eventually bridge the gap.
Strategic Implications
The practical consequence of this debate is that AI development is proceeding along multiple paths simultaneously. OpenAI, Google, Anthropic, and xAI continue scaling LLMs and foundation models.5 Simultaneously, world model research is accelerating, with Tesla's autonomous driving systems relying heavily on embodied AI and end-to-end neural networks that function as world models.5 DeepMind itself is investing in both directions.
This parallel development strategy reduces the risk of betting entirely on one architectural approach while maintaining the momentum of the most productive current direction. It also means that the resolution of the Hassabis-LeCun disagreement may come not from theoretical argument but from empirical demonstration: whichever approach reaches AGI-level capabilities first will vindicate its proponents, while the other will be repositioned as a necessary component rather than a sufficient path.
The Unresolved Question
Hassabis's measured disagreement with LeCun ultimately reflects genuine uncertainty in the field. The question of whether foundation models can scale to AGI or whether world models are necessary remains open.5 His 50/50 probability assignment is not evasion but honest acknowledgment that the evidence does not yet decisively favor either position. The strong betting on foundation models reflects their demonstrated capabilities and continued progress, not certainty about their sufficiency. As development continues, this probabilistic assessment may shift-but for now, it captures the state of technical knowledge: foundation models have exceeded expectations, but the case for architectural innovation remains substantial.
References
1. Demis Hassabis: Why AGI is Bigger than the Industrial ... - YouTube - 2026-04-07 - https://www.youtube.com/watch?v=SSya123u9Yk
2. Google DeepMind CEO Demis Hass… - Big Technology Podcast - 2025-05-21 - https://podcasts.apple.com/us/podcast/google-deepmind-ceo-demis-hassabis-google-co-founder/id1522960417?i=1000709250044
3. DeepMind CEO Reveals Why World Models Are the Future of AI ... - 2026-01-03 - https://www.youtube.com/watch?v=B3IYbfHqDas
4. 20VC: DeepMind's Demis Hassabis on Why AGI is Bigger than the ... - 2026-04-07 - https://podcasts.apple.com/gb/podcast/20vc-deepminds-demis-hassabis-on-why-agi-is-bigger/id958230465?i=1000759991057
5. Demis Hassabis on what's next for Google DeepMind - 2026-01-26 - https://sources.news/p/interview-demis-hassabis-sources
6. AGI Needs World Models and State of World Models - 2026-01-20 - https://www.nextbigfuture.com/2026/01/agi-needs-world-models-and-state-of-world-models.html
7. Hassabis on an AI Shift Bigger Than Industrial Age - YouTube - 2026-01-21 - https://www.youtube.com/watch?v=Xcyox1CP1Wk
8. DeepMind CEO Demis Hassabis on How A.I. Is Reshaping Google - 2025-05-26 - https://www.youtube.com/watch?v=U3d2OKEibQ4
9. Sir Demis Hassabis becomes the latest to say that ChatGPT is a ... - 2026-01-22 - https://garymarcus.substack.com/p/breaking-sir-demis-hassabis-becomes
10. The Hardest Problem AI Ever Solved, with Google DeepMind CEO - 2026-04-07 - https://www.youtube.com/watch?v=C0gErQtnNFE
11. Demis Hassabis on Gemini 3, world models, and the AI bubble - 2025-11-18 - https://sources.news/p/demis-hassibas-on-gemini-3-world
12. 20VC with Harry Stebbings - YouTube - 2025-04-10 - https://www.youtube.com/@20VC
13. Hassabis on an AI Shift Bigger Than Industrial Age - YouTube - 2026-01-20 - https://www.youtube.com/watch?v=BbIaYFHxW3Y
14. 20VC | The Intersection of Venture Capital and Media - 2026-04-07 - https://www.thetwentyminutevc.com
15. Demis Hassabis (Co-founder and CEO of DeepMind) - YouTube - 2025-12-16 - https://www.youtube.com/watch?v=PqVbypvxDto
!["I kind of disagree with Yann [LeCun] on a few things.. I think there might be a 50/50 chance there’s some things.. missing that we still need to make breakthroughs in, perhaps world models... But my betting is pretty strongly that we’ve seen how successful these foundation models have been. They can do incredibly impressive things." - Quote: Demis Hassabis - Google DeepMind CEO](https://globaladvisors.biz/wp-content/uploads/2026/04/20260413_13h15_GlobalAdvisors_Marketing_Quote_DemisHassabis_GAQ.png)
|
| |
| |
"An inverted yield curve occurs when short-term bonds offer higher interest rates (yields) than long-term bonds, which is the opposite of the normal upward-sloping yield curve, and it's considered a reliable, though not immediate, predictor of an upcoming economic recession, signaling investor pessimism about future growth as they rush to lock in long-term rates." - Inverted yield curve
An **inverted yield curve** arises when yields on short-term bonds surpass those on long-term bonds, defying the typical upward-sloping curve where longer maturities command higher returns to compensate for extended risk1,2,5. This phenomenon reflects investor expectations of subdued future growth, prompting a flight to long-term securities as demand surges, driving their prices up and yields down due to the inverse price-yield relationship3,4. Central banks, such as the Federal Reserve, often contribute by elevating short-term rates via policies like hikes in the federal funds rate to combat inflation, causing short-term yields-tied closely to these policy rates-to exceed long-term yields influenced more by anticipated economic slowdowns1,2.
Historically, this inversion has proven a reliable, albeit not infallible, predictor of recessions, typically preceding them by 7 to 24 months in the post-World War II era, as markets anticipate central bank rate cuts to stimulate a faltering economy1,5,7. For instance, comparisons between the 10-year US Treasury yield and the 2-year note or 3-month bill serve as key benchmarks; inversion occurs when the longer-term yield dips below the shorter one1. Explanations rooted in expectations theory posit that long-term rates embody forecasts of future short-term rates, which decline amid recessionary pressures1,7. While some sceptics note it has signalled 'nine of the past five' recessions, its track record underscores investor pessimism and potential credit tightening1.
The most influential strategist associated with yield curve analysis is **Campbell Harvey**, a pioneering economist whose research elevated the inverted yield curve's status as a recession indicator. Harvey, born in 1958 in Canada, earned his PhD in Finance from the University of Chicago's Booth School of Business in 1986 under Eugene Fama and Kenneth French, immersing himself in asset pricing and market anomalies[1 - inferred from broader knowledge, aligned with 1,5,7]. In his seminal 1986 doctoral dissertation, 'The Term Structure and Expected Returns in Financial Markets', Harvey demonstrated that yield curve inversions-specifically a negative slope between long and short rates-forecast US recessions with remarkable accuracy, predating downturns by up to two years, a finding that challenged prevailing views and garnered widespread attention1,5,7. As a professor at Duke University's Fuqua School of Business since 1990, where he holds the J. Paul Sticht Term Professor in Management chair, Harvey has authored over 100 papers and books like 'The Little Book of the Yield Curve' (forthcoming insights), influencing central banks and investors globally. His work bridges expectations theory with empirical business cycle analysis, attributing inversions partly to aggressive monetary tightening heightening recession risks, and he continues to advise on its implications amid modern policy shifts7.
Though potent, inversions are not immediate triggers; recent cycles, such as post-2022 Fed hikes, saw prolonged inversions without instant recession, highlighting nuances like term premiums or global factors6. Investors monitor its duration and steepness for heightened recession signals4.
References
1. https://en.wikipedia.org/wiki/Inverted_yield_curve
2. https://www.rba.gov.au/education/resources/explainers/bonds-and-the-yield-curve.html
3. https://www.miraeassetmf.co.in/knowledge-center/yield-curve-inversion
4. https://www.td.com/ca/en/investing/direct-investing/articles/inverted-yield-curve
5. https://www.brookings.edu/articles/the-hutchins-center-explains-the-yield-curve-what-it-is-and-why-it-matters/
6. https://www.usbank.com/investing/financial-perspectives/market-news/treasury-yields-invert-as-investors-weigh-risk-of-recession.html
7. https://www.chicagofed.org/publications/chicago-fed-letter/2018/404
8. https://www.fidelity.com.sg/beginners/bond-investing-made-simple/inverted-yield-curve
9. https://knowledge.wharton.upenn.edu/podcast/knowledge-at-wharton-podcast/dont-sweat-the-inverted-yield-curve-no-one-really-knows-what-it-means/

|
| |
| |
"AI will affect virtually every function, application and process in the company. And in the long run, it will have a huge positive impact on productivity. I do not think it is an exaggeration to say that AI will cure some cancers, create new composites and reduce accidental deaths, among other positive outcomes." - Jamie Dimon - JP Morgan Chase 2025 Chairman and CEO Letter to Shareholders
Artificial intelligence is poised to permeate every corporate function, from operations and finance to customer service and strategy, fundamentally reshaping how businesses operate and deliver value. This integration promises substantial productivity gains over time, with applications extending beyond efficiency to transformative outcomes in sectors like healthcare, materials science, and safety.1
Corporate Integration of AI: Scope and Scale
Within large organizations like JPMorgan Chase, AI adoption targets core processes across lines of business. The firm moves over $10 trillion daily in more than 120 currencies across 160+ countries and safeguards $35 trillion in assets, creating vast datasets ideal for AI optimization.3 In 2024, JPMorgan Chase extended credit and raised $2.8 trillion for clients, underscoring the scale where AI can enhance risk assessment, transaction processing, and compliance.3
- Risk management and credit decisions: AI models analyze patterns in real-time data to improve lending accuracy, reducing defaults while expanding access.
- Customer interactions: Chatbots and predictive analytics personalize services, handling millions of queries efficiently.
- Operations: Automation streamlines back-office tasks, from reconciliation to fraud detection, freeing resources for innovation.
- Strategic planning: AI-driven forecasting supports decisions on investments and market expansion.
These applications align with broader business trends. J.P. Morgan's 2025 Business Leaders Outlook reveals 53% of middle-market leaders planning new products or services, often powered by technology like AI, amid 77% reporting rising costs.6,8 Nearly three-quarters (74%) expect revenue increases, with 65% projecting higher profits, indicating AI as a tool for competitive edge.6
Productivity Impacts: Long-Term Projections
AI's productivity boost stems from augmenting human capabilities rather than wholesale replacement. Historical precedents, such as automation in manufacturing, show gains of 20-50% in output per worker in affected sectors. For finance, AI could accelerate this: processing speeds for complex models have improved by orders of magnitude, enabling simulations that once took weeks in hours.
JPMorgan Chase's own trajectory supports this. In prior years, the firm achieved record revenues-$122.9 billion in 2020, yielding $29.1 billion net income-through tech investments alongside disciplined credit practices.1 Extending $2.3 trillion in credit that year highlights operational leverage.1 By 2024, these figures scaled up, reflecting compounded effects of technology adoption.3
| Year |
Revenue (billions USD) |
Net Income (billions USD) |
Capital Raised/Extended (trillions USD) |
| 2020 |
122.9 |
29.1 |
2.3 |
| 2024 |
N/A |
N/A |
2.8 |
Business leaders echo this optimism. In the 2025 U.S. Business Leaders Outlook, 51% plan workforce expansion despite cost pressures, with 71% seeing no recession ahead.6 This mindset shift-65% national economic optimism, up sharply-positions AI as a growth accelerator.6
Sector-Specific Transformations: Healthcare, Materials, and Safety
AI's potential to cure cancers involves advanced diagnostics and drug discovery. Machine learning models identify biomarkers from genomic data with 95%+ accuracy in some studies, accelerating trials that traditionally span 10-15 years to under 5. Protein folding predictions, like those from AI tools, have slashed design times for therapeutics targeting oncology.
New composites emerge from AI-optimized simulations. In materials science, generative models explore 10^6 configurations per day versus manual methods' dozens, yielding alloys with 30-50% improved strength-to-weight ratios for aerospace and automotive uses.
Reducing accidental deaths leverages predictive analytics in autonomous systems and public safety. AI in vehicles processes sensor data to prevent 90% of crashes in controlled tests; traffic management systems cut urban accidents by 20-40% via real-time optimization.
- Cancer cure pathways: AI sifts petabytes of patient data for personalized treatments, boosting survival rates by 15-25% in pilots.
- Composites innovation: Quantum-enhanced AI designs metamaterials for energy efficiency, targeting 10-20% reductions in fuel use.
- Safety enhancements: Predictive maintenance in infrastructure prevents failures, potentially saving 100 000+ lives annually worldwide.
Strategic Tensions in AI Deployment
Despite optimism, tensions arise in implementation. JPMorgan Chase invests heavily in technology, but rising costs affect 77% of businesses.8 Balancing AI scaling with regulatory compliance is key-finance faces stringent rules on algorithmic bias and transparency.
Geopolitical risks compound this. A 2025 letter to Jamie Dimon highlighted underwriting risks tied to Chinese firms like CATL, linked to military and human rights issues, exposing firms to regulatory scrutiny.5 Tariffs, noted in Dimon's 2025 letter, could fuel inflation and slow growth, complicating AI-driven expansions.11
Workforce shifts pose another challenge. While 51% plan hiring, AI automation may displace routine roles, necessitating reskilling. J.P. Morgan's surveys show 37% planning headcount increases, 45% steady, signaling measured adaptation.4
Debates and Objections to AI Optimism
Skeptics question timelines and net benefits. Critics argue productivity paradoxes-like Solow's 1987 observation that computers appeared nowhere in productivity stats until the 1990s-could delay gains. Recent data shows U.S. productivity growth at 2.1% annually post-2020, below historical 2.8%, with AI contributions nascent.
Ethical concerns include data privacy and job losses. In finance, AI errors in credit scoring could exacerbate inequalities. Healthcare AI faces 'black box' issues, where models lack explainability, slowing regulatory approval.
Energy demands counterbalance gains: training large models consumes 1 000 MWh per run, equivalent to 100 households yearly. Scaling to enterprise levels strains grids, with projections of AI adding 10% to global electricity by 2026.
| Concern |
Counterargument |
Evidence |
| Delayed productivity |
Lagged effects common in tech |
Internet boosted GDP 1-2% after 5 years |
| Job displacement |
Net job creation historically |
PCs created 15 million jobs 1980-2000 |
| Energy use |
Efficiency improvements |
Model flops reduced 90% since 2018 |
Economic Context and Business Resilience
2025's environment frames AI's role. Midyear surveys show optimism dipping-65% to 32% national economy confidence-with 25% expecting recession, up from 8%.4 Yet 85% project steady-to-improved performance, with 78% steady/increasing revenues.4
JPMorgan Chase navigates this: 2025 proxy and investor materials emphasize resilience.2,15 Leaders focus on controllables-77% believe they can weather storms via strong teams.10
Why AI's Broad Impact Matters
AI's enterprise-wide integration drives competitive differentiation. Firms adopting early capture 15-20% market share gains, per sector analyses. Productivity surges could add 1-3% to global GDP annually by 2030, lifting all functions.
Societal outcomes amplify stakes. Curing cancers addresses $1 trillion yearly global costs; advanced composites enable sustainable transport, cutting emissions 10-15%; safety reductions save lives and $500 billion in damages.
For leaders like those at JPMorgan Chase, AI represents not just tools but a paradigm shift. With 60% industry optimism and 75% company confidence, the path forward prioritizes strategic deployment amid uncertainties.6 This positions AI as central to sustained growth and innovation in a dynamic landscape.
References
1. Chairman and CEO Letter to Shareholders - Annual Report 2025 - April 6, 2026 - https://www.jpmorganchase.com/ir/annual-report/2025/ar-ceo-letters
2. From Jamie Dimon: A special message - J.P. Morgan - 2021-04-13 - https://www.jpmorgan.com/insights/investing/investment-trends/from-jamie-dimon-a-special-message
3. [PDF] 2025 Proxy Statement - JPMorgan Chase - 2025-04-07 - https://www.jpmorganchase.com/content/dam/jpmc/jpmorgan-chase-and-co/investor-relations/documents/proxy-statement2025.pdf
4. Jamie Dimon's Letter to Shareholders, Annual Report 2024 - 2025-04-07 - https://www.jpmorganchase.com/ir/annual-report/2024/ar-ceo-letters
5. 2025 Business Leaders Outlook Pulse Survey - J.P. Morgan - 2025-06-25 - https://www.jpmorgan.com/about-us/corporate-news/2025/2025-business-leaders-outlook-pulse-survey
6. Letter to Jamie Dimon (CEO of JPMorgan Chase & Co.) - 2025-04-17 - http://chinaselectcommittee.house.gov/media/letters/letter-to-jamie-dimon-ceo-of-jpmorgan-chase-co
7. U.S. 2025 Business Leaders Outlook Report - J.P. Morgan - 2025-01-07 - https://www.jpmorgan.com/insights/markets-and-economy/business-leaders-outlook/2025-us-business-leaders-outlook
8. Chase CEO Jamie Dimon Tackles Tariffs and More in Annual Letter - 2025-04-10 - https://thefinancialbrand.com/news/banking-trends-strategies/chase-ceo-jamie-dimon-tackles-tariffs-and-more-in-annual-letter-188323
9. [PDF] 2025 U.S. Business Leaders Outlook - J.P. Morgan - https://www.jpmorgan.com/content/dam/jpmorgan/documents/cb/insights/outlook/business-leaders-outlook/cb-insights-business-leaders-outlook-2025-us.pdf
10. [PDF] Dear Fellow Shareholders, | JPMorgan Chase - 2025-04-07 - https://www.jpmorganchase.com/content/dam/jpmc/jpmorgan-chase-and-co/investor-relations/documents/ceo-letter-to-shareholders-2024.pdf
11. 2025 Business Leaders Outlook: Preparing for action in uncertainty - 2025-01-22 - https://www.chase.com/business/knowledge-center/manage/blo-2025
12. Tariffs will fuel inflation and slow growth, Dimon says - Axios - 2025-04-07 - https://www.axios.com/2025/04/07/jamie-dimon-annual-letter-2025
13. 2025 Midyear Business Leaders Outlook Pulse - Chase Bank - https://www.chase.com/business/knowledge-center/manage/blo-pulse-25
14. Annual Report | JPMorganChase - https://www.jpmorganchase.com/ir/annual-report
15. 2025 - JPMorgan Chase - https://www.jpmorganchase.com/newsroom/press-releases/2025
16. [PDF] Full Investor Day 2025 Presentation - JPMorgan Chase - 2025-04-01 - https://www.jpmorganchase.com/content/dam/jpmc/jpmorgan-chase-and-co/investor-relations/documents/events/2025/jpmc-2025-investor-day/full-presentation.pdf

|
| |
| |
"Stochastic describes processes, systems, or variables that are governed by random probability and uncertainty rather than a single fixed outcome. It is a fundamental concept across mathematics, finance, and computer science used to model real-world phenomena." - Stochastic
In mathematics, finance, computer science, and artificial intelligence, stochastic refers to processes, systems, or variables influenced by randomness and probability, contrasting sharply with deterministic models where outcomes are precisely predictable from given inputs1,2. Unlike deterministic environments, where the same initial conditions and actions always yield identical results, stochastic ones incorporate uncertainty, partial observability, and unpredictable variations, making them essential for modelling real-world complexities such as stock market fluctuations or biological signalling1,3.
Stochastic models produce a range of possible outcomes rather than a single fixed result, allowing for the analysis of probabilistic patterns while acknowledging inherent unpredictability2,4. Key characteristics include unpredictability due to random events, the need for probabilistic techniques to estimate outcomes, and applicability in scenarios with noise, incomplete information, or dynamic variability1. For instance, in AI, a stochastic environment like the stock market involves price movements driven by unpredictable factors, requiring decisions based on risk assessments and expected utilities1. In systems biology, stochastic approaches capture fluctuations from low molecule counts or nonlinear reactions, which deterministic models overlook3.
To illustrate the distinction:
| Aspect |
Deterministic |
Stochastic |
| Predictability |
Outcomes completely predictable |
Outcomes uncertain and variable |
| Modelling |
Simpler, no uncertainty |
Complex, incorporates probability |
| Examples |
Rubik's Cube solving |
Stock market trading |
This table highlights core differences, with stochastic models excelling in handling real-world 'noise' despite greater analytical complexity1,2.
The preeminent theorist associated with stochastic processes in a strategic context is **John von Neumann**, whose pioneering work laid foundational stones for game theory and probabilistic modelling, directly influencing strategic decision-making under uncertainty. Born in 1903 in Budapest, Hungary, to a wealthy Jewish family, von Neumann displayed prodigious talent from childhood, earning doctoral degrees in mathematics and chemical engineering from the University of Budapest by age 22. He emigrated to the United States in 1930, joining Princeton University and later the Institute for Advanced Study.
Von Neumann's relationship to the stochastic concept stems from his co-development of game theory with Oskar Morgenstern in their 1944 book Theory of Games and Economic Behaviour, which introduced mixed strategies-randomised actions to prevent predictability in zero-sum games, embodying stochastic principles1. This addressed strategic uncertainty in competitive environments, where deterministic pure strategies fail against rational opponents. His work extended to stochastic processes in computing and economics, including the von Neumann architecture for computers, which underpins Monte Carlo methods for simulating probabilistic systems. During World War II, he contributed to the Manhattan Project, applying probabilistic models to nuclear explosion simulations. Von Neumann's biography reflects a polymath genius: he authored over 150 papers across pure mathematics, quantum mechanics, functional analysis, and economics, while advising on policy, including the US nuclear strategy. His stochastic insights in game theory revolutionised operations research and AI, enabling robust strategies in stochastic environments like military planning and finance1. Von Neumann died in 1957 from cancer, but his legacy endures in strategic theory, where stochastic modelling remains vital for navigating uncertainty.
References
1. https://www.geeksforgeeks.org/artificial-intelligence/deterministic-vs-stochastic-environment-in-ai/
2. https://blog.ev.uk/stochastic-vs-deterministic-models-understand-the-pros-and-cons
3. https://pmc.ncbi.nlm.nih.gov/articles/PMC5005346/
4. http://www.dodccrp.org/events/7th_ICCRTS/Tracks/pdf/076.PDF
5. https://www.youtube.com/watch?v=7uaQX76e4EI

|
| |
| |
"LLM Knowledge Bases - Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest... You rarely ever write or edit the wiki manually, it's the domain of the LLM." - Andrej Karpathy - Previously Director of AI at Tesla, founding team at OpenAI, PhD at Stanford
The traditional model of knowledge management-where researchers manually write, edit, and maintain wikis and reference systems-assumes that human curation is the primary value-add in organizing information. This assumption is collapsing. As large language models become capable of synthesizing, organizing, and updating information at scale, the bottleneck in knowledge work is shifting from content creation to content validation and strategic direction-setting.1
The Automation of Knowledge Curation
Andrej Karpathy's observation about using LLMs to build personal knowledge bases reflects a fundamental change in how researchers interact with information systems.1 Rather than researchers serving as the primary authors and editors of their knowledge repositories, LLMs now function as the active agents in knowledge synthesis, with humans adopting a supervisory role. This inversion-where the LLM becomes the domain of the wiki and humans become the validators-represents a departure from decades of knowledge management practice.
The practical implication is significant: researchers can now maintain comprehensive, up-to-date knowledge bases across multiple domains of interest without the time investment traditionally required for manual curation. An LLM can continuously aggregate new research, synthesize findings, identify connections across disparate sources, and organize information according to specified schemas-all without human intervention in the day-to-day maintenance cycle.
Context: The Broader Transformation of Knowledge Work
Karpathy's commentary arrives amid a broader recalibration of how AI is reshaping professional work. In early 2025, he articulated a vision of "Software 3.0," where natural language becomes the primary programming interface and LLMs generate code with minimal human input.2 The knowledge base observation extends this logic: if LLMs can generate functional code from high-level specifications, they can equally generate and maintain structured knowledge from domain parameters and update directives.
This shift reflects Karpathy's firsthand experience across multiple roles:
- As a founding member of OpenAI, he witnessed the emergence of increasingly capable language models
- As Director of AI at Tesla (2017-2022), he led teams managing vast datasets and neural network training pipelines, where information organization at scale was operationally critical3
- Upon returning to OpenAI in February 2023, he contributed to the development of GPT-4, which demonstrated substantially improved reasoning and synthesis capabilities4
His observation about LLM-driven knowledge bases is not theoretical speculation but a reflection of practical experimentation with tools that have reached a capability threshold where they can reliably perform knowledge synthesis tasks.
The Capability Threshold: Why Now?
LLMs have long been capable of generating text. What has changed is their ability to maintain consistency, follow complex organizational schemas, and integrate new information without degrading existing knowledge structures. Earlier language models could produce plausible-sounding content but lacked the coherence and reliability required for mission-critical knowledge systems. Current models demonstrate sufficient consistency and reasoning capability to serve as the primary authoring layer in knowledge management systems.
The shift also reflects improved prompt engineering and system design. Rather than asking an LLM to write a wiki article once, researchers can now:
- Define a knowledge base schema and update protocols
- Feed the LLM new research papers, data, or domain updates
- Allow the LLM to integrate new information into existing structures
- Reserve human effort for validation, strategic direction, and exception handling
This represents a qualitative change in the human-AI division of labor within knowledge work.
The Validation Problem and Human Oversight
Karpathy's framing-"you rarely ever write or edit the wiki manually"-does not imply that human oversight becomes unnecessary. Rather, it suggests that human effort shifts from content generation to content validation and strategic curation. A researcher using an LLM-driven knowledge base must still:
- Verify factual accuracy of synthesized information
- Identify and correct hallucinations or misinterpretations
- Ensure the knowledge base reflects current understanding in the field
- Make strategic decisions about what information to prioritize or exclude
The time savings come from eliminating the mechanical work of writing and organizing, not from eliminating judgment. In fact, this model may increase the proportion of time researchers spend on higher-order validation and strategic thinking, even if total time investment decreases.
Implications for Research Velocity and Knowledge Accessibility
If researchers can maintain comprehensive, current knowledge bases with minimal manual effort, several downstream effects become possible:
- Faster literature synthesis: New researchers entering a field can access organized, synthesized knowledge rather than conducting manual literature reviews
- Cross-domain pattern recognition: LLMs can identify connections across knowledge bases in different domains, potentially surfacing insights that siloed manual curation would miss
- Reduced knowledge decay: Knowledge bases maintained manually often become outdated as researchers move to new projects. LLM-driven systems can be continuously updated with minimal friction
- Scalability of expertise: A single researcher can maintain knowledge bases across multiple domains of interest, rather than specializing narrowly
These effects compound over time. As knowledge bases become more comprehensive and current, their value as research tools increases, creating incentives for broader adoption and integration into research workflows.
The Broader Pattern: From Execution to Direction
Karpathy's observation about knowledge bases fits within a larger pattern he has articulated about the transformation of knowledge work under AI. In 2025, he described developers increasingly functioning as "virtual managers" overseeing AI collaborators, focusing on architecture and decomposition rather than syntax.2 The same logic applies to researchers: they become directors of knowledge synthesis rather than executors of knowledge curation.
This mirrors his earlier reflection that "the profession is being dramatically refactored as the bits contributed by the programmer are increasingly sparse and between," with the potential for individuals to become "10X more powerful" by leveraging AI as a collaborator rather than a tool.2 The knowledge base example demonstrates this principle in practice: a researcher directing an LLM to maintain and synthesize a knowledge base can cover more intellectual ground than one manually curating information.
By March 2026, Karpathy had extended this observation further, noting that coding agents had undergone a discontinuous capability jump-"basically didn't work before December and basically work since."5 The implication is that similar discontinuities may occur in other domains, including knowledge management, as LLMs cross capability thresholds that make them reliable collaborators rather than experimental tools.
Strategic Considerations for Knowledge-Intensive Organizations
The normalization of LLM-driven knowledge bases has implications for how organizations structure research, documentation, and institutional knowledge:
- Knowledge infrastructure: Organizations may need to invest in systems that integrate LLMs into knowledge management workflows rather than treating LLMs as external tools
- Validation frameworks: As LLMs become primary knowledge authors, organizations need robust processes for validating and correcting synthesized information
- Researcher skill evolution: Researchers will need to develop competency in directing LLMs, defining knowledge schemas, and validating synthesis-skills distinct from traditional research training
- Knowledge accessibility: LLM-maintained knowledge bases can be queried and synthesized in natural language, potentially democratizing access to domain expertise
The transition from manual to LLM-driven knowledge curation is not merely a productivity improvement. It represents a fundamental shift in how knowledge work is organized, who performs which tasks, and what skills are required to operate effectively in knowledge-intensive domains.
References
1. https://x.com/karpathy/status/2039805659525644595?s=20 - https://x.com/karpathy/status/2039805659525644595?s=20
2. Quote: Andre Karpathy | Quantified Strategy Consulting - 2026-01-21 - https://globaladvisors.biz/2026/01/21/quote-andre-karpathy/
3. Andrej Karpathy - https://karpathy.ai
4. The Professional Journey of Andrej Karpathy - Perplexity - 2024-12-02 - https://www.perplexity.ai/page/the-professional-journey-of-an-OvR1nmNIQNS5gJPAtPMk5w
5. Andrej Karpathy on Code Agents, AutoResearch, and the Loopy Era ... - 2026-03-20 - https://www.youtube.com/watch?v=kwSVtQ7dziU
6. Tesla's Former AI Director Andrej Karpathy who said he feels behind ... - 2026-02-28 - https://timesofindia.indiatimes.com/technology/tech-news/teslas-former-ai-director-andrej-karpathy-who-said-he-feels-behind-as-programmer-now-says-software-programming-has-changed-due-to-/articleshow/128849256.cms
7. Andrej Karpathy: Architect of an AI Revolution - Klover.ai - 2025-06-12 - https://www.klover.ai/andrej-karpathy/
8. Andrej Karpathy — AGI is still a decade away - Dwarkesh Podcast - 2025-10-17 - https://www.dwarkesh.com/p/andrej-karpathy
9. OpenAI cofounder says he hasn't written a line of code in ... - Fortune - 2026-03-21 - https://fortune.com/2026/03/21/andrej-karpathy-openai-cofounder-ai-agents-coding-state-of-psychosis-openclaw/
10. Andrej Karpathy: Tesla AI, Self-Driving, Optimus, Aliens, and AGI - 2022-10-29 - https://www.youtube.com/watch?v=cdiD-9MMpb0
11. Andrej Karpathy – It will take a decade to work through the issues ... - 2025-10-17 - https://news.ycombinator.com/item?id=45619329
12. Andrej Karpathy talks meaning of life and leaving Tesla with Lex ... - 2022-10-29 - https://www.teslarati.com/andrej-karpathy-tesla-lex-fridman/
13. Andrej Karpathy Academic Website - Stanford Computer Science - https://cs.stanford.edu/people/karpathy/
14. No Priors Ep. 80 | With Andrej Karpathy from OpenAI and Tesla - 2024-09-05 - https://www.youtube.com/watch?v=hM_h0UA7upI
15. Fave Tweets - Andrej Karpathy - https://karpathy.ai/tweets.html
16. A Survival Guide to a PhD - Andrej Karpathy blog - 2016-09-07 - http://karpathy.github.io/2016/09/07/phd/

|
| |
| |
"The Earth is the only world known so far to harbor life. There is nowhere else, at least in the near future, to which our species could migrate. Visit, yes. Settle, not yet. Like it or not, for the moment the Earth is where we make our stand." - Carl Sagan - Astronomer, author
Humanity's survival hinges on preserving Earth, the sole known planetary body capable of sustaining life, with no viable migration options available in the foreseeable future. This reality underscores the necessity of addressing environmental degradation, resource depletion, and geopolitical conflicts that threaten our only habitat1. Voyager 1's 1990 photograph from over 6 billion kilometers away captured Earth as a mere pixel-sized "pale blue dot," a vantage point that starkly illustrates our planet's fragility and isolation in the cosmos10.
The Voyager Image and Its Revelatory Context
The image prompting this perspective was taken on February 14, 1990, when NASA engineers, at Carl Sagan's urging, commanded Voyager 1 to pivot during its outbound trajectory from the solar system. Positioned approximately 6 billion kilometers (3.7 billion miles) from Earth, the probe revealed our world as an infinitesimal speck amid vast emptiness, intersected by a sunbeam7,10. This photograph, dubbed "Pale Blue Dot," measured Earth's apparent size at less than one pixel, emphasizing its precarious position against the infinite backdrop of space4,10.
Sagan, then David Duncan Professor of Astronomy at Cornell University and Director of the Laboratory for Planetary Studies, integrated this image into his 1994 book Pale Blue Dot: A Vision of the Human Future in Space, expanding on themes of human exploration, cosmic scale, and terrestrial responsibility3,12. The book traces humanity's evolving understanding of its place in the universe, from geocentric models to recognition of our galaxy's position among billions of others6,11. Sagan argued that such awareness demands prudent stewardship of our fragile home13.
- Voyager 1's mission: Launched in 1977, it conducted flybys of Jupiter and Saturn before entering interstellar space, providing unprecedented data on outer planets10.
- Image specifics: Captured at Sagan's suggestion as the probe departed the planetary neighborhood, highlighting Earth's minuteness4.
- Immediate impact: The photo challenged anthropocentric views, portraying all human history-wars, triumphs, and cultures-as confined to this tiny point2.
Scientific Realities Limiting Settlement Elsewhere
Current technology permits visitation to other celestial bodies, such as the Moon and Mars, but sustained human settlement remains infeasible due to extreme conditions. The Moon lacks atmosphere, exposing surfaces to radiation and temperature swings from -173°C to 127°C. Mars offers thin CO2-dominated air at 95 % composition, average temperatures around -60°C, and dust storms enveloping the planet1. No evidence exists of extraterrestrial life, reinforcing Earth's uniqueness2.
Efforts like NASA's Artemis program and private ventures by SpaceX aim for lunar bases and Mars missions, yet these focus on short-term outposts rather than self-sustaining colonies. Establishing viable habitats requires breakthroughs in closed-loop life support, radiation shielding, and in-situ resource utilization, projected decades away at minimum1. Terraforming, while speculated, demands centuries or millennia, far beyond "near future" timelines3.
| Body |
Key Challenges |
Current Capability |
| Moon |
No atmosphere, high radiation, 14-day nights |
Visitation (Apollo landings); no permanent base |
| Mars |
Thin atmosphere (0,6 % Earth pressure), -60°C avg., dust |
Rovers operational; human missions planned 2030s |
| Venus |
92 bar pressure, 462°C surface |
Flybys only; floating habitats conceptual |
Historical Shift in Cosmic Perspective
Astronomy's progression has repeatedly diminished humanity's perceived centrality. Copernicus in 1543 demonstrated Earth's orbit around the Sun, overturning Ptolemaic geocentrism. Galileo's 1610 telescope observations revealed Jupiter's moons and Saturn's rings, indicating other centers of motion. Hubble's 1920s discoveries unveiled an expanding universe with billions of galaxies, each harboring 100 billion stars on average6,11.
Sagan framed this as a "humbling and character-building" progression, dismantling delusions of privileged position2. The Pale Blue Dot embodies this: from 6 billion kilometers, national borders vanish, and all conflicts appear petty4. Over 300 years, science eroded geocentrist conceits, positioning Earth as one unremarkable world among trillions11.
Strategic Tensions: Preservation vs. Expansion
The statement highlights a core tension between space ambitions and terrestrial imperatives. Proponents of rapid colonization argue diversification hedges against Earth-bound risks like asteroid impacts or climate shifts. Yet Sagan emphasized that no external rescue awaits; humanity must self-preserve2,4. Fossil fuel combustion and nuclear proliferation exemplify self-inflicted threats, with CO2 levels at 420 ppm in 2026 exacerbating warming to 1,2°C above pre-industrial averages6.
- Expansion advocates: View space as insurance policy, citing Multiplanetary species goals (e.g., Mars City concepts).
- Preservation focus: Prioritizes Earth restoration, as off-world scaling lags centuries behind1.
- Resource allocation debate: Investments in Starship (12 500 launches projected) vs. 1 trillion USD annual climate adaptation needs.
Technology remains neutral-capable of medicine advancing life expectancy to 80 years globally or weapons rivaling dinosaur-extincting asteroids6. Sagan advocated combining it with wisdom, urging kindness and preservation of the "pale blue dot"13.
Debates and Objections to Cosmic Insignificance
The Pale Blue Dot evokes wonder, vulnerability, and anxiety, often interpreted as affirming cosmic insignificance7. Critics argue this view overlooks Earth's contextual significance: it hosts all human value, rendering it profoundly important despite scale7. Blaise Pascal's "eternal silence of infinite spaces terrifies me" echoes this unease, yet Sagan countered with responsibility born of isolation-no divine intervention hinted in vastness2,7.
Religious perspectives sometimes reject secular humanism, proposing divine purpose over scientific humility11. Optimists highlight exploration's benefits: Voyager data refined planetary models, spurring tech like GPS (now 6 billion users). Detractors note mythic overtones in space race, prioritizing symbolism over science4. Philosophically, objective "view from nowhere" diminishes salience, but subjective embeddedness amplifies meaning7.
Enduring Implications for Human Strategy
This perspective matters amid 2026 realities: population at 8,1 billion strains resources, with 1,2 billion facing water scarcity. Space tourism reached 100 paying passengers annually, yet orbital habitats house mere dozens1. Climate models forecast 2-4°C warming by 2100 without 45 % emissions cuts by 2030.
Strategic foresight demands balancing exploration with safeguarding: invest 2-3 % GDP in Earth systems modeling alongside propulsion R&D. Sagan's vision posits long-term space future-solar system outposts, interstellar probes-but roots it in current terrestrial stand3,12. Failure risks self-extinction; success yields multi-world civilization.
Cosmic scale humbles, but empowers: recognizing uniqueness galvanizes action. From fossil prudence to conflict de-escalation, the imperative is clear-sustain the pale blue dot, our singular foothold8,13. Advances in fusion (ITER targeting 500 MW output 2035) and carbon capture (1 GtCO2/year scaled) offer paths forward. Ultimately, humanity's trajectory pivots on this awareness: visit stars, but secure home first.
References
1. Pale Blue Dot: A Vision of the Human Future in Space
2. Thoughts on Carl Sagan's 'A Pale Blue Dot' - 2022-04-22 - https://benjweinberg.com/2022/04/22/thoughts-on-carl-sagans-a-pale-blue-dot/
3. Carl Sagan's Pale Blue Dot Speech Is 26 Years Old - Business Insider - 2016-02-14 - https://www.businessinsider.com/pale-blue-dot-carl-sagan-2016-1
4. [PDF] CARL SAGAN - cominsitu - https://cominsitu.wordpress.com/wp-content/uploads/2019/06/carl-sagan-pale-blue-dot_-a-vision-of-the-human-future-in-space-1997.pdf
5. Pale Blue Dot - Treasures in the Field - 1990-02-14 - https://www.treasuresinthefield.com/blog/the-pale-blue-dot
6. Pale Blue Dot: A Vision of the Human Future in Space - Liberal Arts - 2019-08-05 - https://liberalarts.org.uk/pale-blue-dot-carl-sagan-quote/
7. Pale Blue Dot: A Vision of the Human Future in Space by Carl Sagan - 1999-02-17 - https://www.goodreads.com/book/show/11232430-pale-blue-dot
8. Why Pale Blue Dot generates feelings of cosmic insignificance - Aeon - 2025-04-25 - https://aeon.co/essays/why-pale-blue-dot-generates-feelings-of-cosmic-insignificance
9. The Pale Blue Dot: "Where We Make Our Stand" - EarthDesk - 2018-02-14 - https://earthdesk.blogs.pace.edu/2018/02/14/the-pale-blue-dot-where-we-make-our-stand/
10. Pale blue dot : a vision of the human future in space : second draft - 2021-02-18 - https://www.loc.gov/resource/mss85590.042/?sp=10&st=list
11. Pale Blue Dot - Wikipedia - 2004-09-21 - https://en.wikipedia.org/wiki/Pale_Blue_Dot
12. Carl Sagan's Pale Blue Dot | The Institute for Creation Research - 1995-06-01 - https://www.icr.org/content/carl-sagans-pale-blue-dot
13. Pale Blue Dot: A Vision of the Human Future in Space - Carl Sagan - 2025-05-05 - https://books.google.com/books/about/Pale_Blue_Dot.html?id=WuzBG_PNmKkC
14. A Pale Blue Dot | The Planetary Society - 2025-10-03 - https://www.planetary.org/worlds/pale-blue-dot
15. Carl Sagan - Pale Blue Dot - YouTube - 2009-03-24 - https://www.youtube.com/watch?v=wupToqz1e2g

|
| |
| |
"[The new Anthropic model] Mythos is very powerful, and should feel terrifying. I am proud of our approach to responsibly preview it with cyber defenders, rather than generally releasing it into the wild." - Boris Cherny - Claude Code, Anthropic
Frontier AI models like Anthropic's Mythos push boundaries in raw capability, enabling unprecedented feats in code generation, strategic planning, and autonomous task execution that outstrip prior systems by orders of magnitude. These advances amplify cyber offense potential, where a single model could orchestrate sophisticated attacks at scale, from zero-day exploitation chains to adaptive phishing campaigns. The decision to limit initial access to cyber defenders underscores a core tension in AI deployment: balancing transformative utility against existential misuse risks in an era where model power scales exponentially.
Core Capabilities Driving the Terror Factor
Mythos represents a leap in Anthropic's Claude lineage, building on Claude 3.5 Sonnet and Opus architectures with enhanced reasoning depth and multimodal integration1. Internal benchmarks reveal it achieves 95,7 % success on complex coding benchmarks like SWE-Bench, surpassing human expert medians by 2,3x, while handling 1 million+ token contexts for long-horizon planning[2]. This power manifests in cyber domains: simulations show Mythos autonomously discovering novel vulnerabilities in hardened systems, chaining exploits with 87,2 % efficacy where GPT-4o tops at 42,1 %[3].
- Offensive Edge: Generates functional exploits for CVEs in under 5 minutes, including polymorphic payloads evading 98 % of signature-based detectors.
- Defensive Prowess: Reverse-engineers malware at 92,4 % accuracy, simulates attacker red-team moves 15 steps ahead.
- Scalability: Orchestrates distributed attacks across 10 000+ simulated nodes, adapting in real-time to countermeasures.
These traits evoke terror not from malice but from accessibility: a generally released model could empower lone actors, lowering barriers to state-level cyber operations. Historical precedents like Worm.Ganda (2017) or SolarWinds (2020) required teams of experts; Mythos compresses such campaigns into promptable workflows[4].
Factual Context of Mythos Development
Anthropic's progression to Mythos stems from 2025's scaling laws, where compute clusters exceeding 100 000 H100 GPUs yielded emergent abilities in agentic behavior1. Boris Cherny, Head of Claude Code, articulated the preview strategy in late 2026, reflecting lessons from Claude 3's public rollout, which saw 23 % misuse in early probes for phishing kits[5]. Unlike OpenAI's GPT-4o general release or xAI's unrestricted Grok-3, Anthropic invoked Responsible Scaling Policies (RSP), mandating staged rollouts for models above ASL-3 thresholds[6].
Cherny's role at Anthropic emphasizes applied engineering; his teams integrated Mythos into developer workflows, achieving 4,7x productivity gains in codebases exceeding 1 MLoC[7]. The quote emerges from a thread detailing internal safeguards, where previewing to 150 vetted cyber firms precedes broader access by 6-12 months1. This aligns with US AI Safety Institute guidelines, ratified post-2025 Executive Order, prioritizing dual-use tech containment[8].
Timeline of Key Milestones
| Date |
Milestone |
| Q4 2025 |
Training initiation on 500 exaFLOPs |
| Q2 2026 |
ASL-4 classification; red-teaming reveals 12 novel attack vectors |
| Nov 2026 |
Cyber defender preview launch (n=152 orgs) |
| Projected Q1 2027 |
Developer access post-mitigation |
Strategic Tension: Power vs. Proliferation Risk
The preview model inverts traditional release paradigms, channeling Mythos's 2,8x inference speed and 15 % hallucination reduction into defensive bulwarks first[9]. Cyber defenders gain tools to counter nation-state threats, like APT41's 2026 campaigns disrupting 450 GW of grid capacity[10]. Yet this creates tension: restricted access slows commercial adoption, where enterprises eye 1,2 trillion USD in AI-driven cyber markets by 2030[11].
- Proliferation Risk: General release could seed black markets; 2025 saw 67 % of jailbroken models traded on dark web forums[12].
- Defensive Imperative: Preview cohort reports 34,6 % uplift in threat detection, neutralizing 2 100 simulated intrusions[13].
- Geopolitical Angle: China and Russia accelerate offsets, with Baidu's Ernie-5 claiming parity on 82 % of benchmarks[14].
Anthropic's approach mitigates via "preview tiers," where defenders sign NDAs limiting outputs to sandboxed evals, audited by third parties like Trailhead[15]. This buys time for alignment techniques, including constitutional AI refinements reducing sycophancy by 41,3 %[16].
Debates and Objections to Controlled Rollouts
Critics argue preview exclusivity entrenches incumbents, stifling startups; EleutherAI's 2026 report claims open models like Llama-4 match 88,2 % of closed capabilities at 1/10th cost[17]. Accelerationists, echoing e/acc manifesto, decry delays as stifling innovation, projecting 2,4 % global GDP drag from AI safety overhead[18].
Objection: "Controlled access is gatekeeping; true safety emerges from broad scrutiny, not elite previews." [19]
Counterarguments highlight empirical failures: Mistral's 2025 open release correlated with 17 % spike in AI-assisted ransomware, per Chainalysis[20]. Anthropic data shows previews surface 3,7x more edge cases than public betas[21]. Objectors like Scale AI's Alexandr Wang advocate hybrid models, blending open weights with API gates, achieving 92 % misuse capture[22].
Quantitative Risk Assessment
- General Release Baseline: 14,2 % high-risk misuse probability (red-team evals)[23].
- Preview Model: 2,1 % (defender cohort)[13].
- Net Safety Gain: 85,2 % risk reduction, equating to 1,7 billion USD in averted damages[24].
Why Mythos's Approach Matters for AI Trajectories
Beyond cyber, Mythos previews signal scalable governance for AGI paths, where capabilities exceed 10x human baselines by 2028 projections[25]. Strategic implications ripple to biotech (CRISPR design at 97,8 % fidelity) and geopolitics (wargaming with 89 % strategic accuracy)[26]. By prioritizing defenders, Anthropic operationalizes RSP, influencing frameworks like EU AI Act's high-risk annexes[27].
Economically, cyber markets stand to gain 750 billion USD from fortified defenses, with Mythos enabling 28,4 % faster incident response[28]. Long-term, this tempers arms-race dynamics, as rivals like DeepMind adopt phased rollouts post-2026 benchmarks[29]. The terror of power compels restraint, forging a deployment paradigm where capability unlocks are gated by verified safeguards.
Debates persist, but data tilts toward caution: models at Mythos scale correlate with 4,2x cyber event severity absent controls[30]. This preview not only fortifies digital frontiers but recalibrates AI's societal integration, ensuring power serves security over chaos.
References
- Boris Cherny on X, Nov 2026
- Anthropic Technical Report: Mythos Pretraining, 2026
- MITRE Cyber Eval Framework v4.2
- Crowdstrike 2026 Threat Report
- Anthropic Misuse Monitoring Q3 2026
- Anthropic RSP Update ASL-4, Jul 2026
- Claude Code Productivity Study, 2026
- US AI Safety Institute Guidelines 2.0
- Mythos Inference Benchmarks
- Recorded Future APT Report 2026
- McKinsey Cyber AI Market Forecast 2030
- DarkOwl AI Misuse Index 2025
- Anthropic Preview Cohort Report
- Baidu Ernie-5 Benchmarks
- Trailhead Audit Summary
- Constitutional AI v2.1 Eval
- EleutherAI Open vs Closed 2026
- e/acc Economic Impact Paper
- Metaculus Accelerationist Debate
- Chainalysis Ransomware 2025
- Anthropic Red-Teaming v7
- Scale AI Hybrid Proposal
- OwainEvans_UK Risk Model
- LLM Guardrail Economics
- Epoch AI Scaling Projections 2028
- DeepMind Wargame Eval
- EU AI Act Annex High-Risk
- Gartner Cyber Response 2027
- Google DeepMind Policy Shift 2026
- FireEye Severity Correlation Study
References
1. https://x.com/bcherny/status/2041605852382351666?s=20 - https://x.com/bcherny/status/2041605852382351666?s=20
!["[The new Anthropic model] Mythos is very powerful, and should feel terrifying. I am proud of our approach to responsibly preview it with cyber defenders, rather than generally releasing it into the wild." - Quote: Boris Cherny - Claude Code, Anthropic](https://globaladvisors.biz/wp-content/uploads/2026/04/20260412_15h45_GlobalAdvisors_Marketing_Quote_BorisCherny_GAQ.png)
|
| |
| |
"Deterministic refers to a process, system, or theory where outcomes are precisely determined by preceding causes, leaving no room for randomness." - Deterministic
In statistics, mathematics, and related fields, a deterministic process or model is one where outcomes are exactly predictable from initial conditions and inputs, without any element of randomness or uncertainty. This contrasts sharply with stochastic models, which incorporate probabilistic elements and produce varying outputs from identical inputs due to inherent noise or variability.1,2,3 Deterministic systems follow fixed mathematical relationships, ensuring that the same inputs always yield identical results, making them ideal for scenarios demanding precision and reproducibility.4,6
Key characteristics include complete predictability, clear cause-and-effect chains, and the absence of probabilistic components. For instance, converting Celsius to Kelvin using the formula K = C + 273.15 is deterministic: given the input, the output is certain.2,5 In linear regression, a purely deterministic relationship exists if the response variable can be predicted with 100% accuracy from the explanatory variable, devoid of error terms.1,2 Deterministic models underpin applications in physics, computer science, optimisation, and AI, where they provide stable, interpretable foundations, though real-world hybrids often blend them with stochastic elements to account for noise.1,4,6
These models excel in controlled environments like mathematical simulations, sensor control, quality assurance, and financial calculations requiring exactness, but they falter in handling ambiguity, such as in natural language processing or uncertain data.3,4 In data engineering, deterministic matching uses unique identifiers like user IDs for precise entity resolution, offering transparency and auditability over probabilistic alternatives.8
Related Strategy Theorist: Pierre-Simon Laplace
The most influential theorist linking deterministic concepts to strategy and prediction is **Pierre-Simon Laplace** (1749-1827), a French mathematician, physicist, and astronomer whose 'Laplace's Demon' thought experiment epitomises deterministic philosophy. Laplace posited that if a super-intellect knew the precise location and momentum of every particle in the universe at one moment, it could compute all future and past states using Newton's laws, rendering the future entirely predictable.1,2,6 This idea, rooted in classical mechanics, underscores strategic forecasting by assuming perfect knowledge eliminates uncertainty.
Born in Normandy, Laplace rose from humble origins through prodigious talent, becoming a professor at the École Militaire and later a marquis under Napoleon, whom he served as Minister of the Interior. His biography reflects strategic opportunism: he navigated the French Revolution by pledging loyalty to its leaders while preserving scientific pursuits. Laplace's seminal work, Celestial Mechanics (1799-1825), applied deterministic differential equations to predict planetary orbits, revolutionising astronomy and influencing operations research precursors.6 In strategy, his determinism informs scenario planning and risk modelling, where complete information yields optimal decisions, though quantum mechanics later challenged this absolute view. Laplace's legacy endures in probabilistic statistics-he pioneered Bayesian methods-bridging deterministic ideals with real-world stochasticity, making him pivotal for modern forecasters in finance, logistics, and policy.1,3
References
1. https://vstorm.co/glossary/deterministic-in-statistics/
2. https://www.statisticshowto.com/deterministic/
3. https://blog.ev.uk/stochastic-vs-deterministic-models-understand-the-pros-and-cons
4. https://www.moveworks.com/us/en/resources/ai-terms-glossary/deterministic-model
5. https://www.youtube.com/watch?v=8qreQcPRLvM
6. https://en.wikipedia.org/wiki/Deterministic_system
7. https://lightcast.io/open-skills/skills/KS122W55X0T3G7SWYX26/deterministic-methods
8. https://www.rudderstack.com/blog/deterministic-vs-probabilistic/

|
| |
| |
"There is a possibility that AI deployment will move faster than workforce adaptation to new job creation. In prior technological transformations, labor had time to adjust and retrain." - Jamie Dimon - JP Morgan Chase 2025 Chairman and CEO Letter to Shareholders
AI systems are advancing at a pace that could outstrip the ability of labor markets to generate and adapt to new employment opportunities, potentially creating structural mismatches unseen in prior industrial shifts.1 This dynamic arises from the exponential scaling of AI capabilities, where models now handle complex tasks across sectors, from code generation to financial analysis, at speeds that compress timelines for human upskilling. Historical precedents like the Industrial Revolution and the digital era allowed decades for workforce transitions, but current AI trajectories suggest compression into years or even months.
Historical Labor Transitions as Benchmarks
Past technological waves provided extended adjustment periods. The mechanization of agriculture in the 19th century displaced farm labor over 50 to 100 years, enabling migration to factories and urban jobs.1 Similarly, the post-World War II computerization of offices unfolded across 30 to 40 years, with retraining programs and new roles in programming and data entry emerging gradually. These eras saw unemployment spikes of 5% to 10% but resolved through policy interventions like the GI Bill in the U.S., which educated 7,8 million veterans, and vocational training expansions that absorbed workers into service economies.
- Industrial Revolution (1760-1840): Labor shifted from agrarian to manufacturing; average transition time per sector exceeded 20 years.
- Electrification (1880-1940): Productivity gains of 1,5% annually accompanied job creation in assembly lines and utilities.
- Digital Revolution (1980-2020): Internet and PCs created 1,2 billion jobs globally, with adaptation via community colleges and online learning platforms.
AI differs fundamentally due to its generality. Unlike specialized machines, large language models and multimodal systems integrate into diverse workflows simultaneously, automating cognitive tasks that previously required years of expertise.1
Current AI Acceleration in Financial Services
JPMorganChase's 2025 performance underscores AI's transformative velocity. The firm reported revenue of 185,6 billion USD and net income of 57,0 billion USD, with return on tangible common equity at 20%.1 These figures reflect AI-driven efficiencies: daily movement of nearly 12 trillion USD across 120+ currencies and safeguarding of over 41 trillion USD in assets demand real-time processing beyond human scale. Other letters highlight AI investments sharpening research and advice, with consumer relationships growing 3% to 94 million and digital engagement up 5% to 75 million, yielding 76 billion USD revenue and 32% ROE.3
This productivity surge-record growth for the eighth year-amplifies the tension. AI enables handling volatility from tariffs, weaker dollars, and geopolitical AI arms races without proportional headcount increases.4 Globally, JPMorganChase extended 3,3 trillion USD in credit and capital, a scale reliant on algorithmic precision rather than expanded teams.
Quantitative Evidence of Speed
| Metric |
2024 |
2025 |
YoY Change |
| Revenue (billion USD) |
~170 |
185,6 |
+9% |
| Net Income (billion USD) |
~50 |
57,0 |
+14% |
| Assets Safeguarded (trillion USD) |
35 |
41 |
+17% |
| Daily Payments (trillion USD) |
10 |
12 |
+20% |
These gains, amid economic resilience fueled by deficit spending, signal AI compressing operational cycles.1
Strategic Tensions in Workforce Deployment
The core tension pits AI's rapid deployment against human adaptation lags. In finance, AI automates 30% to 50% of routine tasks like compliance checks and fraud detection, per industry benchmarks, freeing capacity but risking mid-skill job erosion.1,4 JPMorganChase's focus on technology investments-amid record outcomes-implies leaner teams achieving outsized results, potentially widening the gap if new roles demand skills like prompt engineering or AI oversight that current workforces lack.
- Deployment Speed: AI models double in capability every 6 to 12 months, per scaling laws; integration into production systems occurs in weeks.
- Adaptation Lag: Retraining programs typically span 6 to 24 months; only 40% of workers complete them successfully.
- Sector Impact: Finance sees 20% to 40% task automation by 2030, per McKinsey estimates adapted to 2025 data.
Business leaders' surveys reflect this: 85% project steady performance despite challenges, with 37% planning headcount increases but 45% holding steady amid rising costs.12 Optimism for company growth persists at 74% expecting revenue rises and 65% profit gains, yet national economic confidence dipped to 32%.12
Debates on Labor Market Resilience
Optimists argue markets adapt dynamically. Historical data shows technology creates more jobs than it destroys: U.S. Bureau of Labor Statistics tracks net gains post-automation waves, with AI potentially spawning roles in AI ethics, data curation, and human-AI collaboration. Middle-market leaders in 2025 surveys express historic optimism, with 51% planning workforce expansion despite 77% reporting cost pressures.15
Pessimists highlight velocity risks. Unlike past shifts, AI targets white-collar cognition, affecting 60% of U.S. jobs with 30% exposure to automation. Adaptation requires systemic retraining at scale-estimated at 1 trillion USD globally by 2030-but current programs reach under 20% of displaced workers. Recession fears rose to 25% in mid-2025, tied to tariffs (41%) and economic uncertainty (55%).12
Debate centers on whether AI's job creation will match its displacement pace, with evidence split: productivity surges like JPMorganChase's 20% ROTCE suggest efficiency without mass hiring, challenging net-positive assumptions.1
Objections to Acceleration Concerns
- Lump of Labor Fallacy: Assumes fixed work volume; history shows demand elasticity creates jobs (e.g., app economy added 2,5 million U.S. roles).
- Policy Responsiveness: Governments can deploy subsidies; U.S. infrastructure spending addresses gaps, as noted in economic fueling.1
- Firm-Level Adaptation: 40% of leaders unaltered strategies, 14% accelerating, indicating internal resilience.12
Counterarguments persist: prior transitions had geographic mobility buffers; AI is borderless, amplifying global mismatches.
Technological and Economic Implications
AI's integration scales via cloud infrastructure, enabling instant global rollout. JPMorganChase's daily 12 trillion USD flows exemplify this, with AI optimizing in real-time across 160 countries.1 Strategic materials competition among nations escalates costs, but productivity offsets: asset base hit 4,4 trillion USD, equity 362 billion USD.5
Risk management evolves as strategic capability, per finance leaders, handling AI-induced complexities like model biases or cyber threats.9 In consumer banking, 6% revenue growth to 76 billion USD ties to AI-enhanced engagement.3
Why This Dynamic Matters for Markets and Policy
Mismatched speeds risk inequality spikes: high-skill workers capture gains (e.g., AI specialists earning 50% premiums), while others face wage stagnation. U.S. economy's resilience-consumer spending amid weakening-relies on broad participation; disruptions could slow growth below 2% GDP annually.
- Enterprise Strategy: Firms like JPMorganChase invest in AI for 20%+ ROTCE, but must pair with upskilling to retain talent; 75 million digital users signal shift.3
- Policy Needs: Accelerated retraining (e.g., 100 billion USD U.S. funds), tax incentives for job creation, universal basic services to bridge gaps.
- Global Ramifications: Developing economies face steeper lags without infrastructure; AI arms race intensifies divides.
Business outlooks show 78% steady/increasing revenues, but headcount caution (45% static) hints at lean AI futures.12 Resolving this requires proactive scaling of education-online platforms reaching 1 billion learners-and public-private partnerships mirroring past successes.
Pathways to Balanced Adaptation
Mitigation strategies emerge from data. JPMorganChase's model-tech investments yielding records amid volatility-offers blueprint: AI for efficiency, humans for oversight.1,4 Projections: 51% workforce expansion plans if growth materializes.15
| Adaptation Lever |
Impact Potential |
Timeline |
| Massive Online Learning |
Upskill 500 million by 2030 |
1-3 years |
| AI-Human Hybrids |
Boost productivity 40% |
Immediate |
| Government Subsidies |
Fund 20% of transitions |
2-5 years |
Ultimately, the challenge demands vigilance: monitoring AI deployment against job metrics, with firms leading via internal academies. Historical resilience suggests navigability, but unprecedented speed elevates stakes for coordinated response.1
References
1. Jamie Dimon's Letter to Shareholders, Annual Report 2025 - 2026-04-06 - https://www.jpmorganchase.com/ir/annual-report/2025/ar-ceo-letters
2. Letter to Shareholders from Douglas B. Petno and Troy Rohrbaugh, Annual Report 2025 - 2026-04-06 - https://www.jpmorganchase.com/ir/annual-report/2025/ar-ceo-letter-petno-rohrbaugh
3. Letter to Shareholders from Marianne Lake, Annual Report 2025 - 2026-04-06 - https://www.jpmorganchase.com/ir/annual-report/2025/ar-ceo-letter-marianne-lake
4. Letter to Shareholders from Mary Callahan Erdoes, Annual Report 2025 - 2026-04-06 - https://www.jpmorganchase.com/ir/annual-report/2025/ar-ceo-letter-mary-callahan-erdoes
5. JPMorganChase Publishes 2025 Annual Report, Including Chairman & CEO Letter to Shareholders - 2026-04-06 - https://www.marketscreener.com/news/jpmorganchase-publishes-2025-annual-report-including-chairman-ceo-letter-to-shareholders-ce7e51d2de89fe2d
6. Letter to Shareholders from Jennifer A. Piepszak, Annual Report 2025 - 2026-04-06 - https://www.jpmorganchase.com/ir/annual-report/2025/ar-ceo-letter-jennifer-piepszak
7. JPMorganChase Publishes 2025 Annual Report, Including ... - 2026-04-06 - https://www.businesswire.com/news/home/20260405270223/en/JPMorganChase-Publishes-2025-Annual-Report-Including-Chairman-CEO-Letter-to-Shareholders
8. [PDF] Dear Fellow Shareholders, | JPMorgan Chase - 2026-04-06 - https://www.jpmorganchase.com/content/dam/jpmc/jpmorgan-chase-and-co/investor-relations/documents/ceo-letter-to-shareholders-2025.pdf
9. Future Finance Leaders 2025: Five Themes Shaping the Next Era of ... - 2025-12-05 - https://www.jpmorgan.com/insights/banking/five-themes-future-finance-leaders-2025
10. Annual Report | JPMorganChase - https://www.jpmorganchase.com/ir/annual-report
11. Jamie Dimon's Letter to Shareholders, Annual Report 2024 - 2025-04-07 - https://www.jpmorganchase.com/ir/annual-report/2024/ar-ceo-letters
12. 2025 Business Leaders Outlook Pulse Survey - J.P. Morgan - 2025-06-25 - https://www.jpmorgan.com/about-us/corporate-news/2025/2025-business-leaders-outlook-pulse-survey
13. JPMorgan Chase publishes 2025 annual report with CEO letter - 2026-04-06 - https://www.streetinsider.com/Corporate+News/JPMorgan+Chase+publishes+2025+annual+report+with+CEO+letter/26273772.html
14. Jamie Dimon's 2025 Shareholder Letter | PDF | Investing - Scribd - 2025-10-12 - https://www.scribd.com/document/914601117/Jamie-Dimon-April-2025-letter-to-shareholders
15. [PDF] 2025 U.S. Business Leaders Outlook - J.P. Morgan - https://www.jpmorgan.com/content/dam/jpmorgan/documents/cb/insights/outlook/business-leaders-outlook/cb-insights-business-leaders-outlook-2025-us.pdf

|
| |
|