ARTIFICIAL INTELLIGENCE
An AI-native strategy firmGlobal Advisors: a consulting leader in defining quantified strategy, decreasing uncertainty, improving decisions, achieving measureable results.
A Different Kind of Partner in an AI World
AI-native strategy
consulting
Experienced hires
We are hiring experienced top-tier strategy consultants
Quantified Strategy
Decreased uncertainty, improved decisions
Global Advisors is a leader in defining quantified strategies, decreasing uncertainty, improving decisions and achieving measureable results.
We specialise in providing highly-analytical data-driven recommendations in the face of significant uncertainty.
We utilise advanced predictive analytics to build robust strategies and enable our clients to make calculated decisions.
We support implementation of adaptive capability and capacity.
Our latest
Thoughts
Podcast – The Real AI Signal from Davos 2026
While the headlines from Davos were dominated by geopolitical conflict and debates on AGI timelines and asset bubbles, a different signal emerged from the noise. It wasn’t about if AI works, but how it is being ruthlessly integrated into the real economy.
In our latest podcast, we break down the “Diffusion Strategy” defining 2026.
3 Key Takeaways:
- China and the “Global South” are trying to leapfrog: While the West debates regulation, emerging economies are treating AI as essential infrastructure.
- China has set a goal for 70% AI diffusion by 2027.
- The UAE has mandated AI literacy in public schools from K-12.
- Rwanda is using AI to quadruple its healthcare workforce.
- The Rise of the “Agentic Self”: We aren’t just using chatbots anymore; we are employing agents. Entrepreneur Steven Bartlett revealed he has established a “Head of Experimentation and Failure” to use AI to disrupt his own business before competitors do. Musician will.i.am argued that in an age of predictive machines, humans must cultivate their “agentic self” to handle the predictable, while remaining unpredictable themselves.
- Rewiring the Core: Uber’s CEO Dara Khosrowshahi noted the difference between an “AI veneer” and a fundamental rewire. It’s no longer about summarising meetings; it’s about autonomous agents resolving customer issues without scripts.
The Global Advisors Perspective: Don’t wait for AGI. The current generation of models is sufficient to drive massive value today. The winners will be those who control their “sovereign capabilities” – embedding their tacit knowledge into models they own.
Read our original perspective here – https://with.ga/w1bd5
Listen to the full breakdown here – https://with.ga/2vg0z

Strategy Tools
Strategy Tools: The 7S Framework – A Comprehensive Guide
By John Khova Global Advisors digital consultant Introduction The McKinsey 7S Framework is one of the most enduring and widely recognised management models in strategic consulting and organisational design. It posits that organisational effectiveness depends not on...
Fast Facts
Fast Fact: Great returns aren’t enough
Key insights
It’s not enough to just have great returns – top-line growth is just as critical.
In fact, S&P 500 investors rewarded high-growth companies more than high-ROIC companies over the past decade.
While the distinction was less clear on the JSE, what is clear is that getting a balance of growth and returns is critical.
Strong and consistent ROIC or RONA performers provide investors with a steady flow of discounted cash flows – without growth effectively a fixed-income instrument.
Improvements in ROIC through margin improvements, efficiencies and working-capital optimisation provide point-in-time uplifts to share price.
Top-line growth presents a compounding mechanism – ROIC (and improvements) are compounded each year leading to on-going increases in share price.
However, without acceptable levels of ROIC, the benefits of compounding will be subdued and share price appreciation will be depressed – and when ROIC is below WACC value will be destroyed.
Maintaining high levels of growth is not as sustainable as maintaining high levels of ROIC – while both typically decline as industries mature, growth is usually more affected.
Getting the right balance between ROIC and growth is critical to optimising shareholder value.
Selected News
Term: Recursive Language Model (RLM)
“A Recursive Language Model (RLM) is an AI inference strategy where a large language model (LLM) is granted the ability to programmatically interact with and recursively call itself or smaller helper models to solve complex tasks and process extremely long inputs.” – Recursive Language Model (RLM)
A **Recursive Language Model (RLM)** is an innovative inference strategy that empowers large language models (LLMs) to treat input contexts not as static strings but as dynamic environments they can actively explore, decompose, and recursively process.1,3,4 This approach fundamentally shifts AI from passive text processing to active problem-solving, enabling the handling of extremely long inputs, complex reasoning tasks, and structured outputs without being constrained by traditional context window limits.1,6
At its core, an RLM operates within a Python Read-Eval-Print Loop (REPL) environment where the input context is stored as a programmable variable.1,2,3 The model begins with exploration and inspection, using tools like string slicing, regular expressions, and keyword searches to scan and understand the data structure actively rather than passively reading it.1 It then performs task decomposition, breaking the problem into smaller subtasks that fit within standard context windows, with the model deciding the splits based on its discoveries.1,3
The hallmark is recursive self-calls, where the model invokes itself (or smaller helper models) on each subtask, forming a tree of reasoning that aggregates partial results into variables within the REPL.1,4 This is followed by aggregation and synthesis, combining outputs programmatically into lists, tables, or documents, and verification and self-checking through re-runs or cross-checks for reliability.1 Unlike traditional LLMs that process a single forward pass on tokenised input, RLMs grant the model ‘hands and eyes’ to query itself programmatically, such as result = rlm_query(sub_prompt), transforming context from ‘input’ to ‘environment’.1,3
RLMs address key limitations like ‘context rot’-degradation in long-context performance-and scale to effectively unlimited lengths (over 10 million tokens tested), outperforming baselines by up to 114% on benchmarks without fine-tuning, via prompt engineering alone.3,6,2 They differ from agentic systems by decomposing context adaptively rather than predefined tasks, and from reasoning models by scaling through recursive decomposition.6
Key Theorist: Alex L. Zhang and the MIT Origins
The primary theorist behind RLMs is **Alex L. Zhang**, a researcher affiliated with MIT, who co-authored the seminal work proposing RLMs as a general inference framework.3,4,8 In his detailed blog and the arXiv paper ‘Recursive Language Models’ (published around late 2025), Zhang articulates the vision: enabling LLMs to ‘recursively call themselves or other LLMs’ to process unbounded contexts and mitigate degradation.3,4 His implementation uses GPT-5 or GPT-5-mini in a Python REPL, allowing adaptive chunking and recursion at test time.3
Alex L. Zhang’s biography reflects a deep expertise in AI scaling and inference innovations. Active in 2025 through platforms like his GitHub blog (alexzhang13.github.io), he focuses on practical advancements in language model capabilities, particularly long-context handling.3 While specific early career details are sparse in available sources, his work builds on MIT’s disruptive ethos-echoed in proposals like ‘why not let the model read itself?’-positioning him as a key figure in the 2026 paradigm shift towards recursive AI architectures.1,8 Zhang’s contributions emphasise test-time compute scaling, distinguishing RLMs from mere architectural changes by framing them as a ‘thin wrapper’ around standard LLMs that reframes them as stateful programmes.5
Experimental validations in Zhang’s framework demonstrate RLMs’ superiority, such as dramatically improved accuracy on pairwise comparison tasks (from near-zero to over 58%) and spam classification in massive prompts.2,4 His ideas have sparked widespread discussion, with sources hailing RLMs as ‘the ultimate evolution of AI’ and a ‘game-changer for 2026’.1,2,7
References
1. https://gaodalie.substack.com/p/rlm-the-ultimate-evolution-of-ai
3. https://alexzhang13.github.io/blog/2025/rlm/
4. https://arxiv.org/html/2512.24601v1
5. https://datasciencedojo.com/blog/what-are-recursive-language-models/
7. https://www.primeintellect.ai/blog/rlm

Polls
What determines your success?
We need your help!
We’re testing how people think about their successes and failures. It would be great if you would take 2 minutes to give a simple multiple choice answer and share the poll with your friends.
Take the poll – “What determines your success?”
Then read So You Think You’re Self Aware?
Services
Global Advisors is different
We help clients to measurably improve strategic decision-making and the results they achieve through defining clearly prioritised choices, reducing uncertainty, winning hearts and minds and partnering to deliver.
Our difference is embodied in our team. Our values define us.
Corporate portfolio strategy
Define optimal business portfolios aligned with investor expectations
BUSINESS UNIT STRATEGY
Define how to win against competitors
Reach full potential
Understand your business’ core, reach full potential and grow into optimal adjacencies
Deal advisory
M&A, due diligence, deal structuring, balance sheet optimisation
Global Advisors Digital Data Analytics
14 years of quantitative and data science experience
An enabler to delivering quantified strategy and accelerated implementation
Digital enablement, acceleration and data science
Leading-edge data science and digital skills
Experts in large data processing, analytics and data visualisation
Developers of digital proof-of-concepts
An accelerator for Global Advisors and our clients
Join Global Advisors
We hire and grow amazing people
Consultants join our firm based on a fit with our values, culture and vision. They believe in and are excited by our differentiated approach. They realise that working on our clients’ most important projects is a privilege. While the problems we solve are strategic to clients, consultants recognise that solutions primarily require hard work – rigorous and thorough analysis, partnering with client team members to overcome political and emotional obstacles, and a large investment in knowledge development and self-growth.
Get In Touch
16th Floor, The Forum, 2 Maude Street, Sandton, Johannesburg, South Africa
+27114616371
