Select Page

ARTIFICIAL INTELLIGENCE

An AI-native strategy firm

Global Advisors: a consulting leader in defining quantified strategy, decreasing uncertainty, improving decisions, achieving measureable results.

Learn MoreGlobal Advisors AI

A Different Kind of Partner in an AI World

AI-native strategy
consulting

Experienced hires

We are hiring experienced top-tier strategy consultants

Quantified Strategy

Decreased uncertainty, improved decisions

Global Advisors is a leader in defining quantified strategies, decreasing uncertainty, improving decisions and achieving measureable results.

We specialise in providing highly-analytical data-driven recommendations in the face of significant uncertainty.

We utilise advanced predictive analytics to build robust strategies and enable our clients to make calculated decisions.

We support implementation of adaptive capability and capacity.

Our latest

Thoughts

Global Advisors’ Thoughts: Leading a deliberate life

Global Advisors’ Thoughts: Leading a deliberate life

By Marc Wilson
Marc is a partner at Global Advisors and based in Johannesburg, South Africa

Download this article at https://globaladvisors.biz/blog/2018/06/26/leading-a-deliberate-life/.

Picket fences. Family of four. Management position.

Mid-life crisis. Meaning. Purpose.

Someone once said that, “At 18, I had all the answers. At 35, I realised I didn’t know the question.”

Serendipity has a lot going for it. Many people might sail through life taking what comes and enjoying the moment. Others might be open to chance and have nothing go right for them.

Some people might strive to achieve, realise rare successes and be bitterly unhappy. Others might be driven and enjoy incredible success and fulfilment.

Perhaps the majority of us become beholden to the momentum of our lives.

We might study, start a career, marry, buy a dream house, have children, send them to a top school. Those steps make up components of many of our dreams. They are steps that may define each subsequent choice. As I discussed this with a friend recently, he remarked that few of these steps had been subject of deliberations in his life – increasingly these steps were the outcome of momentum. Each will shape every step he takes for the rest of his life. He would not have things any other way, but if he knew what he knows now, he might have been more deliberate about choice and consequence…..

Read more at https://globaladvisors.biz/blog/2018/06/26/leading-a-deliberate-life/

.

read more

Strategy Tools

PODCAST: Strategy Tools: Growth, Profit or Returns?

PODCAST: Strategy Tools: Growth, Profit or Returns?

Our Spotify podcast explores the relationship between Return on Net Assets (RONA) and growth, arguing that both are essential for shareholder value creation. The hosts contend that focusing solely on one metric can be detrimental, and propose a framework for evaluating business portfolios based on their RONA and growth profiles. This approach involves plotting business units on a “market-cap curve” to identify value-accretive and value-destructive segments.

The podcast also addresses the impact of economic downturns on portfolio management, suggesting strategies for both offensive and defensive approaches. The core argument is that companies should aim to achieve a balance between RONA and growth, acknowledging that both are essential for long-term shareholder value creation.

Read more from the original article – https://globaladvisors.biz/2020/08/04/strategy-tools-growth-profit-or-returns/

read more

Fast Facts

Selected News

Term: Mixture of Experts (MoE)

Term: Mixture of Experts (MoE)

“Mixture of Experts (MoE) is an efficient neural network architecture that uses multiple specialised sub-models (experts) and a gating network (router) to dynamically select and activate only the most relevant experts for a given input.” – Mixture of Experts (MoE)

This architectural approach divides a large artificial intelligence model into separate sub-networks, each specialising in processing specific types of input data. Rather than activating the entire network for every task, MoE models employ a gating mechanism-often called a router-that intelligently selects which experts should process each input. This selective activation introduces sparsity into the network, meaning only a fraction of the model’s total parameters are used for any given computation.1,3

Core Architecture and Components

The fundamental structure of MoE consists of two essential elements:4

  • Expert networks: Multiple specialised sub-networks, typically implemented as feed-forward neural networks (FFNs), each with its own set of learnable parameters. These experts become skilled at handling specific patterns or types of data during training.1
  • Gating network (router): A trainable mechanism that evaluates each input and determines which expert or combination of experts is best suited to process it. This routing function is computationally efficient, enabling the model to make rapid decisions about expert selection.1,3

In practical implementations, such as the Mixtral 8x7B language model, each layer contains multiple experts-for instance, eight separate feedforward blocks with 7 billion parameters each. For every token processed, the router selects only a subset of these experts (in Mixtral’s case, two out of eight) to perform the computation, then combines their outputs before passing the result to the next layer.3

How MoE Achieves Efficiency

MoE models leverage conditional computation to reduce computational burden without sacrificing model capacity.3 This approach enables several efficiency gains:

  • Models can scale to billions of parameters whilst maintaining manageable inference costs, since not all parameters are activated for every input.1,3
  • Training can occur with significantly less compute, allowing researchers to either reduce training time or expand model and dataset sizes.4
  • Experts can be distributed across multiple devices through expert parallelism, enabling efficient large-scale deployments.1

The gating mechanism ensures that frequently selected experts receive continuous updates during training, improving their performance, whilst load balancing mechanisms attempt to distribute computational work evenly across experts to prevent bottlenecks.1

Historical Development and Key Theorist: Noam Shazeer

Noam Shazeer stands as the primary architect of modern MoE systems in deep learning. In 2017, Shazeer and colleagues-including the legendary Geoffrey Hinton and Google’s Jeff Dean-introduced the Sparsely-Gated Mixture-of-Experts Layer for recurrent neural language models.1,4 This seminal work fundamentally transformed how researchers approached scaling neural networks.

Shazeer’s contribution was revolutionary because it reintroduced the mixture of experts concept, which had existed in earlier machine learning literature, into the deep learning era. His team scaled this architecture to a 137-billion-parameter LSTM model, demonstrating that sparsity could maintain very fast inference even at massive scale.4 Although this initial work focused on machine translation and encountered challenges such as high communication costs and training instabilities, it established the theoretical and practical foundation for all subsequent MoE research.4

Shazeer’s background as a researcher at Google positioned him at the intersection of theoretical machine learning and practical systems engineering. His work exemplified a crucial insight: that not all parameters in a neural network need to be active simultaneously. This principle has since become foundational to modern large language model design, influencing architectures used by leading AI organisations worldwide. The Sparsely-Gated Mixture-of-Experts Layer introduced the trainable gating network concept that remains central to MoE implementations today, enabling conditional computation that balances model expressiveness with computational efficiency.1

Applications and Performance

MoE architectures have demonstrated faster training and comparable or superior performance to dense language models on many benchmarks, particularly in multi-domain tasks where different experts can specialise in different knowledge areas.1 Applications span natural language processing, computer vision, and recommendation systems.2

Challenges and Considerations

Despite their advantages, MoE systems present implementation challenges. Load balancing remains critical-when experts are distributed across multiple devices, uneven expert selection can create memory and computational bottlenecks, with some experts handling significantly more tokens than others.1 Additionally, distributed training complexity and the need for careful tuning to maintain stability and efficiency require sophisticated engineering approaches.1

References

1. https://neptune.ai/blog/mixture-of-experts-llms

2. https://www.datacamp.com/blog/mixture-of-experts-moe

3. https://www.ibm.com/think/topics/mixture-of-experts

4. https://huggingface.co/blog/moe

5. https://newsletter.maartengrootendorst.com/p/a-visual-guide-to-mixture-of-experts

6. https://www.youtube.com/watch?v=sYDlVVyJYn4

7. https://arxiv.org/html/2503.07137v1

8. https://cameronrwolfe.substack.com/p/moe-llms

"Mixture of Experts (MoE) is an efficient neural network architecture that uses multiple specialised sub-models (experts) and a gating network (router) to dynamically select and activate only the most relevant experts for a given input." - Term: Mixture of Experts (MoE)

read more

Polls

No Results Found

The page you requested could not be found. Try refining your search, or use the navigation above to locate the post.

Services

Global Advisors is different

We help clients to measurably improve strategic decision-making and the results they achieve through defining clearly prioritised choices, reducing uncertainty, winning hearts and minds and partnering to deliver.

Our difference is embodied in our team. Our values define us.

Corporate portfolio strategy

Define optimal business portfolios aligned with investor expectations

BUSINESS UNIT STRATEGY

Define how to win against competitors

Reach full potential

Understand your business’ core, reach full potential and grow into optimal adjacencies

Deal advisory

M&A, due diligence, deal structuring, balance sheet optimisation

Global Advisors Digital Data Analytics

14 years of quantitative and data science experience

An enabler to delivering quantified strategy and accelerated implementation

Digital enablement, acceleration and data science

Leading-edge data science and digital skills

Experts in large data processing, analytics and data visualisation

Developers of digital proof-of-concepts

An accelerator for Global Advisors and our clients

Join Global Advisors

We hire and grow amazing people

Consultants join our firm based on a fit with our values, culture and vision. They believe in and are excited by our differentiated approach. They realise that working on our clients’ most important projects is a privilege. While the problems we solve are strategic to clients, consultants recognise that solutions primarily require hard work – rigorous and thorough analysis, partnering with client team members to overcome political and emotional obstacles, and a large investment in knowledge development and self-growth.

Get In Touch

16th Floor, The Forum, 2 Maude Street, Sandton, Johannesburg, South Africa
+27114616371

Global Advisors | Quantified Strategy Consulting