Google Research has introduced Nested Learning, a machine learning approach designed to address the problem of catastrophic forgetting in continual learning.
This new method has been detailed in the paper “Nested Learning: The Illusion of Deep Learning Architectures” at NeurIPS 2025.
Access deeper industry intelligence
Experience unmatched clarity with a single platform that combines unique data, AI, and human expertise.
It frames neural models as sets of interconnected optimisation problems that operate at multiple levels, each with distinct context flows and update frequencies.
Catastrophic forgetting remains a recognised limitation in current large language models (LLMs), where adaptation to new tasks can result in the loss of previously acquired knowledge.
Conventional strategies tend to treat architectural design and optimisation algorithms as separate entities.
Nested Learning challenges this separation by assuming that both are different layers within a unified system of nested optimisation tasks, according to Google.
US Tariffs are shifting - will you react or anticipate?
Don’t let policy changes catch you off guard. Stay proactive with real-time data and expert analysis.
By GlobalDataIn this view, learning occurs across a spectrum of modules, each managing its own internal information and update cycle.
Google Research’s team tested these principles by developing Hope, a self-modifying recurrent architecture built on Titans memory modules but augmented with continuum memory systems (CMS).
The CMS structure allows for variable update rates across memory components, aiming to more closely align with patterns observed in human neuroplasticity.
Google reported that Hope produces lower perplexity and higher accuracy than standard transformers and recurrent models on several public language modelling and reasoning benchmarks.
Nested Learning generalises both optimisers and key architectural elements as associative memory systems, formalising their role as mapping functions between data points and error signals or sequence relationships.
The Hope architecture exploits nested optimisation by enabling memory components to update at multiple frequencies, forming what the authors refer to as a continuum memory system.
This allows for self-referential modification and integration of new data without discarding existing information.
Google is seeking further exploration of this approach within the broader machine learning community.
Recently, the US Department of Justice (DOJ) concluded its antitrust review of Google’s $32bn Wiz acquisition, removing a major regulatory barrier for the Alphabet subsidiary as it seeks to advance in the cloud security market.
The Federal Trade Commission (FTC) confirmed on its website that early termination for the DOJ antitrust review was granted on 24 October 2025.
