AIResearch AIResearch
Back to articles
Science

AI Agents Learn Like Humans Without Forgetting

A new framework helps AI systems continuously learn new tasks while retaining old skills, outperforming current methods in benchmarks and enabling smarter robots.

AI Research
November 14, 2025
3 min read
AI Agents Learn Like Humans Without Forgetting

Artificial intelligence systems often struggle to learn new tasks without forgetting previous ones, a problem known as catastrophic forgetting. This limitation hinders their ability to adapt in real-world environments, such as robotics or data analysis, where continuous learning is essential. Researchers have introduced a novel approach called Eigentasks, which enables AI agents to acquire and transfer skills across tasks while maintaining performance on old ones, achieving state-of-the-art results in supervised learning and reinforcement learning benchmarks.

The key finding is that Eigentasks combine generative models with skills in a way that separates tasks into distinct components, allowing for selective knowledge transfer. This method avoids the common issue where learning a new task erases memory of past tasks, and it improves the agent's ability to quickly adapt to related new challenges. For example, in experiments, the framework demonstrated positive forward transfer, meaning agents performed better on new tasks by leveraging previously learned skills.

Methodologically, the researchers developed the Eigentasks framework using a wake-sleep cycle. During the wake phase, agents solve incoming tasks and store experiences in a buffer. In the sleep phase, they consolidate memories through generative replay, where a model generates synthetic data from past tasks to reinforce learning without retraining on original data. A specific instantiation, the Open World Variational Auto Encoder (OWVAE), uses variational autoencoders as generators and includes a rejection sampling strategy to enhance replay quality by filtering out low-confidence samples. This approach was applied to both supervised classification tasks, like image recognition, and reinforcement learning scenarios, such as video game environments.

Results from the paper show that OWVAE outperformed state-of-the-art methods in continual learning benchmarks. On a combined dataset of MNIST and FashionMNIST, OWVAE achieved an average accuracy of up to 77.43%, compared to lower accuracies in baseline methods like generative replay alone. In reinforcement learning tests using Starcraft II mini-games, agents using Eigentasks showed a jump-start in performance, reaching high rewards 1.5 to 10 times faster than single-task learning. For instance, in combat tasks like DefeatZerglingsAndBanelings, transfer from similar tasks led to immediate performance gains, while navigation tasks benefited from adjusted exploration strategies.

This advancement matters because it brings AI closer to human-like learning, where knowledge accumulates over time without loss. In practical terms, it could lead to more adaptable robots that learn new behaviors in dynamic environments or AI systems in healthcare that continuously update without forgetting critical diagnostic skills. For everyday users, this means smarter, more reliable AI assistants and automation tools that improve with experience.

Limitations noted in the paper include the need for further research to close the wake-sleep loop in reinforcement learning fully and improve out-of-distribution detection. The framework currently relies on predefined task boundaries in some experiments, and its performance can vary with dissimilar tasks, indicating areas for future refinement to handle more complex, real-world scenarios.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn