ArXiv TLDR

Attention to task structure for cognitive flexibility

🐦 Tweet
2604.13281

Xiaoyu K. Zhang, Mehdi Senoussi, Tom Verguts

cs.NEq-bio.NC

TLDR

This paper explores how environmental task structure and attention models influence cognitive flexibility, stability, and generalization in multi-task learning.

Key contributions

  • Designed a multi-task learning environment where tasks are defined by cue dimensions and characterized using graph theory.
  • Introduced gating-based and concatenation-based attention models for task decomposition and sequential attention.
  • Demonstrated that richer environments improve cognitive generalization and stability in multi-task learning.
  • Showed task connectivity (graph theory based) significantly modulates stability and generalization, especially for attention models.

Why it matters

This research highlights the crucial, often overlooked, role of environmental task structure in cognitive flexibility, not just model architecture. It provides insights for designing more robust and adaptable AI agents by considering how tasks are interconnected. This work can inform future multi-task learning system development.

Original Abstract

Humans and artificial agents must often learn and switch between multiple tasks in dynamic environments. Success in such settings requires cognitive flexibility: the ability to retain prior knowledge (cognitive stability) while also transferring it to novel tasks (cognitive generalization). Cognitive flexibility research has largely focused on the role of model architecture to achieve these complementary goals. However, it is less well understood how the structure of the environment itself influences cognitive flexibility, and how it interacts with model architecture. To address this gap, we design a multi-task learning environment in which tasks are defined by a combination of two cue dimensions, allowing us to characterize the environment with graph-theory methods. We also introduce gating-based (multiplicative) and concatenation-based attention models that can decompose tasks into components and can sequentially allocate attention to them. We compare the attention-based models' performance in the multi-task learning environment to multilayer perceptrons. Generalization and stability are systematically evaluated across environments that vary in richness and task connectivity. We observe that richer environments improve both generalization and stability. In addition, a critical novel observation is that (graph theory based) connectivity between the tasks in the environment strongly modulates both stability and generalization, with especially pronounced benefits for attention-based models. These findings underscore the importance of considering not only cognitive architectures but also environmental structure and their interaction in shaping multi-task learning, generalization, and stability.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.