Discovering the neural mechanisms underpinning cognition is one of the grand challenges of neuroscience. However, previous approaches for building models of RNN dynamics that explain behaviour required iterative refinement of architectures and/or optimisation objectives, resulting in a piecemeal, and mostly heuristic, human-in-the-loop process. Here, we offer an alternative approach that automates the discovery of viable RNN mechanisms by explicitly training RNNs to reproduce behaviour, including the same characteristic errors and suboptimalities, that humans and animals produce in a cognitive task. Achieving this required two main innovations. First, as the amount of behavioural data that can be collected in experiments is often too limited to train RNNs, we use a non-parametric generative model of behavioural responses to produce surrogate data for training RNNs. Second, to capture all relevant statistical aspects of the data, we developed a novel diffusion model-based approach for training RNNs. To showcase the potential of our approach, we chose a visual working memory task as our test-bed, as behaviour in this task is well known to produce response distributions that are patently multimodal (due to swap errors). The resulting network dynamics correctly qualitative features of macaque neural data. Importantly, these results were not possible to obtain with more traditional approaches, i.e., when only a limited set of behavioural signatures (rather than the full richness of behavioural response distributions) were fitted, or when RNNs were trained for task optimality (instead of reproducing behaviour). Our approach also yields novel predictions about the mechanism of swap errors, which can be readily tested in experiments. These results suggest that fitting RNNs to rich patterns of behaviour provides a powerful way to automatically discover mechanisms of important cognitive functions.
翻译:揭示认知背后的神经机制是神经科学领域的重大挑战之一。然而,以往构建能够解释行为的循环神经网络(RNN)动力学模型的方法,需要对架构和/或优化目标进行迭代优化,导致这一过程呈现碎片化、主要依赖启发式且需要人工介入的特点。本文提出一种替代性方法,通过明确训练RNN来复现人类和动物在认知任务中表现出的行为(包括相同的特征性错误和次优表现),从而实现可行RNN机制的自动发现。实现这一目标需要两项主要创新。首先,由于实验中可收集的行为数据量通常不足以训练RNN,我们采用非参数生成模型对行为响应进行建模,以生成用于训练RNN的替代数据。其次,为捕捉数据的所有相关统计特征,我们开发了一种基于扩散模型的新型RNN训练方法。为展示本方法的潜力,我们选择视觉工作记忆任务作为测试平台,因为该任务中的行为已知会产生明显多模态的响应分布(由交换错误导致)。训练得到的网络动力学正确再现了猕猴神经数据的定性特征。重要的是,这些结果无法通过更传统的方法获得——例如仅拟合有限的行为特征(而非行为响应分布的完整丰富性),或训练RNN追求任务最优性(而非复现实际行为)。我们的方法还产生了关于交换错误机制的新预测,这些预测可在实验中直接验证。这些结果表明,将RNN拟合到丰富的行为模式中,为自动发现重要认知功能的机制提供了一种强有力的途径。