We show that asymmetric deep recurrent neural networks, enhanced with additional sparse excitatory couplings, give rise to an exponentially large, dense accessible manifold of internal representations which can be found by different algorithms, including simple iterative dynamics. Building on the geometrical properties of the stable configurations, we propose a distributed learning scheme in which input-output associations emerge naturally from the recurrent dynamics, without any need of gradient evaluation. A critical feature enabling the learning process is the stability of the configurations reached at convergence, even after removal of the supervisory output signal. Extensive simulations demonstrate that this approach performs competitively on standard AI benchmarks. The model can be generalized in multiple directions, both computational and biological, potentially contributing to narrowing the gap between AI and computational neuroscience.
翻译:我们证明,通过引入额外的稀疏兴奋性耦合增强的深度非对称循环神经网络,能够产生一个指数级庞大、密集的可访问内部表示流形,该流形可通过不同算法(包括简单的迭代动力学)发现。基于稳定构型的几何特性,我们提出一种分布式学习方案,其中输入-输出关联从循环动力学中自然涌现,无需任何梯度计算。使学习过程得以实现的一个关键特征是,即使在移除监督输出信号后,收敛时达到的构型仍保持稳定。大量仿真实验表明,该方法在标准人工智能基准测试中表现出竞争力。该模型可在计算与生物学等多个方向进行推广,有望为缩小人工智能与计算神经科学之间的差距作出贡献。