It is common practice to use large computational resources to train neural networks, as is known from many examples, such as reinforcement learning applications. However, while massively parallel computing is often used for training models, it is rarely used for searching solutions for combinatorial optimization problems. In this paper, we propose a novel massively parallel Monte-Carlo Tree Search (MP-MCTS) algorithm that works efficiently for 1,000 worker scale, and apply it to molecular design. This is the first work that applies distributed MCTS to a real-world and non-game problem. Existing work on large-scale parallel MCTS show efficient scalability in terms of the number of rollouts up to 100 workers, but suffer from the degradation in the quality of the solutions. MP-MCTS maintains the search quality at larger scale, and by running MP-MCTS on 256 CPU cores for only 10 minutes, we obtained candidate molecules having similar score to non-parallel MCTS running for 42 hours. Moreover, our results based on parallel MCTS (combined with a simple RNN model) significantly outperforms existing state-of-the-art work. Our method is generic and is expected to speed up other applications of MCTS.
翻译:使用大量计算资源来培训神经网络是常见的做法,这一点从许多例子中可以知道,例如强化学习应用等许多例子。然而,虽然大量平行计算常常用于培训模型,但很少用于寻找组合优化问题的解决办法。在本文中,我们提议采用一个全新的大规模平行的蒙特-卡洛树搜索(MP-MCTS)算法,该算法对1,000名工人有效,并适用于分子设计。这是将MCTS应用于现实世界和非游戏问题的首次工作。大规模平行MCTS的现有工作显示,在向100名工人推出的数量方面是有效的,但因解决方案质量下降而受到影响。MP-MCTS在更大程度上保持了搜索质量,通过在256 CPU核心上只运行10分钟的MP-MCTS,我们获得了与运行42小时的非平行 MCTS相近分数的候选分子。此外,我们基于平行的MCTS(与简单的RNNMTS模型相结合)的现有工作结果大大超过现有状态MTS速度。我们所期望的其他方法是通用的,其他应用速度。