Molecular discovery has brought great benefits to the chemical industry. Various molecule design techniques are developed to identify molecules with desirable properties. Traditional optimization methods, such as genetic algorithms, continue to achieve state-of-the-art results across multiple molecular design benchmarks. However, these techniques rely solely on random walk exploration, which hinders both the quality of the final solution and the convergence speed. To address this limitation, we propose a novel approach called Gradient Genetic Algorithm (Gradient GA), which incorporates gradient information from the objective function into genetic algorithms. Instead of random exploration, each proposed sample iteratively progresses toward an optimal solution by following the gradient direction. We achieve this by designing a differentiable objective function parameterized by a neural network and utilizing the Discrete Langevin Proposal to enable gradient guidance in discrete molecular spaces. Experimental results demonstrate that our method significantly improves both convergence speed and solution quality, outperforming cutting-edge techniques. For example, it achieves up to a 25% improvement in the top-10 score over the vanilla genetic algorithm. The code is publicly available at https://github.com/debadyuti23/GradientGA.
翻译:分子发现为化学工业带来了巨大益处。为识别具有理想性质的分子,已开发出多种分子设计技术。传统优化方法,如遗传算法,在多个分子设计基准测试中持续取得最先进的结果。然而,这些技术仅依赖于随机游走探索,这限制了最终解的质量和收敛速度。为解决这一局限,我们提出了一种称为梯度遗传算法的新方法,它将目标函数的梯度信息融入遗传算法中。每个提出的样本不再进行随机探索,而是通过沿梯度方向迭代地向最优解推进。我们通过设计一个由神经网络参数化的可微目标函数,并利用离散朗之万提议在离散分子空间中实现梯度引导来实现这一点。实验结果表明,我们的方法显著提高了收敛速度和解的质量,超越了前沿技术。例如,在top-10得分上,相比原始遗传算法实现了高达25%的提升。代码公开于https://github.com/debadyuti23/GradientGA。