In Machine Learning, optimization mostly has been done by using a gradient descent method to find the minimum value of the loss. However, especially in deep learning, finding a global minimum from a nonconvex loss function across a high dimensional space is an extraordinarily difficult task. Recently, a generalization learning algorithm, Sharpness-Aware Minimization (SAM), has made a great success in image classification task. Despite the great performance in creating convex space, proper direction leading by SAM is still remained unclear. We, thereby, propose a creating a Unit Vector space in SAM, which not only consisted of the mathematical instinct in linear algebra but also kept the advantages of adaptive gradient algorithm. Moreover, applying SAM in unit gradient brings models competitive performances in image classification datasets, such as CIFAR - {10, 100}. The experiment showed that it performed even better and more robust than SAM.
翻译:在机器学习中,优化主要是通过使用梯度下降法找到损失的最小值。然而,特别是在深层学习中,从高维空间的非混凝土损失函数中找到一个全球最低值是一项极为困难的任务。最近,一般化学习算法,即夏普内斯-软件最小化(SAM)在图像分类任务中取得了巨大成功。尽管在创建曲线空间方面表现出色,但SAM的正确领导方向仍然不明确。因此,我们提议在SAM中建立一个单位矢量空间,它不仅包括线性代数的数学本能,而且还保留了适应性梯度算法的优点。此外,在单位梯度中应用SAM在图像分类数据集中带来了模型的竞争性性能,如CIFAR - {10, 100}。实验表明,它的表现比SAM更好、更有力。