We present a comparison of several online machine learning techniques for tactical learning and proving in the Coq proof assistant. This work builds on top of Tactician, a plugin for Coq that learns from proofs written by the user to synthesize new proofs. Learning happens in an online manner, meaning that Tactician's machine learning model is updated immediately every time the user performs a step in an interactive proof. This has important advantages compared to the more studied offline learning systems: (1) it provides the user with a seamless, interactive experience with Tactician and, (2) it takes advantage of locality of proof similarity, which means that proofs similar to the current proof are likely to be found close by. We implement two online methods, namely approximate k-nearest neighbors based on locality sensitive hashing forests and random decision forests. Additionally, we conduct experiments with gradient boosted trees in an offline setting using XGBoost. We compare the relative performance of Tactician using these three learning methods on Coq's standard library.
翻译:我们比较了用于战术学习的几种在线机器学习技术,并在 Coq 校对助理中进行验证。 这项工作以Tactician (Coq 的插件) 为基础, 该插件从用户写的证据中学习, 以合成新的证据。 学习以在线方式进行, 这意味着Tactician 的机器学习模式每次用户在互动证据中迈出一步时都会立即更新。 与更多研究的离线学习系统相比,这具有重要的优势:(1) 它为用户提供了与Tactician 的无缝互动经验,(2) 它利用了相似的证明地点, 也就是说, 类似当前证据的证据很可能在附近找到。 我们实施了两种在线方法, 即基于地方敏感草原森林和随机决定森林的近邻。 此外, 我们用 XGBoost 在离线环境中用梯度增生树进行实验。 我们比较了Tactician 在Coq 标准图书馆使用这三种学习方法的相对表现。