Recommender retrievers aim to rapidly retrieve a fraction of items from the entire item corpus when a user query requests, with the representative two-tower model trained with the log softmax loss. For efficiently training recommender retrievers on modern hardwares, inbatch sampling, where the items in the mini-batch are shared as negatives to estimate the softmax function, has attained growing interest. However, existing inbatch sampling based strategies just correct the sampling bias of inbatch items with item frequency, being unable to distinguish the user queries within the mini-batch and still incurring significant bias from the softmax. In this paper, we propose a Cache-Augmented Inbatch Importance Resampling (XIR) for training recommender retrievers, which not only offers different negatives to user queries with inbatch items, but also adaptively achieves a more accurate estimation of the softmax distribution. Specifically, XIR resamples items for the given mini-batch training pairs based on certain probabilities, where a cache with more frequently sampled items is adopted to augment the candidate item set, with the purpose of reusing the historical informative samples. XIR enables to sample query-dependent negatives based on inbatch items and to capture dynamic changes of model training, which leads to a better approximation of the softmax and further contributes to better convergence. Finally, we conduct experiments to validate the superior performance of the proposed XIR compared with competitive approaches.
翻译:建议检索器力求在用户询问请求时迅速从整个项目库中检索部分物品。 代表的二至二楼模型经过对日志软模损失的训练。 为了高效培训现代硬件的推荐人检索员, 将微型批量中的物品作为负数共享, 用于估算软模功能, 兴趣日益增长。 但是, 现有的以批量取样为基础的战略仅仅纠正了以项目频率批量项目的抽样偏差, 无法在小型批量中区分用户查询, 并且仍然与软模有重大偏差。 在本文中, 我们提议为培训推荐人提供高效培训建议人进行重要性抽查( XIR), 不仅为用户询问软模组功能提供不同的负数, 而且适应性地对软模组分布作出更准确的估计。 具体地说, XIR 类批量培训配对基于某种概率, 与更经常的软模集项目缓存一起, 以更高级的比重来提升候选项目的比重比重抽比重, 使得历史信息性样本更精确地转换为最后的样品, 。