Conversational Recommender Systems (CRS) has become an emerging research topic seeking to perform recommendations through interactive conversations, which generally consist of generation and recommendation modules. Prior work on CRS tends to incorporate more external and domain-specific knowledge like item reviews to enhance performance. Despite the fact that the collection and annotation of the external domain-specific information needs much human effort and degenerates the generalizability, too much extra knowledge introduces more difficulty to balance among them. Therefore, we propose to fully discover and extract internal knowledge from the context. We capture both entity-level and contextual-level representations to jointly model user preferences for the recommendation, where a time-aware attention is designed to emphasize the recently appeared items in entity-level representations. We further use the pre-trained BART to initialize the generation module to alleviate the data scarcity and enhance the context modeling. In addition to conducting experiments on a popular dataset (ReDial), we also include a multi-domain dataset (OpenDialKG) to show the effectiveness of our model. Experiments on both datasets show that our model achieves better performance on most evaluation metrics with less external knowledge and generalizes well to other domains. Additional analyses on the recommendation and generation tasks demonstrate the effectiveness of our model in different scenarios.
翻译:交流建议系统(CRS)已成为一个新兴研究课题,寻求通过互动对话提出建议,这些互动对话通常由生成和建议模块组成。关于CRS的先前工作往往包括更多的外部和特定领域知识,如项目审查,以提高绩效。尽管外部领域特定信息的收集和说明需要大量人力工作,并削弱了一般性,但过多的额外知识给平衡带来了更多困难。因此,我们提议从背景中充分发现和提取内部知识。我们从实体一级和背景层面的表述中收集到共同模拟建议用户偏好,以便共同模拟该建议的用户偏好,在其中,注意时间,强调最近在实体一级表述中出现的项目。我们进一步利用经过事先培训的BART启动生成模块,以缓解数据稀缺情况,加强背景建模。除了对大众数据集(ReDal)进行实验外,我们还包含一个多数据数据集(OpenDialKG),以显示我们模型的有效性。我们两个数据集的实验表明,我们模型在大多数评价指标上取得了更好的业绩,外部知识较少,其他模型也很好地展示了我们的不同模型。