Motivated by scenarios of information diffusion and advertising in social media, we study an influence maximization problem in which little is assumed to be known about the diffusion network or about the model that determines how information may propagate. In such a highly uncertain environment, one can focus on multi-round diffusion campaigns, with the objective to maximize the number of distinct users that are influenced or activated, starting from a known base of few influential nodes. During a campaign, spread seeds are selected sequentially at consecutive rounds, and feedback is collected in the form of the activated nodes at each round. A round's impact (reward) is then quantified as the number of newly activated nodes. Overall, one must maximize the campaign's total spread, as the sum of rounds' rewards. In this setting, an explore-exploit approach could be used to learn the key underlying diffusion parameters, while running the campaign. We describe and compare two methods of contextual multi-armed bandits, with upper-confidence bounds on the remaining potential of influencers, one using a generalized linear model and the Good-Turing estimator for remaining potential (GLM-GT-UCB), and another one that directly adapts the LinUCB algorithm to our setting (LogNorm-LinUCB). We show that they outperform baseline methods using state-of-the-art ideas, on synthetic and real-world data, while at the same time exhibiting different and complementary behavior, depending on the scenarios in which they are deployed.
翻译:以社交媒体信息传播和广告的假想为动力,我们研究影响最大化问题,在这种假想中,人们很少认为对信息传播网络或决定信息传播方式的模式了解多少。在这种高度不确定的环境中,人们可以把重点放在多轮传播运动上,目的是从几个有影响力的节点的已知基地开始,使影响或激活的不同用户数量最大化。在一次运动中,连续几轮连续挑选传播种子,以每个回合的激活节点的形式收集反馈。然后将一轮的影响(向上)量化为新激活节点的数量。总体而言,必须最大限度地扩大运动的总传播,作为回合的奖励之和。在这一环境中,可以使用探索性方法学习关键的基本传播参数,同时从运动中开始。我们描述并比较了两种背景背景多臂匪徒的方法,对影响者剩余的潜力有上下限,其中一种是使用一般线性模型,对剩余潜力的预估值(GLM-GT-UCB)进行量化。总体而言,必须最大限度地扩大运动的总传播范围,作为回合的回报。在这个环境中,我们使用不同的基线方法来直接调整 L-L 以显示它们本身的模型。