Online convex optimization is a framework where a learner sequentially queries an external data source in order to arrive at the optimal solution of a convex function. The paradigm has gained significant popularity recently thanks to its scalability in large-scale optimization and machine learning. The repeated interactions, however, expose the learner to privacy risks from eavesdropping adversary that observe the submitted queries. In this paper, we study how to optimally obfuscate the learner's queries in first-order online convex optimization, so that their learned optimal value is provably difficult to estimate for the eavesdropping adversary. We consider two formulations of learner privacy: a Bayesian formulation in which the convex function is drawn randomly, and a minimax formulation in which the function is fixed and the adversary's probability of error is measured with respect to a minimax criterion. We show that, if the learner wants to ensure the probability of accurate prediction by the adversary be kept below $1/L$, then the overhead in query complexity is additive in $L$ in the minimax formulation, but multiplicative in $L$ in the Bayesian formulation. Compared to existing learner-private sequential learning models with binary feedback, our results apply to the significantly richer family of general convex functions with full-gradient feedback. Our proofs are largely enabled by tools from the theory of Dirichlet processes, as well as more sophisticated lines of analysis aimed at measuring the amount of information leakage under a full-gradient oracle.
翻译:在线 convex 优化是一个框架, 学习者可以据此按顺序询问外部数据源, 以便找到最佳的剖面功能解决方案。 范例最近因其在大规模优化和机器学习中的伸缩性而获得显著受欢迎。 然而, 反复的交互作用使学习者暴露于监听提交查询的对手的隐私风险。 在本文中, 我们研究如何在第一阶在线平面优化中最好地模糊学习者询问的精密数据源, 以便其所学的最佳值很难为偷听对手估计最佳值。 我们考虑两种学习者隐私的配方: 一种是贝伊斯配方的配方, 其调控功能是随机绘制的, 以及一种小型模量的配方, 其功能是固定的, 其鼠标的误差概率是参照一个最小值标准。 我们的研究者希望确保对手准确预测的概率保持在1美元/ L$以下, 那么, 查询时的顶端复杂度在最小值中以 $ $ 美元 来添加 。 但是, 我们的理论性 将常规的精度分析结果应用到我们普通的直观模型, 的直观 学习中, 。