Arrow Research search
Back to IJCAI

IJCAI 2017

Efficient Private ERM for Smooth Objectives

Conference Paper Multidisciplinary Topics and Applications Artificial Intelligence

Abstract

In this paper, we consider efficient differentially private empirical risk minimization from the viewpoint of optimization algorithms. For strongly convex and smooth objectives, we prove that gradient descent with output perturbation not only achieves nearly optimal utility, but also significantly improves the running time of previous state-of-the-art private optimization algorithms, for both $\epsilon$-DP and $(\epsilon, \delta)$-DP. For non-convex but smooth objectives, we propose an RRPSGD (Random Round Private Stochastic Gradient Descent) algorithm, which provably converges to a stationary point with privacy guarantee. Besides the expected utility bounds, we also provide guarantees in high probability form. Experiments demonstrate that our algorithm consistently outperforms existing method in both utility and running time.

Authors

Keywords

  • Machine Learning: Machine Learning
  • Multidisciplinary Topics and Applications: AI&Security and Privacy

Context

Venue
International Joint Conference on Artificial Intelligence
Archive span
1969-2025
Indexed papers
14525
Paper id
216384536937534130