Convex nondifferentiable stochastic optimization: A local randomized smoothing technique

Farzad Yousefian, Angelia Nedić, Uday V. Shanbhag

Research output: Chapter in Book/Report/Conference proceedingConference contribution

6 Scopus citations

Abstract

We consider a class of stochastic nondifferentiable optimization problems where the objective function is an expectation of a random convex function, that is not necessarily differentiable. We propose a local smoothing technique, based on random local perturbations of the objective function, that lead to differentiable approximations of the function. Under the assumption that the local randomness originates from a uniform distribution, we establish a Lipschitzian property for the gradient of the approximation. This facilitates the development of a stochastic approximation framework, which now requires sampling in the product space of the original measure and the artificially introduced distribution. We show that under suitable assumptions, the resulting diminishing steplength stochastic subgradient algorithm, with two samples per iteration, converges to an optimal solution of the problem when the subgradients are bounded.

Original languageEnglish (US)
Title of host publicationProceedings of the 2010 American Control Conference, ACC 2010
PublisherIEEE Computer Society
Pages4875-4880
Number of pages6
ISBN (Print)9781424474264
DOIs
StatePublished - 2010
Externally publishedYes

Publication series

NameProceedings of the 2010 American Control Conference, ACC 2010

Keywords

  • Local smoothing technique
  • Nondifferentiable convex minimization
  • Stochastic approximation method

ASJC Scopus subject areas

  • Control and Systems Engineering

Fingerprint

Dive into the research topics of 'Convex nondifferentiable stochastic optimization: A local randomized smoothing technique'. Together they form a unique fingerprint.

Cite this