Random algorithms for convex minimization problems

Research output: Contribution to journalArticle

50 Scopus citations

Abstract

This paper deals with iterative gradient and subgradient methods with random feasibility steps for solving constrained convex minimization problems, where the constraint set is specified as the intersection of possibly infinitely many constraint sets. Each constraint set is assumed to be given as a level set of a convex but not necessarily differentiable function. The proposed algorithms are applicable to the situation where the whole constraint set of the problem is not known in advance, but it is rather learned in time through observations. Also, the algorithms are of interest for constrained optimization problems where the constraints are known but the number of constraints is either large or not finite. We analyze the proposed algorithm for the case when the objective function is differentiable with Lipschitz gradients and the case when the objective function is not necessarily differentiable. The behavior of the algorithm is investigated both for diminishing and non-diminishing stepsize values. The almost sure convergence to an optimal solution is established for diminishing stepsize. For non-diminishing stepsize, the error bounds are established for the expected distances of the weighted averages of the iterates from the constraint set, as well as for the expected sub-optimality of the function values along the weighted averages.

Original languageEnglish (US)
Pages (from-to)225-253
Number of pages29
JournalMathematical Programming
Volume129
Issue number2
DOIs
StatePublished - Oct 1 2011
Externally publishedYes

Keywords

  • Convex minimization
  • Error bounds
  • Gradient algorithms
  • Random algorithms
  • Subgradient algorithms

ASJC Scopus subject areas

  • Software
  • Mathematics(all)

Fingerprint Dive into the research topics of 'Random algorithms for convex minimization problems'. Together they form a unique fingerprint.

  • Cite this