### Abstract

This paper deals with iterative gradient and subgradient methods with random feasibility steps for solving constrained convex minimization problems, where the constraint set is specified as the intersection of possibly infinitely many constraint sets. Each constraint set is assumed to be given as a level set of a convex but not necessarily differentiable function. The proposed algorithms are applicable to the situation where the whole constraint set of the problem is not known in advance, but it is rather learned in time through observations. Also, the algorithms are of interest for constrained optimization problems where the constraints are known but the number of constraints is either large or not finite. We analyze the proposed algorithm for the case when the objective function is differentiable with Lipschitz gradients and the case when the objective function is not necessarily differentiable. The behavior of the algorithm is investigated both for diminishing and non-diminishing stepsize values. The almost sure convergence to an optimal solution is established for diminishing stepsize. For non-diminishing stepsize, the error bounds are established for the expected distances of the weighted averages of the iterates from the constraint set, as well as for the expected sub-optimality of the function values along the weighted averages.

Original language | English (US) |
---|---|

Pages (from-to) | 225-253 |

Number of pages | 29 |

Journal | Mathematical Programming |

Volume | 129 |

Issue number | 2 |

DOIs | |

State | Published - Oct 1 2011 |

Externally published | Yes |

### Keywords

- Convex minimization
- Error bounds
- Gradient algorithms
- Random algorithms
- Subgradient algorithms

### ASJC Scopus subject areas

- Software
- Mathematics(all)