Mini-Batch Stochastic Three-Operator Splitting for Distributed Optimization

B. Franci*, M. Staudigl

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

We consider a network of agents, each with its own private cost consisting of a sum of two possibly nonsmooth convex functions, one of which is composed with a linear operator. At every iteration each agent performs local calculations and can only communicate with its neighbors. The challenging aspect of our study is that the smooth part of the private cost function is given as an expected value and agents only have access to this part of the problem formulation via a heavy-tailed stochastic oracle. To tackle such sampling-based optimization problems, we propose a stochastic extension of the triangular pre-conditioned primal-dual algorithm. We demonstrate almost sure convergence of the scheme and validate the performance of the method via numerical experiments.
Original languageEnglish
Pages (from-to)2882-2887
Number of pages6
JournalIEEE Control Systems Letters
Volume6
DOIs
Publication statusPublished - 2022

Keywords

  • Random variables
  • Convergence
  • Stochastic processes
  • Approximation algorithms
  • Machine learning
  • Costs
  • Cost function
  • Stochastic systems
  • optimization
  • VARIANCE REDUCTION
  • CONSTRAINTS

Cite this