標題: Distributed Sub-gradient Algorithms with Limited Communications
作者: Rini, Stefano
Rao, Milind
Goldsmith, Andrea
交大名義發表
National Chiao Tung University
關鍵字: Distributed convex optimization;Decentralized methods;Gradient descent methods;Dual averaging;Nesterov gradient acceleration;Random projections
公開日期: 1-Jan-2019
摘要: We consider the distributed convex optimization scenario in which nodes in a network collectively find the minimum of a function utilizing only local communications and computations. Various sub-gradient algorithms have been developed for this optimization setting for the case in which the global function factorizes as the sum of local functions to be distributed to the nodes in network, including the distributed (i) online gradient descent, (ii) Nesterov gradient descent, and (iii) dual averaging algorithms. Generally speaking, these subgradient algorithms assume that, in each communication round, nodes can exchange messages of size comparable to that of the optimization variable. For many high-dimensional optimization problems, this communication requirement is beyond the capabilities of the communication network supporting the distributed optimization. For this reason, we propose a dimensionality reduction technique based on random projections to adapt these distributed optimization algorithms to networks with communication links of limited capacity. In particular, we analyze the performance of the proposed dimensionality reduction technique for the three algorithms above, i.e. (i)-(iii). Numerical simulations are presented that illustrate the performance of the proposed approach for these three algorithms.
URI: http://hdl.handle.net/11536/155066
ISBN: 978-1-7281-4300-2
ISSN: 1058-6393
期刊: CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS
起始頁: 2171
結束頁: 2175
Appears in Collections:Conferences Paper