
Researchers: Deniz Ustebay (Ph.D. Student), Boris Oreshkin (Ph.D. Student), Prof. Mark Coates & Prof. Mike Rabbat
Description:
Distributed consensus has been identified as a canonical problem in both the distributed signal processing and control communities. The prototypical example of a consensus problem is computation of the average consensus: in a network, initially each node has a scalar data value and the goal is to find a distributed algorithm that asymptotically computes the average at every node. Such an algorithm can further be used for computing linear functions of the data and can be generalized for averaging vectorial data.
In this study we propose a new randomized gossip algorithm, greedy gossip with eavesdropping (GGE), for the computation of average consensus. Unlike previous randomized gossip algorithms, which perform updates completely at random, GGE takes advantage of the broadcast nature of wireless communications and implements a greedy neighbor selection procedure. We assume a broadcast transmission model such that all neighbors within range of a transmitting node receive the message. Thereby, in addition to keeping track of its own value, each node tracks its neighbors values by eavesdropping on their transmissions. At each iteration, the activated node uses this information to greedily choose the neighbor with which it will gossip, selecting the neighbor whose value is most different from its own. Accelerating convergence in this myopic way does not bias computation and does not rely on geographic location information, which may change in networks of mobile nodes.
Although GGE is a powerful yet simple variation on gossipstyle algorithms, analyzing its convergence behavior is nontrivial. The main reason is that each GGE update depends explicitly on the values at each node (via the greedy decision of with which neighbor to gossip). Thus, the standard approach to proving convergence to the average consensus solution (i.e., expressing updates in terms of a linear recursion and then imposing properties on this recursion) cannot be applied to guarantee convergence of GGE. To prove convergence, we demonstrate that GGE updates correspond to iterations of a distributed randomized incremental subgradient optimization algorithm. Similarly, analysis of the convergence rate of GGE requires a different approach than the standard approach of examining the mixing time of a related Markov chain. We develop a bound relating the rate of convergence of GGE to the rate of standard randomized gossip. The bound indicates that GGE always converges faster than randomized gossip, a finding supported by simulation results. We also provide a worstcase bound on the rate of convergence of GGE. For other gossip algorithms the rate of convergence is generally characterized as a function of the second largest eigenvalue of a related stochastic matrix. In the case of GGE, our worstcase bound characterizes the rate of convergence in terms of a constant that is completely determined by the network topology. We investigate the behavior of this constant empirically for random geometric graph topologies and derive lower bounds that provide some characterization of its scaling properties.
We also show that the improvement achieved using GGE over standard randomized gossip (i.e., exchanging information equally often with all neighbors) is proportional to the maximum node degree. Thus, for network topologies such as random geometric graphs, where node degree grows with the network size, the improvements of GGE scale with network size, but for gridlike topologies, where the node degree remains constant, GGE yields limited improvement. We propose an extension to GGE, which is called multihop GGE, that improves the rate of convergence for gridlike topologies. Multihop GGE relies on increasing artificially neighborhood size by performing greedy updates with nodes beyond one hop neighborhoods. We show that multihop GGE converges to the average consensus and illustrate via simulation that multihop GGE improves the performance of GGE for different network topologies.
Publications:
D. Ustebay, B. Oreshkin, M.J. Coates and, M. Rabbat, Multihop Greedy Gossip with Eavesdropping, to appear, in Proc. Int. Conf. on Information Fusion, Seatle, WA, USA, July 2009.
D. Ustebay, B. Oreshkin, M.J. Coates and, M. Rabbat, Greedy Gossip with Eavesdropping, Technical Report, Department of Electrical and Computer Engineering, McGill University, Montreal, QC, Canada, 2009.
D. Ustebay, B. Oreshkin, M. Coates and, M. Rabbat, The Speed of Greed: Characterizing Myopic Gossip through Network Voracity, in Proc. of IEEE Int. Conf. on Acoustics, Speech, and Signal Processing (ICASSP), Taipei, Taiwan, April 2009.
D. Ustebay, B. Oreshkin, M. Coates and, M. Rabbat, Rates of Convergence for Greedy Gossip with Eavesdropping, in Proceedings of the FortySixth Annual Allerton Conference on Communication, Control, and Computing, IL, USA, September 2008.
D. Ustebay, M. Coates and, M. Rabbat, Greedy Gossip with Eavesdropping, in Proceedings of the IEEE International Symposium on Wireless Pervasive Computing, Santorini, Greece, May 2008.
