Distributed Dynamic No-Regret Learning in Two-Network Zero-Sum Games

  • Lan Liao
  • , Deming Yuan*
  • , Daniel W. C. Ho
  • , Wei Xing Zheng
  • , Baoyong Zhang
  • , Zhan Yu
  • *Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

Abstract

This paper considers the time-varying zero-sum game between two multi-agent networks with different topologies. The two networks are modeled as adversarial players, where agents within each network communicate with their local neighbors while simultaneously gathering information from the opposing network through a bipartite interaction framework. The payoffs of the agents can be quantified by time-varying cost functions, and the objective is to design an online distributed algorithm to optimize the payoff for each network in the game. We use dynamic Nash equilibrium regret and duality gap as performance metrics and propose a projection-free algorithm called Online Distributed Frank-Wolfe in Two-Network (ODFW-TN). We assess the convergence of Algorithm ODFW-TN through regret analysis, and establish sublinear bounds for dynamic Nash equilibrium regret and duality gap of Algorithm ODFW-TN with respect to T, where T is the total number of games. Moreover, we validate the effectiveness of Algorithm ODFW-TN via a numerical experiment of a time-varying bilinear matrix game.
Original languageEnglish
Pages (from-to)1-15
Number of pages15
JournalIEEE Transactions on Automatic Control
Volume71
Issue number1
DOIs
Publication statusE-pub ahead of print - 8 Jul 2025

User-Defined Keywords

  • Frank-Wolfe
  • no-regret learning
  • online distributed optimization
  • zero-sum game
  • nash equilibrium

Fingerprint

Dive into the research topics of 'Distributed Dynamic No-Regret Learning in Two-Network Zero-Sum Games'. Together they form a unique fingerprint.

Cite this