TY - GEN
T1 - Simultaneous Gradient Descent-Ascent for GANs Minimax Optimization using Sinkhorn Divergence
AU - Adnan, Risman
AU - Adi Saputra, Muchlisin
AU - Fadlil, Junaidillah
AU - Iqbal, Muhamad
AU - Basaruddin, Tjan
N1 - Publisher Copyright:
© 2020 ACM.
PY - 2020/12/12
Y1 - 2020/12/12
N2 - The Sinkhorn divergence, a smooth and symmetric normalization version of entropy-regularized optimal transport (EOT) is a promising tool for Generative Adversarial Networks (GANs). However, understanding the dynamic of gradient algorithms for Sinkhorn-based GANs remains a big challenge. In this work, we consider the GANs minimax optimization problem using Sinkhorn divergence, in which smoothness and convexity properties of the objective function are critical factors for convergence and stability. We prove that GANs with convex-concave Sinkhorn divergence can converge to local Nash equilibrium using first-order simultaneous stochastic gradient descent-ascent (SimSGDA) algorithm under certain approximations. We further present a nonasymptotic analysis for the convergence rate of the SimSGDA using structural similarity index measure (SSIM). Our experiments suggest a convergence rate proportional to the inverse number of SGDA iterations tested on tiny-colored datasets (Cats and CelebA) and advanced neural architectures (DCGAN and ResNet). We demonstrate that SSIM is potential tool to measure convergence rate of the SimSGDA algorithm empirically.
AB - The Sinkhorn divergence, a smooth and symmetric normalization version of entropy-regularized optimal transport (EOT) is a promising tool for Generative Adversarial Networks (GANs). However, understanding the dynamic of gradient algorithms for Sinkhorn-based GANs remains a big challenge. In this work, we consider the GANs minimax optimization problem using Sinkhorn divergence, in which smoothness and convexity properties of the objective function are critical factors for convergence and stability. We prove that GANs with convex-concave Sinkhorn divergence can converge to local Nash equilibrium using first-order simultaneous stochastic gradient descent-ascent (SimSGDA) algorithm under certain approximations. We further present a nonasymptotic analysis for the convergence rate of the SimSGDA using structural similarity index measure (SSIM). Our experiments suggest a convergence rate proportional to the inverse number of SGDA iterations tested on tiny-colored datasets (Cats and CelebA) and advanced neural architectures (DCGAN and ResNet). We demonstrate that SSIM is potential tool to measure convergence rate of the SimSGDA algorithm empirically.
KW - Convex Optimization
KW - Entropy Regularization
KW - Generative Adversarial Networks
KW - Generative Models
KW - Gradient Descent-Ascent
KW - Minimax Game
KW - Optimal Transport
KW - Sinkhorn Divergence
KW - Stochastic Gradient Descent-Ascent
UR - http://www.scopus.com/inward/record.url?scp=85109532115&partnerID=8YFLogxK
U2 - 10.1145/3448326.3448328
DO - 10.1145/3448326.3448328
M3 - Conference contribution
AN - SCOPUS:85109532115
T3 - ACM International Conference Proceeding Series
SP - 6
EP - 17
BT - 2020 2nd International Conference on Artificial Intelligence, Robotics and Control, AIRC 2020
PB - Association for Computing Machinery
T2 - 2nd International Conference on Artificial Intelligence, Robotics and Control, AIRC 2020
Y2 - 12 December 2020 through 14 December 2020
ER -