no code implementations • 2 May 2024 • Sicong Wang, Kuo Gai, Shihua Zhang
Overall, this study extends NC to PFC to model the collapse phenomenon of intermediate layers and its dependence on the input data, shedding light on the theoretical understanding of ResNet in classification problems.
no code implementations • 18 Feb 2021 • Kuo Gai, Shihua Zhang
In a word, we conclude a mathematical principle of deep learning is to learn the geodesic curve in the Wasserstein space; and deep learning is a great engineering realization of continuous transformation in high-dimensional space.
no code implementations • 20 May 2020 • Kuo Gai, Shihua Zhang
Non-adversarial generative models such as variational auto-encoder (VAE), Wasserstein auto-encoders with maximum mean discrepancy (WAE-MMD), sliced-Wasserstein auto-encoder (SWAE) are relatively easy to train and have less mode collapse compared to Wasserstein auto-encoder with generative adversarial network (WAE-GAN).
no code implementations • 25 Nov 2019 • Chihao Zhang, Kuo Gai, Shihua Zhang
However, most of the existing methods only assume that the noise is correlated in the feature space while there may exist two-way structured noise.