Generative models have received considerable interest in modern machine learning and statistics as a method for data generation and representation learning. Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs) are the two important classes of implicit generative modeling methods, which model the transformation between the latent variable and data variable to simulate the sampling process without specifying probability distributions explicitly. Owing to the recent development of deep learning, generative models have yielded remarkable empirical performance in a wide range of applications. Despite the empirical success of generative models, their theoretical properties were less justified, especially those of GANs. This motivates the first thrust of this thesis, which is statistical analysis of f-divergence GANs. Our theory gives rise to a new class of GAN algorithms with higher statistical efficiency and sheds light on the statistical problems including the relationship between the modern algorithm (GAN) and the classical method (maximum likelihood estimation) as well as how various f-divergences behave. We also provide a unified view of GAN and VAE under the principled framework of bidirectional generative models. In addition, we extensively adapt our proposed methods to practical tasks in computer vision and natural language processing and achieve state-of-the-art performance. In particular, we present a new model architecture and learning formulation based on our efficient GAN approach for coherent and diverse text generation. Structures are pervasive and inherent in human’s recognition and understanding of the real world. The second part of this thesis shifts the focus to the structural properties of generative models. An emerging field regarding this is disentangled representation learning that starts with the premise that real-world data is generated by a few explanatory factors and aims at recovering the generative factors as well as their underlying structure. Disentangled representations of data have numerous benefits in the interpretability of deep learning models, downstream learning tasks, and controllable generation. The difficulty of disentanglement depends on the amount of supervision available as well as the complexity of the underlying structures. It is acknowledged that disentanglement is impossible in a fully unsupervised setting. Existing disentanglement literature mostly considers simple structures such as independence or conditional independence given some observed auxiliary variables, while a more general (and challenging) structure is the causal structure where the underlying factors are connected by a causal graph. We formalize the failure of previous methods in the causal case and propose a method for disentangling the causal factors based on a bidirectional generative model with a causal prior. We provide theoretical justification on the identifiability and asymptotic convergence of the proposed algorithm. Finally, we develop a nonparametric method to learn causal structures from observational data.
| Date of Award | 2022 |
|---|
| Original language | English |
|---|
| Awarding Institution | - The Hong Kong University of Science and Technology
|
|---|
| Supervisor | Tong ZHANG (Supervisor) & Kani CHEN (Supervisor) |
|---|
Statistical and structural properties of generative models
SHEN, X. (Author). 2022
Student thesis: Doctoral thesis