Advances in Generative Models

The field of generative models is moving towards improving the quality and diversity of generated images and texts. Recent developments focus on enhancing variational autoencoders (VAEs) and diffusion models to achieve better performance and robustness. Notably, researchers are exploring new architectures and techniques to preserve identity consistency and prompt diversity in generated images. Additionally, there is a growing interest in developing models that can generate high-quality images from limited datasets.

Some noteworthy papers in this area include: The Multivariate Variational Autoencoder, which improves reconstruction and calibration by preserving Gaussian tractability and lifting the diagonal posterior restriction. Taming Identity Consistency and Prompt Diversity in Diffusion Models via Latent Concatenation and Masked Conditional Flow Matching, which proposes a LoRA fine-tuned diffusion model to achieve robust identity preservation without architectural modifications. oboro: Text-to-Image Synthesis on Limited Data using Flow-based Diffusion Transformer with MMH Attention, which develops an image generation model from scratch using only copyright-cleared images for training. Improving Conditional VAE with approximation using Normalizing Flows, which explores image generation with conditional Variational Autoencoders to incorporate desired attributes within the images.

Sources

Multivariate Variational Autoencoder

Taming Identity Consistency and Prompt Diversity in Diffusion Models via Latent Concatenation and Masked Conditional Flow Matching

oboro: Text-to-Image Synthesis on Limited Data using Flow-based Diffusion Transformer with MMH Attention

Improving Conditional VAE with approximation using Normalizing Flows

Built with on top of