Advancements in Diffusion Models and Image Processing

The field of diffusion models and image processing is rapidly advancing, with a focus on improving efficiency, fidelity, and versatility. Recent developments have led to the creation of novel frameworks, such as cross-matrix Krylov projection methods, that accelerate score-based diffusion models and achieve significant time reductions. Additionally, advancements in periodicity-enforced neural networks and data-driven predictive modeling are enhancing the design of microfluidic devices for cancer detection and other applications. Furthermore, innovative approaches like UltraFlux and DeCo are pushing the boundaries of text-to-image generation and image compression, enabling high-quality image synthesis and efficient compression. Other notable advancements include the development of efficient pixel diffusion frameworks, frequency-decoupled pixel diffusion models, and hybrid convolution and frequency state space networks for image compression. These innovations are driving progress in various areas, including image generation, compression, and processing, and are expected to have a significant impact on the field. Noteworthy papers include Efficient Score Pre-computation for Diffusion Models via Cross-Matrix Krylov Projection, which achieves a 15.8% to 43.7% time reduction over standard sparse solvers, and Periodicity-Enforced Neural Network for Designing Deterministic Lateral Displacement Devices, which improves critical diameter error by 85.4% over baseline methods.

Sources

Efficient Score Pre-computation for Diffusion Models via Cross-Matrix Krylov Projection

Periodicity-Enforced Neural Network for Designing Deterministic Lateral Displacement Devices

Data-Driven Predictive Modeling of Microfluidic Cancer Cell Separation Using a Deterministic Lateral Displacement Device

UltraFlux: Data-Model Co-Design for High-quality Native 4K Text-to-Image Generation across Diverse Aspect Ratios

Versatile Recompression-Aware Perceptual Image Super-Resolution

Sphinx: Efficiently Serving Novel View Synthesis using Regression-Guided Selective Refinement

CoD: A Diffusion Foundation Model for Image Compression

DiP: Taming Diffusion Models in Pixel Space

DeCo: Frequency-Decoupled Pixel Diffusion for End-to-End Image Generation

Efficiency vs. Fidelity: A Comparative Analysis of Diffusion Probabilistic Models and Flow Matching on Low-Resource Hardware

Hybrid Convolution and Frequency State Space Network for Image Compression

TReFT: Taming Rectified Flow Models For One-Step Image Translation

FREE: Uncertainty-Aware Autoregression for Parallel Diffusion Transformers

PixelDiT: Pixel Diffusion Transformers for Image Generation

Inversion-Free Style Transfer with Dual Rectified Flows

RAVQ-HoloNet: Rate-Adaptive Vector-Quantized Hologram Compression

MobileI2V: Fast and High-Resolution Image-to-Video on Mobile Devices

Built with on top of