Personalization and Evaluation in Text-to-Image Generation

The field of text-to-image generation is moving towards personalized and explainable models that can align with human perception and preferences. Researchers are developing new methods for evaluating and optimizing generated images to better match individual user tastes. This includes the introduction of datasets and benchmarks that capture diverse user preferences and the development of models that can dynamically generate user-conditioned evaluation dimensions. Noteworthy papers in this area include IE-Critic-R1, which introduces a comprehensive and explainable quality assessment metric for text-driven image editing, and MagicWand, a universal generation and evaluation agent that enhances prompts based on user preferences. Additionally, PIGReward, a personalized reward model, and RubricRL, a simple and general framework for rubric-based reward design, demonstrate promising approaches to personalized text-to-image generation and evaluation.

Sources

IE-Critic-R1: Advancing the Explanatory Measurement of Text-Driven Image Editing for Human Perception Alignment

MagicWand: A Universal Agent for Generation and Evaluation Aligned with User Preference

Personalized Reward Modeling for Text-to-Image Generation

DesignPref: Capturing Personal Preferences in Visual Design Generation

RubricRL: Simple Generalizable Rewards for Text-to-Image Generation

Generative AI Compensates for Age-Related Cognitive Decline in Decision Making: Preference-Aligned Recommendations Reduce Choice Difficulty

Seeing Twice: How Side-by-Side T2I Comparison Changes Auditing Strategies

Built with on top of