Advances in Personalized Preference Learning and Recommendation Systems

The field of personalized preference learning and recommendation systems is rapidly evolving, with a focus on developing innovative methods to capture diverse human preferences and improve decision-making in complex environments. Recent developments have highlighted the importance of adaptive reward modeling, context-aware routing, and mixture modeling to enhance personalized preference learning. Additionally, there is a growing interest in integrating large language models into recommendation tasks, leveraging their strong semantic understanding and prompt flexibility.

Noteworthy papers in this area include ChARM, which proposes a character-based act-adaptive reward model for advanced role-playing language agents, and MiCRo, which introduces a two-stage framework for personalized preference learning using large-scale binary preference datasets. Other notable papers include Descriptive History Representations, which focuses on learning representations by answering questions, and Towards Human-like Preference Profiling in Sequential Recommendation, which proposes a preference optimization framework to emulate human-like prioritization in sequential recommendation.

Sources

ChARM: Character-based Act-adaptive Reward Modeling for Advanced Role-Playing Language Agents

MiCRo: Mixture Modeling and Context-aware Routing for Personalized Preference Learning

Descriptive History Representations: Learning Representations by Answering Questions

Towards Human-like Preference Profiling in Sequential Recommendation

TransAct V2: Lifelong User Action Sequence Modeling on Pinterest Recommendation

MidPO: Dual Preference Optimization for Safety and Helpfulness in Large Language Models via a Mixture of Experts Framework

MMM4Rec: An Transfer-Efficient Framework for Multi-modal Sequential Recommendation

Optimization of Epsilon-Greedy Exploration

Learning Fair And Effective Points-Based Rewards Programs

User Altruism in Recommendation Systems

Reason-to-Recommend: Using Interaction-of-Thought Reasoning to Enhance LLM Recommendation

Built with on top of