When Personalization Meets Reality: A Multi-Faceted Analysis of...View PDF#Personalization#Preferences#Paper#PDF#Cohere·arxiv.org·today at 6:21 PMWhen Personalization Meets Reality: A Multi-Faceted Analysis of...
Beyond Preferences in AI Alignment#AI#Preferences#Alignment#Paper#pddf·arxiv.org·Sep 8, 2024Beyond Preferences in AI Alignment
Iterative Reasoning Preference Optimization#Reasoning#Preferences#Paper#PDF#Meta#Large Language Models#Algorithms#Chain of Thought·arxiv.org·May 1, 2024Iterative Reasoning Preference Optimization
Self-Rewarding Language ModelsDownload PDF#AI#Meta#Paper#PDF#Large Language Models#Preferences#Autonomous·arxiv.org·Jan 20, 2024Self-Rewarding Language Models
Diffusion Model Alignment Using Direct Preference OptimizationDownload PDF#Fine-Tuning#Stable Diffusion#Preferences#Large Language Models#Paper#PDF·arxiv.org·Nov 24, 2023Diffusion Model Alignment Using Direct Preference Optimization
The Jiminy Advisor: Moral Agreements among Stakeholders Based on Norms and Argumentation | Journal of Artificial Intelligence Research#Preferences#Reasoning#Stakeholders#Paper#PDF·jair.org·Jul 12, 2023The Jiminy Advisor: Moral Agreements among Stakeholders Based on Norms and Argumentation | Journal of Artificial Intelligence Research
Human Preference Score v2: A Solid Benchmark for Evaluating Human Preferences of Text-to-Image Synthesis#Text-to-Image#Benchmark#Preferences#Prompt Engineering#Paper#PDF·arxiv.org·Jul 5, 2023Human Preference Score v2: A Solid Benchmark for Evaluating Human Preferences of Text-to-Image Synthesis
Emotion prediction as computation over a generative theory of mind | Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences#Emotion#Prediction#Paper#PDF#Neuroscience#Computation#Preferences#Probability#Planning#Generative Grammar#Theory of Mind·royalsocietypublishing.org·Jun 7, 2023Emotion prediction as computation over a generative theory of mind | Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences
Direct Preference Optimization: Your Language Model is Secretly a Reward ModelPDF#Large Language Models#Preferences#Reward#Training#Paper#PDF·arxiv.org·Jun 4, 2023Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Pick-a-Pic: An Open Dataset of User Preferences for Text-to-Image Generation#Preferences#Datasets#Paper#PDF#Text-to-Image#Generative Models·arxiv.org·May 3, 2023Pick-a-Pic: An Open Dataset of User Preferences for Text-to-Image Generation