-
RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Paper • 2312.00849 • Published • 8 -
RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness
Paper • 2405.17220 • Published -
ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation
Paper • 2304.05977 • Published • 1 -
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Experts
Paper • 2406.12845 • Published • 1
Robert Wijaya
wijayarobert
AI & ML interests
Computer Vision
Organizations
None yet
Collections
1
models
3
datasets
None public yet