-
RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Paper • 2312.00849 • Published • 8 -
RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness
Paper • 2405.17220 • Published -
ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation
Paper • 2304.05977 • Published • 1 -
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-Experts
Paper • 2406.12845 • Published • 1
Robert Wijaya
wijayarobert
AI & ML interests
Computer Vision
Recent Activity
liked
a Space
about 1 month ago
librarian-bots/recommend_similar_papers
liked
a model
about 2 months ago
lmms-lab/llava-onevision-qwen2-7b-ov
Organizations
None yet
Collections
1
models
3
datasets
None public yet