Command Palette
Search for a command to run...
RLAIF-V: Open-Source AI Feedback Leads to Super GPT-4V Trustworthiness

Abstract
Traditional feedback learning for hallucination reduction relies on labor-intensive manual labeling or expensive proprietary models. This leaves the community without foundational knowledge about how to build high-quality feedback with open-source MLLMs. In this work, we introduce RLAIF-V, a novel framework that aligns MLLMs in a fully open-source paradigm. RLAIF-V maximally explores open-source MLLMs from two perspectives, including high-quality feedback data generation for preference learning and self-feedback guidance for inference-time scaling. Extensive experiments on six benchmarks in both automatic and human evaluation show that RLAIF-V substantially enhances the trustworthiness of models at both preference learning and inference time. RLAIF-V 7B reduces object hallucination by 80.7\% and overall hallucination by 33.7\%. Remarkably, RLAIF-V 12B further reveals the self-alignment potential of open-source MLLMs, where the model can learn from feedback of itself to achieve super GPT-4V trustworthiness.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| image-captioning-on-object-halbench | RLAIF-V 12B | chair_i: 1.8 chair_s: 3.3 |
| image-captioning-on-object-halbench | RLAIF-V 7B | chair_i: 4.3 chair_s: 8.5 |
| visual-question-answering-on-amber | RLAIF-V 12B | Accuracy: 88 F1: 90.9 |
| visual-question-answering-on-mmhal-bench | RLAIF-V 7B | Hallucination Rate: 29.2 Score: 3.06 |
| visual-question-answering-on-mmhal-bench | RLAIF-V 12B | Hallucination Rate: 29.2 Score: 3.36 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.