-
MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts
Paper • 2407.21770 • Published • 19 -
VILA^2: VILA Augmented VILA
Paper • 2407.17453 • Published • 38 -
The Synergy between Data and Multi-Modal Large Language Models: A Survey from Co-Development Perspective
Paper • 2407.08583 • Published • 10 -
Vision language models are blind
Paper • 2407.06581 • Published • 80
rainningXY
xxyyy123
AI & ML interests
None yet
Organizations
None yet
Collections
3
-
Internal Consistency and Self-Feedback in Large Language Models: A Survey
Paper • 2407.14507 • Published • 44 -
New Desiderata for Direct Preference Optimization
Paper • 2407.09072 • Published • 8 -
Self-Recognition in Language Models
Paper • 2407.06946 • Published • 24 -
MJ-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation?
Paper • 2407.04842 • Published • 52
models
58
xxyyy123/1701221123_Ads_Mistral7B-slimorca_all-Lqv-r4b128
Updated
xxyyy123/final_submit_v3_1700124777
Updated
•
1
xxyyy123/final_submit_v3_1700105042
Updated
xxyyy123/final_submit_v3_1700104898
Updated
xxyyy123/final_submit_v3_1700097826
Updated
xxyyy123/final_submit_v3_1700097805
Updated
xxyyy123/final_submit_v3_1700097607
Updated
xxyyy123/final_submit_v3_1700097567
Updated
xxyyy123/final_submit_v3_1700065328
Updated
•
1
xxyyy123/final_submit_v3_1700065217
Updated
datasets
None public yet