✨✨Latest Advances on Multimodal Large Language Models
-
Updated
Jun 19, 2025
✨✨Latest Advances on Multimodal Large Language Models
[NeurIPS 2024] An official implementation of ShareGPT4Video: Improving Video Understanding and Generation with Better Captions
[ICML2024 (Oral)] Official PyTorch implementation of DoRA: Weight-Decomposed Low-Rank Adaptation
✨✨[CVPR 2025] Video-MME: The First-Ever Comprehensive Evaluation Benchmark of Multi-modal LLMs in Video Analysis
🔥🔥🔥 A curated list of papers on LLMs-based multimodal generation (image, video, 3D and audio).
The Paper List of Large Multi-Modality Model (Perception, Generation, Unification), Parameter-Efficient Finetuning, Vision-Language Pretraining, Conventional Image-Text Matching for Preliminary Insight.
Resources and paper list for "Thinking with Images for LVLMs". This repository accompanies our survey on how LVLMs can leverage visual information for complex reasoning, planning, and generation.
Curated papers on Large Language Models in Healthcare and Medical domain
[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models
[ECCV 2024] ShareGPT4V: Improving Large Multi-modal Models with Better Captions
A curated list of recent and past chart understanding work based on our IEEE TKDE survey paper: From Pixels to Insights: A Survey on Automatic Chart Understanding in the Era of Large Foundation Models.
[NeurIPS 2024] This repo contains evaluation code for the paper "Are We on the Right Way for Evaluating Large Vision-Language Models"
up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resources
Talk2BEV: Language-Enhanced Bird's Eye View Maps (ICRA'24)
GeoPixel: A Pixel Grounding Large Multimodal Model for Remote Sensing is specifically developed for high-resolution remote sensing image analysis, offering advanced multi-target pixel grounding capabilities.
[ECCV 2024] API: Attention Prompting on Image for Large Vision-Language Models
A curated collection of resources focused on the Mechanistic Interpretability (MI) of Large Multimodal Models (LMMs). This repository aggregates surveys, blog posts, and research papers that explore how LMMs represent, transform, and align multimodal information internally.
This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strategy.
[ICML 2024] Safety Fine-Tuning at (Almost) No Cost: A Baseline for Vision Large Language Models.
✨A curated list of papers on the uncertainty in multi-modal large language model (MLLM).
Add a description, image, and links to the large-vision-language-models topic page so that developers can more easily learn about it.
To associate your repository with the large-vision-language-models topic, visit your repo's landing page and select "manage topics."