Awesome-Multimodal-Large-Language-Models and Awesome-Multimodal-LLM-Autonomous-Driving
These two tools are ecosystem siblings, where B is a specialized application of the broader field surveyed by A, specifically focusing on multimodal large language models within the autonomous driving domain.
About Awesome-Multimodal-Large-Language-Models
BradyFU/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
Comprehensive curated repository of research papers, datasets, and benchmarks covering multimodal LLM advances across instruction tuning, hallucination mitigation, and reasoning tasks. Features proprietary evaluation frameworks (MME, Video-MME, MME-RealWorld) and the VITA series of omni-modal models supporting real-time vision-speech interaction and embodied reasoning. Targets the broader MLLM research ecosystem with extensive documentation of 750+ references and curated resources for model development and evaluation.
About Awesome-Multimodal-LLM-Autonomous-Driving
IrohXu/Awesome-Multimodal-LLM-Autonomous-Driving
[WACV 2024 Survey Paper] Multimodal Large Language Models for Autonomous Driving
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work