awesome-vla-for-ad and Awesome-Multimodal-LLM

awesome-vla-for-ad
50
Established
Maintenance 10/25
Adoption 10/25
Maturity 15/25
Community 15/25
Maintenance 0/25
Adoption 8/25
Maturity 16/25
Community 11/25
Stars: 331
Forks: 31
Downloads:
Commits (30d): 0
Language: HTML
License: MIT
Stars: 69
Forks: 7
Downloads:
Commits (30d): 0
Language:
License: MIT
No Package No Dependents
Stale 6m No Package No Dependents

About awesome-vla-for-ad

worldbench/awesome-vla-for-ad

🌐 Vision-Language-Action Models for Autonomous Driving: Past, Present, and Future

This project offers a comprehensive survey of Vision-Language-Action (VLA) models for autonomous driving. It explains how these models integrate real-world visual data and natural language commands to produce driving actions, moving beyond traditional, error-prone modular systems. Robotics engineers and researchers in autonomous vehicle development would use this to understand the current state and future directions of AI-driven self-driving systems.

autonomous-driving robotics self-driving-cars vision-systems vehicle-intelligence

About Awesome-Multimodal-LLM

vincentlux/Awesome-Multimodal-LLM

Reading list for Multimodal Large Language Models

This is a reading list for anyone deeply involved in or studying advanced AI, specifically focusing on how large language models (LLMs) can process and understand multiple types of data, like images, video, and text, simultaneously. It provides a structured collection of the latest academic papers, tutorials, and datasets in the field. Researchers, academics, and AI practitioners looking to stay current or explore specific areas within multimodal AI would use this resource.

AI-research machine-learning-engineering natural-language-processing computer-vision multimodal-AI

Scores updated daily from GitHub, PyPI, and npm data. How scores work