Awesome-Large-Vision-Language-Model and Awesome-Multimodal-LLM

Maintenance 2/25
Adoption 8/25
Maturity 16/25
Community 10/25
Maintenance 0/25
Adoption 8/25
Maturity 16/25
Community 11/25
Stars: 42
Forks: 4
Downloads:
Commits (30d): 0
Language:
License: MIT
Stars: 69
Forks: 7
Downloads:
Commits (30d): 0
Language:
License: MIT
Stale 6m No Package No Dependents
Stale 6m No Package No Dependents

About Awesome-Large-Vision-Language-Model

SuperBruceJia/Awesome-Large-Vision-Language-Model

Awesome Large Vision-Language Model: A Curated List of Large Vision-Language Model

This resource provides a curated collection of materials for anyone exploring or working with large vision-language models, including medical foundation models. It centralizes key papers, presentations, books, and benchmarks related to integrating visual and linguistic data. Researchers and AI practitioners focused on developing or applying advanced AI systems that understand and process both images and text will find this helpful.

AI-research computer-vision natural-language-processing multimodal-AI medical-AI

About Awesome-Multimodal-LLM

vincentlux/Awesome-Multimodal-LLM

Reading list for Multimodal Large Language Models

This is a reading list for anyone deeply involved in or studying advanced AI, specifically focusing on how large language models (LLMs) can process and understand multiple types of data, like images, video, and text, simultaneously. It provides a structured collection of the latest academic papers, tutorials, and datasets in the field. Researchers, academics, and AI practitioners looking to stay current or explore specific areas within multimodal AI would use this resource.

AI-research machine-learning-engineering natural-language-processing computer-vision multimodal-AI

Scores updated daily from GitHub, PyPI, and npm data. How scores work