nkkbr/ViCA
This is the official implementation of ViCA2 (Visuospatial Cognitive Assistant 2), a multimodal large language model designed for advanced visuospatial reasoning. The repository also provides training scripts for the original ViCA model.
Stars
12
Forks
—
Language
Python
License
—
Category
Last pushed
Dec 15, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/nkkbr/ViCA"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
om-ai-lab/VLM-R1
Solve Visual Understanding with Reinforced VLMs
KimMeen/Time-LLM
[ICLR 2024] Official implementation of " 🦙 Time-LLM: Time Series Forecasting by Reprogramming...
bytedance/SALMONN
SALMONN family: A suite of advanced multi-modal LLMs
fixie-ai/ultravox
A fast multimodal LLM for real-time voice
NVlabs/OmniVinci
OmniVinci is an omni-modal LLM for joint understanding of vision, audio, and language.