WangRongsheng/XrayGLM
🩺 首个会看胸部X光片的中文多模态医学大模型 | The first Chinese Medical Multimodal Model that Chest Radiographs Summarization.
Builds on VisualGLM-6B with LoRA fine-tuning using curated Chinese chest X-ray diagnostic datasets (MIMIC-CXR and OpenI translated via ChatGPT), enabling multi-turn medical dialogue beyond image captioning. Supports inference via CLI and WebUI, with training reproducible on A100 GPUs using the provided shell scripts and preprocessed report pairs.
1,066 stars. No commits in the last 6 months.
Stars
1,066
Forks
143
Language
Python
License
—
Category
Last pushed
Nov 20, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/WangRongsheng/XrayGLM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
Event-AHU/Medical_Image_Analysis
Foundation models based medical image analysis
cambridgeltl/visual-med-alpaca
Visual Med-Alpaca is an open-source, multi-modal foundation model designed specifically for the...
canyuchen/ClinicalBench
Code for the paper "ClinicalBench: Can LLMs Beat Traditional ML Models in Clinical Prediction?"
duyhominhnguyen/Exgra-Med
[NeurIPS 2025] ExGra-Med: Medical Multi-Modal LLM with Extended Context Alignment
synlp/R2-LLM
The official GitHub repository of the AAAI-2024 paper "Bootstrapping Large Language Models for...