dorarad/gansformer
Generative Adversarial Transformers
Combines bipartite transformer architecture with multiplicative integration for efficient high-resolution image synthesis, enabling long-range spatial interactions with linear computational complexity. The model alternates information flow between latent variables and visual features to encourage compositional object representations, generalizing StyleGAN's region-based modulation approach. Supports both TensorFlow and PyTorch implementations with pretrained weights for resolutions up to 1024×1024, trained convergence in 5-7x fewer steps than StyleGAN2.
1,346 stars. No commits in the last 6 months.
Stars
1,346
Forks
151
Language
Python
License
MIT
Category
Last pushed
Jun 14, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/dorarad/gansformer"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
j-min/VL-T5
PyTorch code for "Unifying Vision-and-Language Tasks via Text Generation" (ICML 2021)
invictus717/MetaTransformer
Meta-Transformer for Unified Multimodal Learning
Yachay-AI/byt5-geotagging
Confidence and Byt5 - based geotagging model predicting coordinates from text alone.
zinengtang/TVLT
PyTorch code for “TVLT: Textless Vision-Language Transformer” (NeurIPS 2022 Oral)
OFA-Sys/OFASys
OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models