LLM-groundedDiffusion and LLM-groundedVideoDiffusion
About LLM-groundedDiffusion
TonyLianLong/LLM-groundedDiffusion
LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models (LLM-grounded Diffusion: LMD, TMLR 2024)
This project helps graphic designers, marketers, and artists create highly specific images from text descriptions. You provide a detailed text prompt, and the system intelligently interprets it to generate an image that precisely matches your vision, including the placement of objects. The output is a high-quality image that accurately reflects your instructions.
About LLM-groundedVideoDiffusion
TonyLianLong/LLM-groundedVideoDiffusion
[ICLR 2024] LLM-grounded Video Diffusion Models (LVD): official implementation for the LVD paper
This project helps video creators and content producers generate short video clips from detailed text descriptions. You input a text prompt, and the system uses an AI planner to create a scene layout with objects and their movements. This layout then guides a video generator to produce a video that accurately reflects your prompt's specifics, making it easier to visualize and create dynamic scenes.
Scores updated daily from GitHub, PyPI, and npm data. How scores work