Rishikesh-Jadhav/Multi-Modal-Image-Generation-using-Grounding-DINO-SAM-and-Stable-Diffusionv2
A project to combine Grounding-DINO with Meta AI's Segment Anything Model (SAM) and Stable Diffusion for image manipulation using prompts. The plan is to integrate these techniques and deploy the model on Hugging Face with a Gradio interface for users to detect, segment regions and inpaint them in images.
No commits in the last 6 months.
Stars
5
Forks
1
Language
Jupyter Notebook
License
—
Category
Last pushed
Oct 31, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/Rishikesh-Jadhav/Multi-Modal-Image-Generation-using-Grounding-DINO-SAM-and-Stable-Diffusionv2"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
huggingface/diffusers
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch.
bghira/SimpleTuner
A general fine-tuning kit geared toward image/video/audio diffusion models.
mcmonkeyprojects/SwarmUI
SwarmUI (formerly StableSwarmUI), A Modular Stable Diffusion Web-User-Interface, with an...
nateraw/stable-diffusion-videos
Create 🔥 videos with Stable Diffusion by exploring the latent space and morphing between text prompts
TheDesignFounder/DreamLayer
Benchmark diffusion models faster. Automate evals, seeds, and metrics for reproducible results.