AmberSahdev/Open-Interface
Control Any Computer Using LLMs.
Implements a vision-based automation loop where LLM backends (GPT-4o, Gemini) analyze screenshots to plan actions, then execute them via keyboard and mouse simulation with real-time course correction. Runs as a cross-platform desktop application (macOS, Linux, Windows) requiring Accessibility and Screen Recording permissions to capture visual feedback and control system input devices.
2,630 stars.
Stars
2,630
Forks
265
Language
Python
License
GPL-3.0
Category
Last pushed
Feb 25, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/AmberSahdev/Open-Interface"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
digiteinfotech/kairon
Agentic AI platform that harnesses Visual LLM Chaining to build proactive digital assistants
withcatai/catai
Run AI ✨ assistant locally! with simple API for Node.js 🚀
bigsk1/voice-chat-ai
🎙️ Speak with AI - Run locally using Ollama, OpenAI, Anthropic or xAI - Speech uses SparkTTS,...
syxanash/maxheadbox
Tiny truly local voice-activated LLM Agent that runs on a Raspberry Pi
second-state/echokit_server
Open Source Voice Agent Platform