mcp-client-for-ollama and mcp-ollama-python
About mcp-client-for-ollama
jonigl/mcp-client-for-ollama
A text-based user interface (TUI) client for interacting with MCP servers using Ollama. Features include agent mode, multi-server, model switching, streaming responses, tool management, human-in-the-loop, thinking mode, model params config, MCP prompts, custom system prompt and saved preferences. Built for developers working with local LLMs.
This interactive terminal application helps developers connect their local Large Language Models (LLMs) running on Ollama to external tools and services defined by the Model Context Protocol (MCP). It allows for real-time management of models, tools, and server connections, facilitating advanced LLM-driven automation and experimentation. Developers can input natural language prompts, and the system outputs responses, potentially enhanced by tool calls, all within a user-friendly text interface.
About mcp-ollama-python
pblagoje/mcp-ollama-python
Ollama MCP Server, let's you chat with your local LLM running on your local OLLAMA server.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work