mcp-server-atlassian-confluence and mcp-server-atlassian-jira
These tools are complements, as one provides AI access to Jira for project and issue management, while the other offers AI capabilities for Confluence space and page content, allowing a unified AI interface for both Atlassian products.
About mcp-server-atlassian-confluence
aashari/mcp-server-atlassian-confluence
Node.js/TypeScript MCP server for Atlassian Confluence. Provides tools enabling AI systems (LLMs) to list/get spaces & pages (content formatted as Markdown) and search via CQL. Connects AI seamlessly to Confluence knowledge bases using the standard MCP interface.
Supports token-optimized TOON output format (reducing API response sizes by 30-60% compared to JSON) and JMESPath filtering for selective field extraction. Implements five generic HTTP tools (`conf_get`, `conf_post`, `conf_put`, `conf_patch`, `conf_delete`) providing direct access to the Confluence v2 API, enabling AI assistants to create/update pages and comments alongside read operations. Integrates with Claude Desktop, Cursor AI, and any MCP-compatible assistant via stdio transport, with flexible credential management through environment variables, `.env` files, or `~/.mcp/configs.json`.
About mcp-server-atlassian-jira
aashari/mcp-server-atlassian-jira
Node.js/TypeScript MCP server for Atlassian Jira. Equips AI systems (LLMs) with tools to list/get projects, search/get issues (using JQL/ID), and view dev info (commits, PRs). Connects AI capabilities directly into Jira project management and issue tracking workflows.
Implements MCP (Model Context Protocol) with stdio transport, allowing seamless integration with Claude Desktop, Cursor AI, and other compatible LLM clients through a single standardized interface. Exposes five generic HTTP tools (GET, POST, PUT, PATCH, DELETE) covering the full Jira REST API v3, enabling AI systems to perform CRUD operations on projects, issues, comments, and worklogs. Features token-efficient TOON output format (30-60% reduction vs JSON), JMESPath filtering for response refinement, and automatic truncation handling for large responses exceeding 40k characters.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work