neurolink and mcp
About neurolink
juspay/neurolink
Universal AI Development Platform with MCP server integration, multi-provider support, and professional CLI. Build, test, and deploy AI applications with multiple ai providers.
Abstracts multi-provider LLM communication as composable token streams using a pipe-based architecture, unifying 13 AI providers (OpenAI, Anthropic, Google, AWS Bedrock, Azure, etc.) under a single TypeScript API. Built-in features include 64+ MCP server tools, Redis-backed persistent memory with LLM-powered condensation, context window auto-compaction with per-provider token estimation, RAG with hybrid search and reranking, and multi-provider failover for cost optimization. Deployable via professional CLI or as HTTP servers (Hono, Express, Fastify, Koa) with full observability hooks for existing OpenTelemetry instrumentation.
About mcp
mondaycom/mcp
Enable AI agents to work reliably - giving them secure access to structured data, tools to take action, and the context needed to make smart decisions.
Implements the Model Context Protocol (MCP) as a standardized interface for AI agents to query monday.com's work OS via GraphQL API, with both hosted and self-hosted deployment options. Provides TypeScript-based server and agent toolkit packages compatible with Claude, ChatGPT, Cursor, and other AI platforms, handling OAuth authentication and workspace-level access controls. Supports real-time board management, task tracking, and workflow automation without requiring agents to understand monday.com's API directly.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work