How Agents Acquire Skills: The Emerging Architecture of Composable Capabilities

Skills, registries, marketplaces, and the trust problem — scored on quality daily. A map of the 2,700+ repos building the capability layer for autonomous agents.

Graham Rowe · April 01, 2026 · Updated daily with live data
agents llm-tools

"Agent skills" is the new vocabulary. Six months ago, the conversation was about tool use and function calling. Now developers are searching for evoskills, antigravity-skills, and skill registries — a higher-level abstraction where agents don't just call tools but acquire composable capabilities they can combine and reuse.

PT-Edge tracks over 2,700 repositories across skill collections, registries, marketplaces, runtime engines, and communication standards. This guide maps the emerging architecture: where skills run, where they're found, how they're composed, and where the critical trust gaps remain.

Skills vs tools vs MCP: what's actually different?

The terminology is still settling, but the distinctions matter:

  • Function calling is the primitive: an LLM invokes a typed function with parameters. One action, one response.
  • MCP (Model Context Protocol) is the transport: a standard way for agents to discover and call tools across servers. It handles connection, authentication, and protocol — but the tools themselves are individual endpoints.
  • Agent skills are higher-level: a skill is a bundle of related capabilities, often with state, context, and multi-step logic. A "research skill" might combine web search, document parsing, summarisation, and citation — presented to the agent as a single composable capability rather than four separate tool calls.

The practical difference: function calling and MCP give agents actions. Skills give agents competencies. The skill layer is where the intelligence about how to combine actions lives.

Where skills run: the platform layer

Skills need runtime environments that handle execution, sandboxing, authentication, and tool orchestration. A few platforms dominate this layer.

ProjectScoreStarsWhat it does
composio 94/100 27,355 Composio powers 1000+ toolkits, tool search, context management, authentication, and a...
agno 96/100 38,655 Build, run, manage agentic software at scale.
agentscope 90/100 18,063 Build and run agents you can see, understand and trust.
agent-squad 73/100 7,506 Flexible and powerful framework for managing multiple AI agents and handling complex...
astron-agent 70/100 10,111 Enterprise-grade, commercial-friendly agentic workflow platform for building...

Composio (94/100, 27,355 stars, 528 commits in the last 30 days) is the quality leader and the closest thing to a universal skill platform. It powers 1,000+ toolkits with built-in tool search, context management, authentication, and a sandboxed workbench. If you're building agents that need to compose many capabilities, Composio is the infrastructure most teams start with.

agno (38,655 stars) has the largest user base — an agent runtime for building, running, and managing agentic software at scale. AgentScope (90/100) focuses on transparency and trust — agents you can see, understand, and verify.

For enterprise deployment, agent-squad from AWS (7,506 stars) provides multi-agent orchestration, and astron-agent (10,111 stars, 238 commits in the last 30 days) offers an enterprise-grade, commercial-friendly workflow platform.

Where skills are found: collections and catalogues

The most immediate way to give your agent new capabilities is to install skills from curated collections. Two dominate the space:

ProjectScoreStarsSkills
antigravity-awesome-skills 70/100 23,847 1,000+
awesome-agent-skills 68/100 11,079 500+
openclaw-master-skills 64/100 1,534 339+

antigravity-awesome-skills (23,847 stars, 533 commits in the last 30 days) is the largest collection: 1,000+ skills for Claude Code, Antigravity, and Cursor, including official skills from Anthropic and Vercel. awesome-agent-skills (11,079 stars) offers 500+ skills compatible across agents.

These collections grew fast because skills are easy to write but hard to discover. A "skill" can be as simple as a Markdown file with instructions, or as complex as a multi-step workflow with API integrations. The catalogues solve the discovery problem, but they don't solve the quality problem — not every skill in a 1,000-item collection is production-ready.

The registry problem: discovery and trust are unsolved

With 292 repos in the agent-skill-registry category (average quality 22.9/100), the ecosystem is clearly looking for a universal registry. Nobody has built one that's won.

Acontext (70/100, 3,154 stars, 174 commits in the last 30 days) takes the most interesting approach: treating agent skills as a memory layer. Instead of a static registry, skills are contextually retrieved based on what the agent is trying to do. registry-broker-skills (113 stars) takes the scale approach: 72,000+ agents across 14 protocols.

For practical skill installation, skillport (338 stars) and flins offer universal installers that work across coding agents. The vision: install a skill once, use it from Claude Code, Cursor, Codex, or any MCP-compatible agent.

The fragmentation is real. Hacker News asked "Any real OpenClaw users? What's your experience?" (121 points, 189 comments) — the community is interested but uncertain about which ecosystem to invest in.

The OpenClaw ecosystem: most visible, most debated

OpenClaw is the most discussed skills implementation on Hacker News. The "Agent Skills" concept post got 544 points. But the community reaction is split: excitement about the concept, concern about the execution.

The concerns are legitimate. "OpenClaw is basically a cascade of LLMs in prime position to mess stuff up" (110 HN points) captures the core worry: skills execute with agent-level permissions. A malicious or buggy skill has the same access as the agent itself. The agent-skill-security category (64 repos) exists specifically to address this, but average quality is just 23.8/100.

The OpenClaw marketplace (clawdtm, openclaw-master-skills) is building curation and rating systems. But the 230 repos in openclaw-skill-integrations average just 14.8/100 — the lowest quality average of any subcategory in this landscape. Most individual skills are trivial wrappers.

Standards: early but necessary

For skills to be truly composable and portable, the ecosystem needs shared standards. A few projects are attempting this:

better-agents from LangWatch (1,507 stars) is building standards for agent construction. agent-capability-standard defines 36 atomic capabilities across 9 cognitive layers — the most formally specified capability model in the space. AgentsMesh (1,123 stars, 208 commits in the last 30 days) is building the orchestration layer that connects different agent runtimes.

The benchmarking conversation is maturing too: SkillsBench (364 HN points) is attempting to measure how well skills actually work across diverse tasks, and the Agent Skills Leaderboard (135 HN points) is ranking them. These are the infrastructure projects that turn a fragmented ecosystem into something you can evaluate.

Three patterns shaping agent skills

1. Skills collections are growing faster than quality controls

The biggest collections already have 1,000+ entries. But the quality distribution is bimodal: a small number of well-maintained skills sit alongside hundreds of untested wrappers. The ecosystem needs better quality signals — which is exactly what scoring and benchmarking projects are trying to provide.

2. The registry layer is fragmenting before consolidating

292 registry repos is too many. The market is searching for a standard discovery mechanism. MCP provides part of the answer (tool discovery via protocol), but skill-level discovery (finding the right competency, not just the right function) is still unsolved. Whoever cracks this becomes the npm of agent skills.

3. Trust is the binding constraint

Skills execute code with agent-level permissions. The trust model — how do you know a skill is safe? — is the most important unsolved problem. Curated collections help. Sandboxing helps. But a systematic, automated approach to skill security vetting doesn't exist yet. The skill-security subcategory (64 repos, avg quality 23.8) is where the answers will come from.

How to use this data

Every project mentioned in this guide has a quality-scored page in our agents directory, updated daily. You can:

Quality scores update daily from live GitHub, PyPI, and npm data. In a space moving this fast, yesterday's star count doesn't tell you if a project is still maintained. The daily scoring does.

Related analysis