feat: add clarvia-aeo-check — score MCP servers for agent-readiness (#402)
Co-authored-by: digitamaz <noreply@digitamaz.com>
This commit is contained in:
133
skills/clarvia-aeo-check/SKILL.md
Normal file
133
skills/clarvia-aeo-check/SKILL.md
Normal file
@@ -0,0 +1,133 @@
|
||||
---
|
||||
name: clarvia-aeo-check
|
||||
description: "Score any MCP server, API, or CLI for agent-readiness using Clarvia AEO (Agent Experience Optimization). Search 15,400+ indexed tools before adding them to your workflow."
|
||||
category: tool-quality
|
||||
risk: safe
|
||||
source: community
|
||||
date_added: "2026-03-27"
|
||||
author: digitamaz
|
||||
tags: [mcp, aeo, tool-quality, agent-readiness, api-scoring, clarvia]
|
||||
tools: [claude, cursor, windsurf, cline]
|
||||
---
|
||||
|
||||
# Clarvia AEO Check
|
||||
|
||||
## Overview
|
||||
|
||||
Before adding any MCP server, API, or CLI tool to your agent workflow, use Clarvia to score its agent-readiness. Clarvia evaluates 15,400+ AI tools across four AEO dimensions: API accessibility, data structuring, agent compatibility, and trust signals.
|
||||
|
||||
## Prerequisites
|
||||
|
||||
Add Clarvia MCP server to your config:
|
||||
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"clarvia": {
|
||||
"command": "npx",
|
||||
"args": ["-y", "clarvia-mcp-server"]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## When to Use This Skill
|
||||
|
||||
- Use when evaluating a new MCP server before adding it to your config
|
||||
- Use when comparing two tools for the same job
|
||||
- Use when building an agent that selects tools dynamically
|
||||
- Use when you want to find the highest-quality tool in a category
|
||||
|
||||
## How It Works
|
||||
|
||||
### Step 1: Score a specific tool
|
||||
|
||||
Ask Claude to score any tool by URL or name:
|
||||
|
||||
```
|
||||
Score https://github.com/example/my-mcp-server for agent-readiness
|
||||
```
|
||||
|
||||
Clarvia returns a 0-100 AEO score with breakdown across four dimensions.
|
||||
|
||||
### Step 2: Search tools by category
|
||||
|
||||
```
|
||||
Find the top-rated database MCP servers using Clarvia
|
||||
```
|
||||
|
||||
Returns ranked results from 15,400+ indexed tools.
|
||||
|
||||
### Step 3: Compare tools head-to-head
|
||||
|
||||
```
|
||||
Compare supabase-mcp vs firebase-mcp using Clarvia
|
||||
```
|
||||
|
||||
Returns side-by-side score breakdown with a recommendation.
|
||||
|
||||
### Step 4: Check leaderboard
|
||||
|
||||
```
|
||||
Show me the top 10 MCP servers for authentication using Clarvia
|
||||
```
|
||||
|
||||
## Examples
|
||||
|
||||
### Example 1: Evaluate before installing
|
||||
|
||||
```
|
||||
Before I add this MCP server to my config, score it:
|
||||
https://github.com/example/new-tool
|
||||
|
||||
Use the clarvia aeo_score tool and tell me if it's agent-ready.
|
||||
```
|
||||
|
||||
### Example 2: Find best tool in category
|
||||
|
||||
```
|
||||
I need an MCP server for web scraping. Use Clarvia to find the
|
||||
top-rated options and compare the top 3.
|
||||
```
|
||||
|
||||
### Example 3: CI/CD quality gate
|
||||
|
||||
Add to your CI pipeline using the GitHub Action:
|
||||
|
||||
```yaml
|
||||
- uses: clarvia-project/clarvia-action@v1
|
||||
with:
|
||||
url: https://your-api.com
|
||||
fail-under: 70
|
||||
```
|
||||
|
||||
## AEO Score Interpretation
|
||||
|
||||
| Score | Rating | Meaning |
|
||||
|-------|--------|---------|
|
||||
| 90-100 | Agent Native | Built specifically for agent use |
|
||||
| 70-89 | Agent Friendly | Works well, minor gaps |
|
||||
| 50-69 | Agent Compatible | Works but needs improvement |
|
||||
| 30-49 | Agent Partial | Significant limitations |
|
||||
| 0-29 | Not Agent Ready | Avoid for agentic workflows |
|
||||
|
||||
## Best Practices
|
||||
|
||||
- ✅ Score tools before adding them to long-running agent workflows
|
||||
- ✅ Use Clarvia's leaderboard to discover alternatives you haven't considered
|
||||
- ✅ Re-check scores periodically — tools improve over time
|
||||
- ❌ Don't skip scoring for "well-known" tools — even popular tools can score poorly
|
||||
- ❌ Don't use tools scoring below 50 in production agent pipelines without understanding the limitations
|
||||
|
||||
## Common Pitfalls
|
||||
|
||||
- **Problem:** Clarvia returns "not found" for a tool
|
||||
**Solution:** Try scanning by URL directly with `aeo_score` — Clarvia will score it on-demand
|
||||
|
||||
- **Problem:** Score seems low for a tool I trust
|
||||
**Solution:** Use `get_score_breakdown` to see which dimensions are weak and decide if they matter for your use case
|
||||
|
||||
## Related Skills
|
||||
|
||||
- `@mcp-builder` - Build a new MCP server that scores well on AEO
|
||||
- `@agent-evaluation` - Broader agent quality evaluation framework
|
||||
Reference in New Issue
Block a user