MCP Structural Analysis - December 23, 2025 #7367
Closed
Replies: 1 comment
-
|
This discussion was automatically closed because it was created by an agentic workflow more than 3 days ago. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Today's analysis evaluated 10 GitHub MCP tools across 9 toolsets, measuring both response sizes and structural usefulness for agentic workflows. Key finding: Most tools (60%) achieve excellent ratings (5/5), but
list_code_scanning_alertsremains problematic at 29,700 tokens per call due to embedded documentation.Full Structural Analysis Report
Executive Summary
get_me: 1/5 (403 error)get_label: 30 tokenslist_code_scanning_alerts: 29,700 tokensUsefulness Ratings for Agentic Work
⭐⭐⭐⭐⭐ Excellent (Rating: 5/5)
get_labellist_brancheslist_discussionslist_workflowssearch_repositoriesget_file_contentsAnalysis: These 6 tools represent the gold standard for agentic workflows - minimal overhead, clear structure, immediately actionable data.
⭐⭐⭐⭐ Good (Rating: 4/5)
list_issueslist_pull_requestsAnalysis: These tools provide comprehensive data but at a high token cost. The full body text in issues/PRs can be massive. Agents should use these selectively or consider fetching minimal data first, then requesting full details only when needed.
⭐⭐ Limited (Rating: 2/5)
list_code_scanning_alertsAnalysis: This tool is problematic for agentic workflows. A single call with just a few alerts can consume 30K+ tokens due to embedded documentation that repeats for each alert. Agents should check alert counts first and fetch individual alerts when needed.
⭐ Poor (Rating: 1/5)
get_meAnalysis: This tool is unavailable in workflow contexts using GitHub App tokens. Not useful for this environment.
Schema Analysis
get_labellist_brancheslist_workflowslist_discussionssearch_repositoriesget_file_contentslist_issueslist_pull_requestslist_code_scanning_alertsKey Findings:
Response Size Analysis by Toolset
Tool-by-Tool Detailed Analysis
Excellent Tools (Use Freely)
get_label(30 tokens, 5/5)list_branches(55 tokens, 5/5)list_discussions(130 tokens, 5/5)list_workflows(170 tokens, 5/5)search_repositories(420 tokens, 5/5)get_file_contents(1,100 tokens, 5/5)Good Tools (Use Selectively)
list_issues(4,800 tokens, 4/5)list_pull_requests(9,000 tokens, 4/5)get_pull_requestfor specific PRs instead of listing.Limited Tools (Use Cautiously)
list_code_scanning_alerts(29,700 tokens, 2/5)get_code_scanning_alertto fetch individual alerts instead. Desperately needs a minimal_output option to exclude documentation.Unavailable Tools
get_me(25 tokens error, 1/5)30-Day Trend Summary
get_label(28-35 tokens)list_code_scanning_alerts(5-29,700 tokens)Trend: Response sizes are generally stable, indicating consistent API behavior. The main variability comes from content-dependent tools (issues/PRs with long bodies, security alerts with many findings).
Recommendations
For Optimal Agentic Workflows
1. High-value, low-cost tools (Rating 5, <500 tokens):
get_label- Label operationslist_branches- Branch discoverylist_discussions- Discussion discoverylist_workflows- Workflow discoverysearch_repositories- Repository search2. Context-efficient tools (High rating, reasonable cost):
get_file_contents- File reading (1,100 tokens)3. Use selectively (High cost but valuable):
list_issues- Limit perPage, fetch titles firstlist_pull_requests- Limit perPage, use for specific PRs only4. Avoid or use extreme caution (Very high cost):
list_code_scanning_alerts- Check count first, fetch individuallyTool Selection Strategy for Agents
Discovery Phase (Minimize Context):
list_workflows,list_branches,list_discussions,get_labelsearch_repositorieswith minimal_outputAnalysis Phase (Targeted Fetches):
get_file_contentsfor specific fileslist_issueswith perPage=5-10 for targeted analysisget_pull_request) over list operationsContext Management:
Improvements Needed
Critical:
list_code_scanning_alertsneeds minimal_output optionNice to have: Body truncation option for issues/PRs
Visualizations
Average Response Size by Toolset
This chart shows the dramatic difference between toolsets. Code security tooling averages 29,700 tokens per call due to embedded documentation, while most other toolsets stay under 1,000 tokens.
Usefulness Ratings by Toolset
Six toolsets achieve perfect 5/5 ratings for agentic workflows. Issues and PRs rate 4/5 due to verbosity. Code security rates 2/5 due to extreme overhead.
Daily Token Usage Trend (30 Days)
Token usage remains stable day-to-day, with spikes when code scanning alerts are tested. Average daily testing consumes ~18K tokens.
Token Size vs Usefulness Rating
The ideal tools cluster in the bottom-left: low tokens, high usefulness. Note that usefulness doesn't necessarily correlate with size -
get_file_contentsat 1,100 tokens is as useful asget_labelat 30 tokens.Top 10 Tools by Response Size
list_code_scanning_alertsdominates at 29,700 tokens - 3x larger than the next tool. Most tools stay under 2,000 tokens.References:
Beta Was this translation helpful? Give feedback.
All reactions