Skip to content

Feedback on your agent-browser skill #240

@RichardHightower

Description

@RichardHightower

I was curious about your browser automation approach, but the missing README made it hard to understand the actual use cases—are you targeting web testing, data extraction, or something else entirely?

Links:

TL;DR

You're at 66/100, solidly in D territory—needs work before this is production-ready. The evaluation uses Anthropic's skill best practices, and your strongest area is Spec Compliance (12/15), which means your YAML frontmatter is solid. The weak spot is Utility (11/20) by a fair margin—your skill does the plumbing but doesn't show enough of why someone would use it or what they can actually do with it.

What's Working Well

  • Clean frontmatter: Your YAML metadata is valid and hits all required fields—no errors there
  • Proper naming: agent-browser follows hyphen-case conventions correctly
  • Structured references: You've got 5 reference files (authentication, proxy-support, video-recording, snapshot-refs, session-management) which shows you're thinking about layered documentation
  • Good examples: You've got code blocks sprinkled through, which nets a +2 bonus for exemplary examples

The Big One: Utility is Your Blocker (11/20 → Target 18+)

This is the killer. Your skill describes what it does but barely shows why I'd use it or what I can build with it. You've got 1/4 points for feedback loops and only 5/8 for problem-solving power.

The fix: Add concrete use-case sections to your main SKILL.md. Right now it's too API-reference-heavy. You need something like:

## Common Use Cases

- **Automated testing**: Verify UI behavior across browsers
- **Data extraction**: Scrape dynamic content that requires JavaScript
- **Screenshot monitoring**: Compare visual changes over time
- **Form testing**: Validate complex multi-step workflows

Then add 2-3 minimal working examples showing real problems solved. This alone could bump you from 11 to 15+ on Utility.

Other Things Worth Fixing

  1. Trigger phrases (Description Quality: 3/4) – You've got "1-2 trigger phrases" but LLM-based search needs 4-5. Add: "web automation", "browser testing", "screenshot", "visual regression", "headless browser"

  2. Navigation signals (PDA: 3/5) – Your SKILL.md is 1496 words but lacks a clear table of contents. Add headers with anchor links so users can jump to sections like "Getting Started", "Advanced Configuration", "Examples"

  3. Metadata quality (Ease of Use: 6/10) – You're missing optional fields. Add tags, categories, and requirements to your frontmatter for better discoverability in the marketplace

  4. Feedback loops (Utility: 1/4) – Zero validation keywords detected. Show what success looks like: "Expected output", "Common errors & fixes", "How to verify it works"

Quick Wins

  • Add 4-5 trigger phrases to your description (+1-2 points, 5 min)
  • Create a TOC with anchor links in SKILL.md (+2 points, 10 min)
  • Write 2-3 use-case examples with expected output (+4-5 points, 30 min)
  • Add optional metadata fields (tags, categories) (+1-2 points, 5 min)

These four changes could realistically get you to 76-78/100, solidly in C territory. Focus on Utility first—it's your biggest gap and the most impactful lever to pull.


Checkout your skill here: [SkillzWave.ai](https://skillzwave.ai) | [SpillWave](https://spillwave.com) We have an agentic skill installer that install skills in 14+ coding agent platforms. Check out this guide on how to improve your agentic skills.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions