hypothesis-library

31 stars 7 forks
28
F

Curated repository of experiment hypotheses, assumptions, and historical learnings.

Marketplace

Third-Party Agent Skill: Review the code before installing. Agent skills execute in your AI assistant's environment and can access your files. Learn more about security

Installation for Agentic Skill

View all platforms →
skilz install gtmagents/gtm-agents/hypothesis-library
skilz install gtmagents/gtm-agents/hypothesis-library --agent opencode
skilz install gtmagents/gtm-agents/hypothesis-library --agent codex
skilz install gtmagents/gtm-agents/hypothesis-library --agent gemini

First time? Install Skilz: pip install skilz

Works with 22+ AI coding assistants

Cursor, Aider, Copilot, Windsurf, Qwen, Kimi, and more...

View All Agents
Download Agent Skill ZIP

Extract and copy to ~/.claude/skills/ then restart Claude Desktop

1. Clone the repository:
git clone https://github.com/gtmagents/gtm-agents
2. Copy the agent skill directory:
cp -r gtm-agents/plugins/growth-experiments/skills/hypothesis-library ~/.claude/skills/

Need detailed installation help? Check our platform-specific guides:

Related Agentic Skills

Agentic Skill Details

Repository
gtm-agents
Stars
31
Forks
7
Type
Non-Technical
Meta-Domain
general
Primary Domain
general
Market Score
28

Agent Skill Grade

F
Score: 50/100 Click to see breakdown

Score Breakdown

Spec Compliance
11/15
PDA Architecture
12/30
Ease of Use
11/25
Writing Style
7/10
Utility
8/20
Modifiers: +1

Areas to Improve

  • Description needs trigger phrases
  • Missing Reference Files for Templates
  • Vague Trigger Terms

Recommendations

  • Focus on improving Pda (currently 12/30)
  • Focus on improving Ease Of Use (currently 11/25)
  • Focus on improving Utility (currently 8/20)

Graded: 2026-01-24

Developer Feedback

I've been diving into property-based testing frameworks lately, and your hypothesis-library skill caught my attention—though the 50/100 score suggests there might be some gaps between the concept and execution that are worth digging into.

Links:

TL;DR

You're at 50/100, solidly in F grade territory. This is based on Anthropic's skill evaluation best practices across five pillars. Your Spec Compliance is actually solid (11/15)—the frontmatter is valid and naming conventions are right—but Progressive Disclosure Architecture (12/30) and Utility (8/20) are where you're losing the most points. The core issue: you've got a great framework concept, but it's missing the teeth to be actually useful.

What's Working Well

  • Consistent terminology: You use 'hypothesis', 'learnings', and 'experiment' consistently throughout—no confusing terminology shifts that would make users stumble.
  • Solid metadata schema thinking: The idea of using ID, theme, persona, funnel stage, and metrics is the right foundation for structured experimentation tracking.
  • Logical section flow: "When to Use" → "Framework" → "Templates" → "Tips" follows a reasonable progression that's easy to scan.

The Big One: Missing Reference Files Kills Progressive Disclosure

Your skill lists three templates—intake form, learning card, portfolio dashboard—but provides zero actual content. This is a critical gap. You're telling users "here are templates" without showing them what they look like, which means they either guess or bounce.

The fix: Create three reference files:

  • references/intake-form.md – actual template with example fields
  • references/learning-card.md – structured format showing context/...

Report Security Issue

Found a security vulnerability in this agent skill?