Skills Overview Guide¶
What Are Skills?¶
Skills are specialized commands and guides that extend HtmlGraph with focused, reusable capabilities. Think of them as expert consultants you can summon for specific tasks:
- Skills provide progressive disclosure - Start simple, dive deep only when needed
- Skills integrate with HtmlGraph - Seamlessly coordinate with tracking, delegation, and analytics
- Skills are discoverable - List, search, and learn at your own pace
When you need guidance on a specific area, invoke a skill to access expert-level documentation and decision frameworks—without overwhelming yourself with unnecessary detail.
Available Skills¶
| Skill | Purpose | When to Use |
|---|---|---|
/orchestrator-directives |
Comprehensive orchestration guidance | Planning complex workflows, understanding delegation decisions |
/multi-ai-orchestration |
Multi-model spawner selection & cost optimization | Choosing the right AI model for subagent tasks |
/code-quality |
Linting, type checking, testing workflows | Pre-commit validation, ensuring quality gates pass |
/deployment-automation |
Release management and version workflows | Publishing packages, managing releases |
/debugging-workflow |
Research-first debugging methodology | Systematic problem-solving, documentation-driven diagnosis |
/memory-sync |
Documentation synchronization patterns | Keeping docs consistent across platforms |
Decision Tree: Which Skill Should I Use?¶
Use this flowchart to find the right skill for your task:
"I'm planning a complex workflow"¶
→ Use /orchestrator-directives for:
- Deciding whether to delegate work
- Understanding parent-child session relationships
- Cost optimization strategies
- Complex multi-agent coordination patterns
Example: "I need to run 5 parallel test suites and then deploy. Should I delegate?"
"I need to choose a model for a subagent"¶
→ Use /multi-ai-orchestration for:
- Selecting spawner types (Gemini, Copilot, Codex, Claude)
- Cost vs. capability tradeoffs
- When to use cheap exploratory models vs. expensive reasoning models
- Spawner compatibility with your task
Example: "Should I use Gemini (free) or Claude (expensive) to explore the codebase?"
"I need to run tests/linters before committing"¶
→ Use /code-quality for:
- Running ruff, mypy, pytest in the correct order
- Fixing linting errors systematically
- Type checking and test failures
- Full quality gate workflow
Example: "My type checker is failing. What's the right fix order?"
"I need to release or publish code"¶
→ Use /deployment-automation for:
- Version number management
- PyPI publishing
- Plugin updates (Claude, Gemini, etc.)
- Release checklists
Example: "I need to bump the version and publish to PyPI."
"I'm stuck debugging something"¶
→ Use /debugging-workflow for:
- Research-first methodology (read docs before guessing)
- Systematic root cause analysis
- When to use debugger vs. researcher agents
- Using HtmlGraph spike documentation
Example: "My hooks aren't loading. I've tried 3 fixes already."
"I need to keep multiple docs synchronized"¶
→ Use /memory-sync for:
- Central documentation (single source of truth)
- Platform-specific file generation
- Automated consistency checking
- Documentation maintenance patterns
Example: "I updated AGENTS.md but forgot to update CLAUDE.md and GEMINI.md"
Skill Descriptions¶
/orchestrator-directives¶
Purpose: Complete orchestration guidance for complex workflows
Contains: - Full orchestrator decision framework - When to delegate vs. direct execution - Cost optimization patterns - Session hierarchy management - Multi-agent coordination examples - Advanced patterns (parallel execution, sequential handoff, divide-and-conquer)
Use this when: - Planning a feature that spans multiple agents - Deciding whether to delegate work - Optimizing for cost and speed simultaneously - Understanding how orchestrator tracking works - Creating complex delegation hierarchies
Related reading: Delegation Guide, AGENTS.md Orchestrator Mode section
/multi-ai-orchestration¶
Purpose: Spawner selection and multi-model cost optimization
Contains: - Spawner types and their capabilities - Cost vs. capability matrix - When to use Gemini (free) vs. Copilot vs. Claude - Dynamic spawner composition patterns - Cost calculation examples - Model selection algorithm
Use this when: - You need to choose which AI model to delegate work to - Optimizing budget for complex workflows - Mixing multiple spawner types in one workflow - Balancing speed vs. cost - Understanding spawner compatibility with tools
Related reading: README.md Orchestrator Architecture, AGENTS.md Multi-Agent section
/code-quality¶
Purpose: Linting, type checking, and testing validation workflows
Contains: - Complete quality gate sequence - Ruff fixing and formatting - Mypy type checking troubleshooting - Pytest execution patterns - Batch testing strategies - Pre-commit workflow
Use this when: - Before committing code - Fixing linting or type errors - Running test suites - Ensuring all quality checks pass - Understanding test failure patterns
Related reading: .claude/rules/code-hygiene.md, Deployment Guide
/deployment-automation¶
Purpose: Complete release and deployment workflows
Contains:
- Version number management across all files
- deploy-all.sh script options and flags
- PyPI publishing workflow
- Plugin update patterns (Claude, Gemini, Codex)
- Release checklist
- Rollback strategies
Use this when: - Bumping version numbers - Publishing a release to PyPI - Updating Claude plugin - Managing release tags - Creating GitHub releases
Related reading: .claude/rules/deployment.md, AGENTS.md Deployment & Release section
/debugging-workflow¶
Purpose: Research-first debugging methodology and systematic problem-solving
Contains: - Research-first vs. trial-and-error comparison - Built-in debug tools and agents - Systematic error analysis process - When to use researcher vs. debugger vs. test-runner agents - Integration with HtmlGraph spikes for documentation - Anti-patterns to avoid
Use this when: - You encounter an unfamiliar error - You've tried 2+ fixes without success - Working with Claude Code hooks or plugins - Need systematic root cause analysis - Want to document your debugging process
Related reading: .claude/rules/debugging.md, AGENTS.md Debugging & Quality section
/memory-sync¶
Purpose: Documentation synchronization patterns for platform-specific files
Contains: - Central documentation (AGENTS.md) concept - Platform-specific file generation - Sync checking and validation - Automated consistency workflows - Single source of truth patterns - Multi-platform maintenance
Use this when: - You've updated AGENTS.md and need to sync to other files - Checking if documentation is in sync - Setting up synchronization for a new project - Understanding documentation architecture - Maintaining consistency across Claude, Gemini, Codex docs
Related reading: AGENTS.md Documentation Synchronization, README.md Links section
How to Use Skills¶
Listing Available Skills¶
Invoking a Skill¶
# Simple invocation
/orchestrator-directives
# With context (optional)
/code-quality --for-python-projects
Progressive Disclosure¶
Skills follow a progressive disclosure model:
- Quick summary - Start with the basics
- Decision guide - Help you make choices
- Examples - Real-world usage patterns
- Details - Deep technical information
You don't need to read everything at once. Start with what you need, dive deeper as needed.
Skill Integration with HtmlGraph¶
Tracking Your Skill Usage¶
When you use a skill to solve a problem, document it in HtmlGraph:
from htmlgraph import SDK
sdk = SDK(agent="claude")
# Create a spike documenting your learnings
spike = sdk.spikes.create(
title="Learned debugging workflow - resolved hook loading issue"
).set_findings("""
Used /debugging-workflow skill to systematically diagnose hook loading problem.
Key learnings:
1. Research first - read Claude Code hook documentation
2. Check all hook sources (not just .claude/hooks/)
3. Understand that hooks from multiple sources MERGE
4. Verify with /hooks command
Issue was resolved by: Removing duplicate hooks from hook sources
""").save()
Orchestrator Directives Integration¶
When delegating complex tasks, reference the orchestrator skill:
from htmlgraph import Task
# Orchestrator coordinates based on skill guidance
Task(
subagent_type="general-purpose",
prompt="""Using /orchestrator-directives skill guidance:
Task: Run comprehensive test suite in parallel
Scope: tests/unit/, tests/integration/, tests/e2e/
Success criteria: Report pass/fail counts only
Time limit: 10 minutes total
"""
)
Code Quality Gates¶
Integrate code quality skill with deployment:
# Before deploying, ensure all quality gates pass
# (Follows /code-quality skill workflow)
uv run ruff check --fix && \
uv run ruff format && \
uv run mypy src/ && \
uv run pytest
# Only then deploy
./scripts/deploy-all.sh 0.9.4
Examples¶
Example 1: Planning a Feature with Orchestrator Directives¶
Scenario: You need to implement a complex authentication system
# Step 1: Use orchestrator directives to plan
/orchestrator-directives
# Understand:
# - Should I delegate work?
# - What sessions will be created?
# - How will costs be optimized?
# Step 2: Create feature tracking
# (Use HtmlGraph SDK as shown in delegation.md)
# Step 3: Execute with proper delegation pattern
# Task(subagent_type="...", prompt="...")
Example 2: Debugging a Mysterious Error¶
Scenario: Your tests are failing with an unclear error
# Step 1: Use debugging workflow
/debugging-workflow
# Learn:
# - Research first methodology
# - Systematic error analysis
# - When to use debugging agents
# Step 2: Document findings in spike
sdk.spikes.create("Debug: test failure in X").set_findings(...)
# Step 3: Don't guess - research and understand root cause
Example 3: Release Workflow¶
Scenario: You need to publish version 0.10.0
# Step 1: Ensure quality with /code-quality
/code-quality
# Step 2: Follow deployment automation
/deployment-automation
# Step 3: Execute steps in order
./scripts/deploy-all.sh 0.10.0
# Step 4: Verify publication
# Check PyPI, update plugins as documented
Quick Reference Table¶
| Need | Skill | Reference |
|---|---|---|
| Delegate work? | /orchestrator-directives |
Delegation Guide |
| Choose model | /multi-ai-orchestration |
AGENTS.md Orchestrator section |
| Fix tests/lints | /code-quality |
code-hygiene.md rules |
| Release version | /deployment-automation |
deployment.md rules |
| Diagnose error | /debugging-workflow |
debugging.md rules |
| Sync docs | /memory-sync |
AGENTS.md Sync section |
FAQs¶
Q: How do I know which skill to use?
A: Use the decision tree at the top of this guide. It walks through common scenarios and points you to the right skill.
Q: Can I use multiple skills for one task?
A: Yes! For example, planning a release might use /deployment-automation + /code-quality + /orchestrator-directives for delegating tests.
Q: Where are skills defined?
A: Skills are defined in packages/claude-plugin/skills/ in the HtmlGraph project.
Q: Do I need to memorize all skills?
A: No! Skills are discoverable. Use /help to list them, and this guide to understand what each does.
Q: Can I create custom skills?
A: Yes! See packages/claude-plugin/ for the skill development framework.
Related Reading¶
- Delegation Guide - Deep dive on Task() and orchestration
- AGENTS.md - Orchestrator Mode - Overview and quick start
- CLAUDE.md - Skills Reference - Project-specific skill guidance
.claude/rules/orchestration.md- Complete orchestrator directivespackages/claude-plugin/skills/- Skill definitions and implementations