skill-upgrader
π―Skillfrom parcadei/continuous-claude-v3
skill-upgrader skill from parcadei/continuous-claude-v3
Installation
npx skills add https://github.com/parcadei/continuous-claude-v3 --skill skill-upgraderSkill Details
Upgrade any skill to v5 Hybrid format using decision theory + modal logic
Overview
# Skill Upgrader
Meta-skill that upgrades any SKILL.md to Decision Theory v5 Hybrid format using 4 parallel Ragie-backed agents.
When to Use
- "Upgrade this skill to v5"
- "Formalize this skill with decision theory"
- "Add MDP structure to this skill"
- "Apply the skill-upgrader to X"
Prerequisites
Ragie RAG with indexed books:
- decision-theory partition: LaValle Planning Algorithms, Sutton & Barto RL
- modal-logic partition: Blackburn Modal Logic, Huth & Ryan Logic in CS
Workflow
Step 1: Setup Session
```bash
SESSION=$(date +%Y%m%d-%H%M%S)-upgrade-{skill_name}
mkdir -p thoughts/skill-builds/${SESSION}
```
Step 2: Initialize Blackboard
Create thoughts/skill-builds/{session}/00-blackboard.md:
```markdown
# Skill Upgrade: {skill_name}
Started: {timestamp}
Input Skill
{path_to_skill}
Target Format
Decision Theory v5 Hybrid
Agent Findings
(Agents append below)
---
```
Step 3: Launch 4 Agents in Parallel
Use Task tool to spawn all 4 agents simultaneously. Each agent:
- Reads the input skill
- Queries Ragie for their specific book
- Appends findings to the blackboard
---
Agent 1: LaValle Planner
Book: LaValle's "Planning Algorithms" (decision-theory partition)
Focus: States, Actions, Transitions
```
Task(
subagent_type="general-purpose",
prompt="""
INPUT SKILL: {path}
BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: LaValle's "Planning Algorithms" in Ragie partition 'decision-theory'
TASK: Identify MDP structure in the skill.
Query Ragie:
```bash
uv run python scripts/ragie_query.py -q "MDP state space definition" -p decision-theory
uv run python scripts/ragie_query.py -q "action space sequential decisions" -p decision-theory
uv run python scripts/ragie_query.py -q "POMDP partial observability" -p decision-theory
```
Read the input skill and answer:
- What are the STATES? (phases, modes, tracked info)
- What are the ACTIONS? (what can agent do in each state)
- How do TRANSITIONS work? (deterministic or stochastic)
- Is this POMDP or fully observable?
WRITE to blackboard section: ## Agent 1: States, Actions & Transitions
Format as plain English with LaValle chapter citations.
"""
)
```
---
Agent 2: Sutton & Barto Optimizer
Book: Sutton & Barto's "Reinforcement Learning" (decision-theory partition)
Focus: Policy, Termination, Value
Depends on: Agent 1
```
Task(
subagent_type="general-purpose",
prompt="""
INPUT SKILL: {path}
BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: Sutton & Barto's "Reinforcement Learning" in Ragie partition 'decision-theory'
WAIT: Read Agent 1's findings from blackboard first.
TASK: Design policy and termination conditions.
Query Ragie:
```bash
uv run python scripts/ragie_query.py -q "policy deterministic stochastic" -p decision-theory
uv run python scripts/ragie_query.py -q "episodic termination conditions" -p decision-theory
uv run python scripts/ragie_query.py -q "reward function design" -p decision-theory
```
Using Agent 1's states and actions, answer:
- What's the POLICY? (state β action rules)
- When does it END? (terminal states, success/failure)
- What are REWARDS? (goals +, costs -)
- Which states are HIGH/LOW value?
WRITE to blackboard section: ## Agent 2: Policy & Values
Format as plain English with Sutton & Barto section citations.
"""
)
```
---
Agent 3: Blackburn Modal Logician
Book: Blackburn's "Modal Logic" (modal-logic partition)
Focus: Constraints (temporal, epistemic, deontic)
```
Task(
subagent_type="general-purpose",
prompt="""
INPUT SKILL: {path}
BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: Blackburn's "Modal Logic" in Ragie partition 'modal-logic'
TASK: Extract constraints from the skill.
Query Ragie:
```bash
uv run python scripts/ragie_query.py -q "temporal logic LTL operators" -p modal-logic
uv run python scripts/ragie_query.py -q "epistemic logic knowledge" -p modal-logic
uv run python scripts/ragie_query.py -q "deontic logic obligations" -p modal-logic
```
Read the input skill and identify:
- TEMPORAL: "must do X before Y" β β‘, β, U
- EPISTEMIC: "must know X" β K operator
- DEONTIC: "must/forbidden/may" β O, F, P
- DYNAMIC: "action causes effect" β [action]
WRITE to blackboard section: ## Agent 3: Constraints
For each constraint:
- Plain English description
- Modal logic notation
- Why it matters
- Blackburn chapter citation
"""
)
```
---
Agent 4: Huth & Ryan Verifier
Book: Huth & Ryan's "Logic in Computer Science" (modal-logic partition)
Focus: Validation, Safety, Liveness
Depends on: Agents 1-3
```
Task(
subagent_type="general-purpose",
prompt="""
INPUT SKILL: {path}
BLACKBOARD: thoughts/skill-builds/{session}/00-blackboard.md
YOUR BOOK: Huth & Ryan's "Logic in Computer Science" in Ragie partition 'modal-logic'
WAIT: Read Agents 1-3 findings from blackboard first.
TASK: Verify consistency and completeness.
Query Ragie:
```bash
uv run python scripts/ragie_query.py -q "safety properties verification" -p modal-logic
uv run python scripts/ragie_query.py -q "liveness properties eventually" -p modal-logic
uv run python scripts/ragie_query.py -q "model checking CTL" -p modal-logic
```
Check:
- SAFETY: What bad things never happen? β‘Β¬(bad)
- LIVENESS: What good things eventually happen? β(good)
- CONSISTENCY: Any contradictions between agents?
- COMPLETENESS: Any gaps in coverage?
WRITE to blackboard section: ## Agent 4: Verification
Report with β/β for each property.
Overall verdict: PASS or NEEDS_WORK
Huth & Ryan section citations.
"""
)
```
---
Step 4: Synthesize Final Skill
After all agents complete, read the blackboard and create:
Output: thoughts/skill-builds/{session}/SKILL-upgraded.md
Use v5 Hybrid template:
```yaml
---
name: {original_name}
description: {original_description}
version: 5.1-hybrid
---
# Option: {name}
Initiation (I)
[From original + Agent 1 state analysis]
Observation Space (Y)
[From Agent 1 POMDP analysis]
Action Space (U)
[From Agent 1 actions]
Policy (pi)
[From Agent 2 stateβaction rules]
Termination (beta)
[From Agent 2 episode structure]
Q-Heuristics
[From Agent 2 value guidance]
Constraints
[From Agent 3 modal logic]
Verification
[From Agent 4 safety/liveness]
```
---
Example Usage
```
User: "Upgrade .claude/skills/implement_plan/SKILL.md to v5 Hybrid"
Claude:
- Creates session directory
- Initializes blackboard
- Launches 4 agents in parallel (Task tool)
- Waits for completion
- Reads blackboard
- Synthesizes upgraded skill
- Reports: "Upgraded skill at thoughts/skill-builds/.../SKILL-upgraded.md"
```
Ragie Query Reference
```bash
# Decision theory partition
uv run python scripts/ragie_query.py -q "your question" -p decision-theory
# Modal logic partition
uv run python scripts/ragie_query.py -q "your question" -p modal-logic
# With reranking for better results
uv run python scripts/ragie_query.py -q "your question" -p decision-theory --rerank
```
Files Created
After upgrade:
```
thoughts/skill-builds/{session}/
βββ 00-blackboard.md # Agent collaboration
βββ SKILL-upgraded.md # Final v5 Hybrid skill
βββ validation-report.md # Agent 4 verification
```