🎯

analytics-flow

🎯Skill

from vladm3105/aidoc-flow-framework

VibeIndex|
What it does

analytics-flow skill from vladm3105/aidoc-flow-framework

analytics-flow

Installation

Install skill:
npx skills add https://github.com/vladm3105/aidoc-flow-framework --skill analytics-flow
5
AddedJan 27, 2026

Skill Details

SKILL.md

Analytics and data analysis workflow skill

Overview

# analytics-flow

Description: Project metrics, documentation analytics, progress tracking, and trend analysis

Category: Data Analysis & Reporting

Complexity: Medium (data aggregation + visualization)

---

Purpose

Provide visibility into project progress, documentation quality, requirement coverage, and development velocity. Generate actionable insights from SDD artifacts to identify bottlenecks and optimize workflows.

---

Capabilities

1. Documentation Coverage Metrics

  • Total documents by type (BRD, PRD, REQ, BDD, etc.)
  • Token count distribution
  • Documentation completeness percentage
  • Outdated document identification
  • Documentation debt tracking

2. Requirement Tracking

  • Total requirements by category
  • Requirement status (draft, approved, implemented, tested)
  • Requirement coverage (REQ β†’ BDD β†’ TASK mapping)
  • Orphaned requirements
  • Requirement age and staleness

3. Traceability Analysis

  • Complete traceability chains (BRD β†’ ... β†’ Code)
  • Broken traceability links
  • Traceability coverage percentage
  • Gap identification
  • Circular dependency detection

4. Implementation Progress

  • Tasks completed vs total
  • Implementation phase progress
  • Blocked vs active tasks
  • Timeline adherence
  • Milestone tracking

5. Velocity Tracking

  • Requirements completed per sprint/week
  • Implementation velocity
  • Documentation velocity
  • Test coverage velocity
  • Trend analysis over time

6. Quality Metrics

  • Code review scores over time
  • Test coverage trends
  • Security vulnerability trends
  • Code complexity trends
  • Documentation quality scores

7. Bottleneck Identification

  • Stuck requirements (no progress)
  • Blocked tasks
  • Resource constraints
  • Dependency chains blocking progress
  • Review bottlenecks

8. Custom Report Generation

  • Executive summaries
  • Sprint/weekly reports
  • Milestone reports
  • Team performance reports
  • Risk assessment reports

---

Analytics Workflow

```mermaid

graph TD

A[SDD Artifacts] --> B[Data Collection]

B --> C[Parse Documents]

C --> D[Extract Metadata]

D --> E[Requirement Analysis]

D --> F[Traceability Analysis]

D --> G[Progress Analysis]

D --> H[Quality Analysis]

E --> I[Requirement Metrics]

F --> J[Traceability Matrix]

G --> K[Progress Dashboard]

H --> L[Quality Dashboard]

I --> M[Aggregate Data]

J --> M

K --> M

L --> M

M --> N[Calculate KPIs]

N --> O[Identify Trends]

O --> P[Detect Anomalies]

P --> Q[Generate Insights]

Q --> R{Report Type}

R -->|Executive| S[Executive Summary]

R -->|Technical| T[Technical Report]

R -->|Sprint| U[Sprint Report]

S --> V[Distribute Reports]

T --> V

U --> V

```

---

Usage Instructions

Generate Project Dashboard

```bash

analytics-flow dashboard --output reports/dashboard.html

```

Generated dashboard sections:

```

Project Overview Dashboard

==========================

Overall Health: 78/100 (Good)

Documentation Coverage:

β”œβ”€ BRD Documents: 3/3 (100%) βœ“

β”œβ”€ PRD Documents: 2/3 (67%) ⚠️

β”œβ”€ Requirements: 146/150 (97%) βœ“

β”œβ”€ BDD Scenarios: 89/146 (61%) ⚠️

β”œβ”€ ADR Documents: 12/12 (100%) βœ“

β”œβ”€ SPEC Documents: 5/8 (63%) ⚠️

└─ TASK Documents: 45/89 (51%) ❌

Implementation Progress:

β”œβ”€ Phase 1: 100% (Completed) βœ“

β”œβ”€ Phase 2: 75% (In Progress) β†’

β”œβ”€ Phase 3: 0% (Not Started) β—‹

└─ Overall: 58% complete

Traceability Coverage:

β”œβ”€ Forward Traceability: 94% βœ“

β”œβ”€ Backward Traceability: 91% βœ“

└─ Complete Chains: 87% βœ“

Quality Metrics:

β”œβ”€ Code Coverage: 87% βœ“

β”œβ”€ Code Review Score: 82/100 βœ“

β”œβ”€ Security Issues: 3 (1 Critical) ⚠️

└─ Documentation Quality: 89% βœ“

Velocity (Last 4 Weeks):

β”œβ”€ Requirements: 12/week (avg)

β”œβ”€ Implementation: 8 tasks/week

β”œβ”€ Tests Written: 45 tests/week

└─ Trend: β†— Improving

Bottlenecks Identified:

β”œβ”€ REQ-AUTH-015: Stuck for 3 weeks

β”œβ”€ BDD Scenarios: 39% gap vs requirements

└─ Phase 2 Dependency: Blocked by external API

```

Generate Requirement Report

```bash

analytics-flow requirements --format json --output reports/requirements.json

```

Output:

```json

{

"summary": {

"total_requirements": 146,

"by_priority": {

"MUST": 130,

"SHOULD": 12,

"MAY": 4

},

"by_status": {

"draft": 5,

"approved": 85,

"implemented": 45,

"tested": 38,

"deployed": 30

},

"by_category": {

"authentication": 15,

"authorization": 12,

"data_processing": 45,

"api": 38,

"ml_models": 32,

"infrastructure": 4

}

},

"coverage": {

"with_bdd_scenarios": 89,

"with_tasks": 78,

"with_tests": 67,

"complete_traceability": 127

},

"age_analysis": {

"average_age_days": 45,

"oldest_requirement": {

"id": "REQ-AUTH-01",

"age_days": 120,

"status": "implemented"

},

"stale_requirements": [

{"id": "REQ-ML-015", "age_days": 90, "status": "draft"}

]

},

"orphaned_requirements": [

"REQ-AUTH-015",

"REQ-DATA-023"

]

}

```

Generate Traceability Matrix

```bash

analytics-flow traceability --format mermaid --output reports/traceability.mmd

```

Generated Mermaid diagram:

```mermaid

graph LR

BRD001[BRD-SYSTEM-001] --> PRD001[PRD-FEATURES-001]

PRD001 --> REQ001[REQ-AUTH-01]

PRD001 --> REQ002[REQ-AUTH-002]

REQ001 --> BDD001[BDD-LOGIN-001]

REQ001 --> BDD002[BDD-LOGIN-002]

REQ002 --> BDD003[BDD-LOGOUT-001]

BDD001 --> TASK001[TASK-AUTH-001]

BDD002 --> TASK002[TASK-AUTH-002]

BDD003 --> TASK003[TASK-AUTH-003]

TASK001 --> CODE001[auth/login.py]

TASK002 --> CODE001

TASK003 --> CODE002[auth/logout.py]

CODE001 --> TEST001[test_login.py]

CODE002 --> TEST002[test_logout.py]

style BRD001 fill:#e1f5ff

style PRD001 fill:#e8f5e9

style REQ001 fill:#fff3e0

style REQ002 fill:#fff3e0

style BDD001 fill:#f3e5f5

style BDD002 fill:#f3e5f5

style BDD003 fill:#f3e5f5

```

Generate Sprint Report

```bash

analytics-flow sprint \

--start-date 2025-01-01 \

--end-date 2025-01-14 \

--output reports/sprint-report.md

```

Generated report:

```markdown

# Sprint Report: Jan 1-14, 2025

Sprint Summary

  • Duration: 2 weeks
  • Team Size: 5 developers
  • Sprint Goal: Complete Phase 2 authentication features

Accomplishments

βœ“ Completed: 12 requirements

βœ“ Implemented: 18 tasks

βœ“ Tests Written: 67 tests

βœ“ Code Coverage: 87% β†’ 91% (+4%)

Requirements Completed

  1. REQ-AUTH-01: User login with JWT βœ“
  2. REQ-AUTH-002: User logout βœ“
  3. REQ-AUTH-003: Password reset βœ“
  4. REQ-AUTH-004: Email verification βœ“

... (8 more)

Velocity Metrics

  • Requirements: 6/week (target: 5/week) β†—
  • Tasks: 9/week (target: 8/week) β†—
  • Tests: 33.5/week (target: 25/week) β†—

Quality Metrics

  • Code Review Score: 85/100 (+3 from last sprint)
  • Security Issues: 2 (down from 5)
  • Test Coverage: 91% (up from 87%)

Blockers Resolved

  1. External API integration (blocked 5 days) β†’ Resolved
  2. Database migration issue β†’ Resolved

Current Blockers

  1. Waiting for design review on user profile page (2 days)

Risks

⚠️ Phase 3 dependency on external service (mitigation: mock implementation)

Next Sprint Goals

  • Complete remaining 8 authentication requirements
  • Begin authorization framework
  • Improve test coverage to 93%

```

Track Velocity Trends

```bash

analytics-flow velocity --weeks 12 --output reports/velocity.json

```

Output with trend analysis:

```json

{

"period": "12 weeks",

"metrics": [

{

"week": "2025-W01",

"requirements_completed": 8,

"tasks_completed": 15,

"tests_written": 42,

"code_coverage": 85

},

{

"week": "2025-W02",

"requirements_completed": 12,

"tasks_completed": 18,

"tests_written": 67,

"code_coverage": 91

}

],

"trends": {

"requirements": {

"average": 10,

"trend": "increasing",

"velocity_change": "+15%"

},

"tasks": {

"average": 16.5,

"trend": "stable",

"velocity_change": "+2%"

},

"tests": {

"average": 54.5,

"trend": "increasing",

"velocity_change": "+25%"

}

},

"forecast": {

"next_sprint_requirements": 13,

"completion_date_phase_2": "2025-02-15",

"confidence": "high"

}

}

```

---

Key Performance Indicators (KPIs)

Documentation KPIs

  1. Documentation Coverage

```

Coverage = (Completed Documents / Total Required Documents) Γ— 100%

Target: β‰₯95%

```

  1. Documentation Debt

```

Debt = Outdated Documents + Missing Documents

Target: <5% of total documents

```

  1. Average Document Age

```

Age = Days since last update

Target: <30 days for active documents

```

Requirement KPIs

  1. Requirement Coverage

```

Coverage = (Requirements with Tests / Total Requirements) Γ— 100%

Target: 100% for MUST requirements

```

  1. Requirement Cycle Time

```

Cycle Time = Days from draft to deployed

Target: <14 days (2 sprints)

```

  1. Orphaned Requirements

```

Orphaned = Requirements without forward/backward links

Target: 0

```

Implementation KPIs

  1. Implementation Velocity

```

Velocity = Tasks completed / Time period

Track: Weekly, calculate 4-week moving average

```

  1. Task Completion Rate

```

Completion = (Completed Tasks / Total Tasks) Γ— 100%

Target: β‰₯90% per sprint

```

  1. Blocked Task Percentage

```

Blocked % = (Blocked Tasks / Active Tasks) Γ— 100%

Target: <10%

```

Traceability KPIs

  1. Forward Traceability

```

Forward = (Requirements with downstream links / Total Requirements) Γ— 100%

Target: 100%

```

  1. Backward Traceability

```

Backward = (Tasks with upstream links / Total Tasks) Γ— 100%

Target: 100%

```

  1. Complete Traceability Chains

```

Complete = (Requirements with full BRD→Code chain / Total) × 100%

Target: β‰₯95%

```

Quality KPIs

  1. Code Coverage Trend

```

Track: Weekly, target: β‰₯90%

Alert: If drops >5% week-over-week

```

  1. Code Review Score

```

Track: Per PR, target: β‰₯80/100

Alert: If average <70 for 2 consecutive weeks

```

  1. Security Vulnerability Density

```

Density = Critical vulnerabilities / KLOC

Target: 0 critical vulnerabilities

```

---

Report Templates

Executive Summary Template

```markdown

# Executive Project Summary

Date: {date}

Period: {period}

Overall Status: {status_indicator}

Progress

  • Phase Completion: {phase_progress}%
  • On Schedule: {schedule_status}
  • Budget Status: {budget_status}

Key Achievements

  1. {achievement_1}
  2. {achievement_2}
  3. {achievement_3}

Risks & Mitigation

| Risk | Impact | Probability | Mitigation |

|------|--------|-------------|------------|

| {risk_1} | {impact} | {prob} | {mitigation} |

Next Milestone

  • {milestone_name}: {milestone_date}
  • {deliverables}

Resource Requirements

  • {resource_needs}

```

Technical Report Template

```markdown

# Technical Project Report

Date: {date}

Architecture

  • ADRs Approved: {adr_count}
  • Major Decisions: {decisions_summary}

Implementation

  • Requirements Completed: {req_completed}/{req_total}
  • Code Coverage: {coverage}%
  • Technical Debt: {tech_debt_score}

Quality Metrics

  • Code Review Score: {review_score}/100
  • Security Issues: {security_count}
  • Performance Benchmarks: {performance_data}

Infrastructure

  • Environments: {env_status}
  • Deployments: {deployment_count}
  • Uptime: {uptime}%

Traceability

  • Forward Coverage: {forward_trace}%
  • Backward Coverage: {backward_trace}%
  • Gaps: {gap_count}

```

---

Data Sources

Primary Sources

  • All SDD artifact files (BRD, PRD, REQ, BDD, ADR, SPEC, TASK, CTR)
  • Git commit history
  • Test execution results
  • Code coverage reports
  • Security scan results
  • CI/CD pipeline logs

Derived Metrics

  • Velocity calculations
  • Trend analysis
  • Predictive forecasting
  • Anomaly detection

---

Visualization Options

Chart Types

  1. Burndown Chart: Sprint/project progress
  2. Velocity Chart: Weekly implementation velocity
  3. Coverage Trend: Test coverage over time
  4. Traceability Matrix: Requirement-to-code mapping
  5. Dependency Graph: Task dependencies
  6. Risk Heat Map: Risk probability vs impact
  7. Quality Dashboard: Multi-metric visualization

Export Formats

  • HTML dashboards (interactive)
  • PDF reports (static)
  • JSON data (for integration)
  • Markdown reports (for Git)
  • Mermaid diagrams (for documentation)
  • CSV data (for spreadsheets)

---

Tool Access

Required tools:

  • Read: Read all SDD artifacts
  • Glob: Find all documents
  • Grep: Search for patterns and IDs
  • Bash: Execute analysis scripts

Required libraries:

  • pandas: Data analysis
  • matplotlib/plotly: Visualization
  • jinja2: Report templating
  • pyyaml: YAML parsing

---

Integration Points

With doc-flow

  • Extract metadata from generated documents
  • Track documentation workflow progress
  • Identify documentation gaps

With test-automation

  • Aggregate test coverage data
  • Track test execution trends
  • Correlate tests with requirements

With code-review

  • Track code quality trends
  • Aggregate review scores
  • Identify quality patterns

With project-mngt

  • Validate timeline adherence
  • Track milestone progress
  • Resource utilization analysis

---

Best Practices

  1. Update frequency: Daily automated collection, weekly reports
  2. Baseline establishment: Set baseline metrics at project start
  3. Trend analysis: Focus on trends, not absolute numbers
  4. Actionable insights: Every metric should inform decisions
  5. Anomaly detection: Alert on unusual patterns
  6. Historical data: Maintain 6+ months of history
  7. Stakeholder alignment: Customize reports per audience
  8. Continuous improvement: Evolve metrics based on feedback

---

Limitations

  1. Requires consistent SDD artifact structure
  2. Historical analysis limited to available data
  3. Predictive accuracy depends on data quality
  4. Cannot measure subjective quality factors
  5. Requires regular data collection for accuracy

---

Success Criteria

  • Real-time dashboard availability
  • <5 minute report generation time
  • 100% data accuracy
  • All KPIs tracked automatically
  • Trend analysis with 90% forecast accuracy
  • Zero manual data collection required
  • Stakeholder satisfaction β‰₯4/5

---

Notes

  • Reports saved to reports/analytics/
  • Historical data stored in metrics/history.json
  • Dashboard auto-refreshes every 15 minutes
  • Alerts configured for KPI threshold breaches
  • Custom metrics configurable via .analytics.yml

More from this repository10

🎯
n8n🎯Skill

Develops custom workflows, nodes, and integrations for n8n automation platform using JavaScript, TypeScript, and multi-service connectors.

🎯
google-adk🎯Skill

Enables rapid development of intelligent, multi-agent AI systems using Google's Agent Development Kit in Python.

🎯
project-init🎯Skill

Initializes new projects with AI Dev Flow framework, creating structured project setup and domain-aware configurations.

🎯
doc-ref🎯Skill

Generates supplementary reference documents (REF) for Business Requirements Documents (BRD) and Architecture Decision Records (ADR) with flexible, non-blocking documentation.

🎯
doc-prd🎯Skill

Generates comprehensive product requirement documents (PRDs) by transforming business requirements into structured, AI-assisted product specifications.

🎯
doc-req🎯Skill

Generates atomic, implementation-ready requirements documents using REQ v3.0 format with comprehensive traceability and readiness scoring across system development layers.

🎯
devops-flow🎯Skill

devops-flow skill from vladm3105/aidoc-flow-framework

🎯
contract-tester🎯Skill

Validates API contracts by performing comprehensive schema checks, detecting breaking changes, and ensuring provider-consumer compatibility across different versions.

🎯
doc-tasks🎯Skill

Decomposes technical specifications into structured, AI-generated TODO tasks for systematic implementation planning.

🎯
doc-flow🎯Skill

doc-flow skill from vladm3105/aidoc-flow-framework