agent-evaluation
π―Skillfrom glennguilloux/context-engineering-kit
Evaluates AI agent performance through comprehensive metrics, benchmarking, and systematic analysis of response quality, accuracy, and contextual understanding.
Same repository
glennguilloux/context-engineering-kit(55 items)
Installation
npx skills add https://github.com/glennguilloux/context-engineering-kit --skill agent-evaluationNeed more details? View full documentation on GitHub β
More from this repository10
Guides developers through writing automated tests before code, ensuring robust, modular software design and preventing regressions.
Rapidly brainstorm and generate innovative concepts across domains by leveraging structured ideation techniques and creative prompting strategies.
Designs scalable system architectures, evaluates design patterns, and provides architectural recommendations for complex software projects.
Provides deep contextual reasoning and explanation generation for complex problems, breaking down root causes and logical connections.
Generates comprehensive unit, integration, and end-to-end test suites with best practices, covering edge cases and ensuring robust code quality across different frameworks.
Enables systematic cognitive decomposition of complex problems through structured reasoning frameworks, logical inference chaining, and meta-cognitive analysis techniques.
Generates comprehensive self-analysis and introspective insights about Claude's current context, reasoning, and potential biases during conversation.
Generates comprehensive project roadmaps with detailed milestones, resource allocation, risk assessment, and actionable implementation strategies.
Evaluates code quality, complexity, and potential issues by analyzing syntax, design patterns, and best practices across multiple programming languages.
Automatically links code reviews to pull requests, streamlining collaboration and tracking feedback across GitHub repositories.