4 results for tag "optimizing-attention-flash"
A large collection of Claude Code skill templates sponsored by Z.AI, providing ready-to-use development skill configurations across various domains.
A Flash Attention optimization skill from the AI Research Engineering Skills Library, providing guidance on optimizing attention mechanisms for efficient transformer inference as part of 83 AI research skills.
Skill for optimizing transformer attention mechanisms with Flash Attention, covering memory-efficient implementations and kernel optimization techniques.