deepspeed
π―Skillfrom orchestra-research/ai-research-skills
Accelerates AI model training and inference with optimized distributed computing, memory efficiency, and GPU utilization techniques
Part of
orchestra-research/ai-research-skills(104 items)
Installation
npx skills add https://github.com/orchestra-research/ai-research-skills --skill deepspeedNeed more details? View full documentation on GitHub β
More from this repository10
Streamlines AI research workflows by providing curated Claude skills for data analysis, literature review, experiment design, and research paper generation.
Assists AI researchers in drafting, structuring, and generating machine learning research papers with academic writing best practices and technical precision.
Streamlines distributed data processing and machine learning workflows using Ray's scalable data loading and transformation capabilities.
Streamlines distributed machine learning training using Ray, optimizing hyperparameter tuning and parallel model execution across compute clusters.
Automates multi-agent AI task workflows with dynamic role assignment, collaborative problem-solving, and intelligent task delegation across complex research scenarios.
Enables high-performance local inference and quantization for large language models using efficient C/C++ implementations and GGML formats.
Implements and evaluates RWKV language model architectures, providing tools for training, fine-tuning, and performance analysis of linear attention transformer alternatives.
Monitors and visualizes LangChain application performance, tracking token usage, latency, and error rates for comprehensive AI workflow insights
Streamlines systematic evaluation of large language models through comprehensive benchmarking, performance metrics, and comparative analysis frameworks.
Perform targeted neural network interventions and activation manipulations to probe model behavior and understand internal representations in PyTorch.