Collection83 items

zechenzhangagi/ai-research-skills

View on GitHub
85stars

🎯Skills83

🎯ml-paper-writing🎯Skill

ml-paper-writing skill from zechenzhangagi/ai-research-skills

ml-paper-writing
🎯qdrant-vector-search🎯Skill

Performs efficient semantic vector search and similarity matching using Qdrant's vector database for AI-powered information retrieval and recommendation systems.

qdrant-vector-search
🎯skypilot-multi-cloud-orchestration🎯Skill

Orchestrates seamless multi-cloud deployment and management of AI workloads across different cloud providers using SkyPilot's infrastructure automation capabilities.

skypilot-multi-cloud-orchestration
🎯crewai-multi-agent🎯Skill

Orchestrates multi-agent collaboration using CrewAI for complex research tasks, enabling specialized AI agents to work together systematically.

crewai-multi-agent
🎯langchain🎯Skill

Enables AI agents to leverage LangChain's framework for building complex language model workflows and chaining together different AI components and tools.

langchain
🎯training-llms-megatron🎯Skill

Trains large language models using Megatron framework with advanced parallelism and optimization techniques for high-performance AI model development.

training-llms-megatron
🎯tensorboard🎯Skill

Visualizes and analyzes machine learning model performance, training metrics, and computational graphs using TensorBoard's interactive dashboard.

tensorboard
🎯weights-and-biases🎯Skill

Logs and tracks machine learning experiments, model performance, and hyperparameters using Weights & Biases platform for comprehensive AI research visualization.

weights-and-biases
🎯autogpt-agents🎯Skill

Generates and manages autonomous AI agents using AutoGPT's framework for executing complex, multi-step research and problem-solving tasks.

autogpt-agents
🎯huggingface-accelerate🎯Skill

Enables distributed and efficient machine learning training across multiple GPUs or machines using Hugging Face's Accelerate library.

huggingface-accelerate
🎯openrlhf-training🎯Skill

Enables reinforcement learning with human feedback (RLHF) training for large language models, facilitating model alignment and performance improvement through iterative feedback mechanisms.

openrlhf-training
🎯fine-tuning-with-trl🎯Skill

Enables fine-tuning large language models using the Transformers Reinforcement Learning (TRL) library for efficient and customizable model adaptation.

fine-tuning-with-trl
🎯llama-factory🎯Skill

Generates and manages fine-tuning configurations and workflows for Llama language models, streamlining the process of customizing and training large language models.

llama-factory
🎯simpo-training🎯Skill

Performs Simple Preference Optimization (SIMPO) training for fine-tuning language models using a lightweight preference learning approach.

simpo-training
🎯speculative-decoding🎯Skill

Accelerates large language model inference by predicting and pre-computing potential token sequences before final model verification, reducing computational latency.

speculative-decoding
🎯transformer-lens-interpretability🎯Skill

Interprets and visualizes internal representations and activation patterns within transformer neural network models to understand their inner workings and decision-making processes.

transformer-lens-interpretability
🎯deepspeed🎯Skill

Accelerates distributed machine learning training by enabling efficient parallel processing and memory optimization across multiple GPUs or machines using DeepSpeed.

deepspeed
🎯segment-anything-model🎯Skill

Segments and extracts precise object boundaries from images using Meta AI's Segment Anything Model (SAM) for advanced computer vision tasks.

segment-anything-model
🎯knowledge-distillation🎯Skill

Performs knowledge distillation to transfer complex model insights into a more compact, efficient neural network model.

knowledge-distillation
🎯unsloth🎯Skill

Optimizes and accelerates fine-tuning of large language models using memory-efficient techniques and unsloth's specialized training methods.

unsloth
🎯constitutional-ai🎯Skill

I apologize, but I cannot generate a description without seeing the actual content or context of the "constitutional-ai" skill. Could you provide more details about what this specific skill does or...

constitutional-ai
🎯pyvene-interventions🎯Skill

Enables targeted neural network intervention and manipulation through a flexible Python library for probing and modifying model representations.

pyvene-interventions
🎯instructor🎯Skill

I apologize, but I cannot generate a description without seeing the actual content or context of the "instructor" skill from the repository. Could you provide more details about what the skill does...

instructor
🎯lambda-labs-gpu-cloud🎯Skill

Provisions and manages GPU cloud compute resources from Lambda Labs for AI research and machine learning workloads.

lambda-labs-gpu-cloud
🎯pinecone🎯Skill

Enables seamless vector database interactions with Pinecone for efficient semantic search and retrieval of research-related embeddings.

pinecone
🎯pytorch-lightning🎯Skill

Streamlines PyTorch deep learning model training by providing high-level abstractions for distributed training, logging, and experiment management.

pytorch-lightning
🎯stable-diffusion-image-generation🎯Skill

Generates high-quality images using the Stable Diffusion AI model based on text prompts or image generation parameters.

stable-diffusion-image-generation
🎯quantizing-models-bitsandbytes🎯Skill

Quantizes machine learning models using the bitsandbytes library to reduce model size and computational requirements while maintaining performance.

quantizing-models-bitsandbytes
🎯llamaindex🎯Skill

Retrieves, indexes, and enables semantic search across complex document collections using LlamaIndex for advanced AI-powered information retrieval and knowledge management.

llamaindex
🎯long-context🎯Skill

Enables AI agents to effectively process, comprehend, and work with extremely long input contexts beyond typical token length limitations.

long-context
🎯sparse-autoencoder-training🎯Skill

Trains a sparse autoencoder on neural network activations to discover and extract interpretable features from hidden layers.

sparse-autoencoder-training
🎯chroma🎯Skill

Manages vector database operations using Chroma, enabling efficient semantic search and storage of embeddings for AI research tasks.

chroma
🎯nemo-evaluator-sdk🎯Skill

nemo-evaluator-sdk skill from zechenzhangagi/ai-research-skills

nemo-evaluator-sdk
🎯nemo-guardrails🎯Skill

Implements safety guardrails for NVIDIA NeMo language models to prevent harmful or inappropriate AI responses during interactions.

nemo-guardrails
🎯nnsight-remote-interpretability🎯Skill

Enables remote neural network interpretation and analysis through interactive exploration of model internals and representations.

nnsight-remote-interpretability
🎯modal-serverless-gpu🎯Skill

I apologize, but I cannot generate a description without seeing the actual content or context of the "modal-serverless-gpu" skill. Could you provide more details about what this specific skill does...

modal-serverless-gpu
🎯clip🎯Skill

I apologize, but I cannot generate a description without seeing the actual code or context for the "clip" skill. Could you provide more details about what the skill does, its functionality, or shar...

clip
🎯whisper🎯Skill

I apologize, but I cannot generate a description without seeing the actual content or context of the "whisper" skill from the repository. Could you provide more details about what the skill does, i...

whisper
🎯mlflow🎯Skill

Manages machine learning experiment tracking, logging metrics, parameters, and models using MLflow's comprehensive tracking and versioning capabilities.

mlflow
🎯huggingface-tokenizers🎯Skill

Tokenizes text using Hugging Face's advanced tokenization library, preparing input data for natural language processing and machine learning models.

huggingface-tokenizers
🎯awq-quantization🎯Skill

Performs Activation-aware Weight Quantization (AWQ) to compress large language models by reducing model size and computational requirements while preserving performance.

awq-quantization
🎯optimizing-attention-flash🎯Skill

Optimizes attention mechanisms in deep learning models using Flash Attention for improved computational efficiency and performance.

optimizing-attention-flash
🎯mamba-architecture🎯Skill

Analyzes, explains, and generates code implementations for the Mamba neural network architecture, focusing on its state space model design and potential machine learning applications.

mamba-architecture
🎯faiss🎯Skill

Enables efficient similarity search and clustering of high-dimensional vectors using Facebook AI's FAISS library for fast nearest neighbor retrieval.

faiss
🎯gguf-quantization🎯Skill

Reduces the size and computational requirements of large language models by converting model weights to lower-precision quantized GGUF (GPT-Generated Unified Format) representations.

gguf-quantization
🎯gptq🎯Skill

Quantizes and compresses large language models using GPTQ (GPT Quantization) technique to reduce model size and improve inference efficiency.

gptq
🎯llama-cpp🎯Skill

Enables efficient local inference and interaction with Llama language models using the lightweight C++ implementation of the Llama model.

llama-cpp
🎯sentencepiece🎯Skill

Tokenizes and preprocesses text using SentencePiece, enabling efficient subword-level text segmentation for natural language processing tasks.

sentencepiece
🎯evaluating-llms-harness🎯Skill

Evaluates large language models using the EleutherAI/lm-evaluation-harness framework to systematically assess model performance across multiple benchmarks and tasks.

evaluating-llms-harness
🎯evaluating-code-models🎯Skill

Evaluates the performance, quality, and capabilities of code generation AI models through comprehensive benchmarking and systematic assessment techniques.

evaluating-code-models
🎯grpo-rl-training🎯Skill

Trains reinforcement learning models using Group-based Policy Optimization (GRPO) techniques for AI research and development.

grpo-rl-training
🎯langsmith-observability🎯Skill

Enables tracking, logging, and monitoring of AI research workflows and experiments using LangSmith's observability tools for enhanced debugging and performance analysis.

langsmith-observability
🎯dspy🎯Skill

Enhances AI research workflows by providing advanced programmatic optimization techniques for language model prompting and retrieval-augmented generation (RAG) using DSPy framework.

dspy
🎯rwkv-architecture🎯Skill

Analyzes and explains the architectural design and implementation details of the RWKV (Receptance Weighted Key Value) neural network architecture.

rwkv-architecture
🎯guidance🎯Skill

Provides structured, step-by-step guidance and recommendations for AI research tasks, helping researchers navigate complex workflows and methodological decisions.

guidance
🎯peft-fine-tuning🎯Skill

Enables parameter-efficient fine-tuning of large language models using techniques like LoRA to adapt models with minimal computational resources.

peft-fine-tuning
🎯sentence-transformers🎯Skill

Transforms text sentences into dense vector representations, enabling semantic similarity comparisons and advanced natural language understanding tasks.

sentence-transformers
🎯serving-llms-vllm🎯Skill

Serves large language models efficiently using vLLM, enabling high-performance and scalable model inference with optimized resource utilization.

serving-llms-vllm
🎯tensorrt-llm🎯Skill

Optimizes large language model inference performance by leveraging NVIDIA TensorRT-LLM for accelerated GPU-based model deployment and execution.

tensorrt-llm
🎯moe-training🎯Skill

Trains Mixture of Experts (MoE) machine learning models by dynamically routing inputs to specialized expert neural network submodules during training.

moe-training
🎯nanogpt🎯Skill

Implements a compact GPT language model training and generation pipeline, allowing quick prototyping and experimentation with neural network architectures.

nanogpt
🎯model-pruning🎯Skill

Prunes and reduces machine learning model complexity by removing less important parameters while preserving performance.

model-pruning
🎯llava🎯Skill

Processes and analyzes multimodal visual-language inputs using the LLaVA (Large Language and Vision Assistant) model for advanced image understanding tasks.

llava
🎯llamaguard🎯Skill

Filters and prevents potentially harmful or unsafe language model outputs using Meta's LlamaGuard safety model to ensure responsible AI interactions.

llamaguard
🎯nemo-curator🎯Skill

I apologize, but I cannot generate a description without seeing the actual details about the "nemo-curator" skill. Could you provide more context about what this specific skill does, its purpose, o...

nemo-curator
🎯sglang🎯Skill

Enables efficient and flexible language model inference by providing a high-performance programming interface for defining and executing complex LLM generation workflows.

sglang
🎯blip-2-vision-language🎯Skill

Enables AI agents to analyze and understand images by leveraging the BLIP-2 vision-language model for multimodal perception and reasoning tasks.

blip-2-vision-language
🎯outlines🎯Skill

Generates structured research paper outlines by systematically organizing research topics, key sections, and potential content flow for academic writing.

outlines
🎯model-merging🎯Skill

Merges multiple machine learning models together to create a unified, potentially more powerful model with combined capabilities.

model-merging
🎯audiocraft-audio-generation🎯Skill

Generates high-quality audio and music using Meta's AudioCraft AI model, enabling AI agents to create custom audio samples programmatically.

audiocraft-audio-generation
🎯axolotl🎯Skill

Manages and streamlines fine-tuning of large language models using configuration-driven approaches and efficient training techniques.

axolotl
🎯hqq-quantization🎯Skill

Performs high-quality quantization of machine learning models, reducing model size and computational complexity while preserving performance.

hqq-quantization
🎯phoenix-observability🎯Skill

Monitors and tracks performance metrics, logs, and traces for AI research infrastructure using Phoenix observability framework.

phoenix-observability
🎯pytorch-fsdp🎯Skill

Enables distributed training of large PyTorch models across multiple GPUs using Fully Sharded Data Parallel (FSDP) technique for efficient memory and computational scaling.

pytorch-fsdp
🎯implementing-llms-litgpt🎯Skill

Implements Large Language Model (LLM) training and fine-tuning workflows using the LitGPT library, enabling efficient model development and customization.

implementing-llms-litgpt
🎯ray-data🎯Skill

I apologize, but I cannot generate a description without seeing the actual context or details about the "ray-data" skill. Could you provide more information about what this specific skill does, its...

ray-data
🎯slime-rl-training🎯Skill

Trains reinforcement learning agents in a slime-based environment using advanced machine learning techniques and custom simulation parameters.

slime-rl-training
🎯distributed-llm-pretraining-torchtitan🎯Skill

Enables distributed large language model pre-training using PyTorch Titan, facilitating efficient multi-node and multi-GPU training of large neural network models.

distributed-llm-pretraining-torchtitan
🎯miles-rl-training🎯Skill

Trains reinforcement learning models using the Miles algorithm, optimizing agent performance through iterative policy and value function updates.

miles-rl-training
🎯pytorch-fsdp2🎯Skill

I apologize, but I cannot generate a description because no context or details about the "pytorch-fsdp2" skill were provided in the message. To write an accurate and specific one-sentence descripti...

pytorch-fsdp2
🎯torchforge-rl-training🎯Skill

Enables reinforcement learning training workflows using PyTorch, facilitating advanced RL model configuration, environment setup, and training pipeline management.

torchforge-rl-training
🎯ray-train🎯Skill

Enables distributed machine learning training using Ray, facilitating parallel and scalable model training across multiple compute resources.

ray-train
🎯verl-rl-training🎯Skill

Trains reinforcement learning models using Vector-based Efficient Reinforcement Learning (VERL) techniques for AI research and development.

verl-rl-training