pytorch-quantization
π―Skillfrom cuba6112/skillfactory
Streamlines model compression and performance optimization in PyTorch by implementing quantization techniques for reduced memory and faster inference
Part of
cuba6112/skillfactory(27 items)
Installation
npx skills add https://github.com/cuba6112/skillfactory --skill pytorch-quantizationNeed more details? View full documentation on GitHub β
More from this repository10
Generates Minecraft Pocket Edition (MCPE) mod and addon build configurations with automated resource packaging and version management
Streamlines PyTorch model training with simplified distributed computing, logging, checkpointing, and advanced training workflows across GPUs and TPUs.
Automates Notion research workflows by extracting, organizing, and structuring research notes, sources, and insights into clean, actionable documentation.
Perform local retrieval-augmented generation using Ollama models, enabling context-aware Q&A and document analysis without external API dependencies.
Automates capturing and organizing web content, articles, and research into structured Notion databases with intelligent tagging and metadata extraction.
Deploys and manages machine learning models using PyTorch's TorchServe framework for scalable and efficient model serving
Streamlines PyTorch model development with optimized tensor operations, neural network layers, and training workflows for deep learning projects.
Fine-tune large language models efficiently using LoRA with optimized memory and performance for machine learning practitioners.
Transforms Notion specifications into executable code, bridging design requirements with technical implementation across programming languages.
Converts PyTorch models to ONNX format for cross-platform inference, optimizing deep learning model deployment and performance.