1 results for tag "model-quantization"
AI model quantization and optimization skill covering 4-bit/8-bit quantization, GGUF format conversion, memory optimization, and quality-performance tradeoffs for deploying LLMs in resource-constrained environments