--- library_name: transformers tags: - quantization - 4-bit - chain-of-zoom - super-resolution - complete - bitsandbytes base_model: Qwen/Qwen2.5-VL-3B-Instruct license: apache-2.0 language: - en pipeline_tag: image-to-image --- # Chain-of-Zoom Complete 4-bit Quantized Pipeline ## 📋 Model Description Complete 4-bit quantized Chain-of-Zoom pipeline with all models This model is part of the **Chain-of-Zoom 4-bit Quantized Pipeline** - a memory-optimized version of the original Chain-of-Zoom super-resolution framework. ## 🎯 Key Features - **4-bit Quantization**: Uses BitsAndBytes NF4 quantization for 75% memory reduction - **Maintained Quality**: Comparable performance to full precision models - **Google Colab Compatible**: Runs on T4 GPU (16GB VRAM) - **Memory Efficient**: Optimized for low-resource environments ## 📊 Quantization Details - **Method**: BitsAndBytes NF4 4-bit quantization - **Compute dtype**: bfloat16/float16 - **Double quantization**: Enabled - **Memory reduction**: ~75% compared to original - **Original memory**: ~12GB → **Quantized**: ~3GB ## 🚀 Usage ```python # Install required packages pip install transformers accelerate bitsandbytes torch # Load quantized model from transformers import BitsAndBytesConfig import torch # 4-bit quantization config bnb_config = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_quant_type="nf4", bnb_4bit_use_double_quant=True, bnb_4bit_compute_dtype=torch.bfloat16 ) # Model-specific loading code here # (See complete notebook for detailed usage) ``` ## 📈 Performance - **Quality**: Maintained performance vs full precision - **Speed**: 2-3x faster inference - **Memory**: 75% reduction in VRAM usage - **Hardware**: Compatible with T4, V100, A100 GPUs ## 🔧 Technical Specifications - **Created**: 2025-06-08 17:12:22 - **Quantization Library**: BitsAndBytes - **Framework**: PyTorch + Transformers - **Precision**: 4-bit NF4 - **Model Size**: 1.0 MB ## 📝 Citation ```bibtex @misc{chain-of-zoom-4bit-complete, title={Chain-of-Zoom 4-bit Quantized Chain-of-Zoom Complete 4-bit Quantized Pipeline}, author={humbleakh}, year={2024}, publisher={Hugging Face}, url={https://huggingface.co/humbleakh/chain-of-zoom-4bit-complete} } ``` ## 🔗 Related Models - [Complete Chain-of-Zoom 4-bit Pipeline](humbleakh/chain-of-zoom-4bit-complete) - [Original Chain-of-Zoom](https://github.com/bryanswkim/Chain-of-Zoom) ## ⚠️ Limitations - Requires BitsAndBytes library for proper loading - May have slight quality differences compared to full precision - Optimized for inference, not fine-tuning ## 📄 License Apache 2.0 - See original model licenses for specific components.