Spaces:
Sleeping
Sleeping
metadata
title: Qwen Training
emoji: 🧙
colorFrom: purple
colorTo: blue
sdk: gradio
sdk_version: 5.49.1
app_file: app.py
pinned: false
license: mit
suggested_hardware: zero-a10g
short_description: app.py is base workkng
PromptWizard Qwen Fine-tuning
This Space fine-tunes Qwen models using Gita dataset wit optimization methodology.
Features
- GPU-Accelerated Training: Uses HuggingFace's GPU infrastructure for fast training
- LoRA Fine-tuning: Efficient parameter-efficient fine-tuning
- GITA Dataset: High-quality use any custiom reasoning dataset
- PromptWizard Integration: Uses Microsoft's PromptWizard evaluation methodology
- Auto Push to Hub: Trained models are automatically uploaded to HuggingFace Hub
How to Use
- Select your base model (default: Qwen/Qwen2.5-7B)
- Configure training parameters:
- Number of epochs (3-5 recommended)
- Batch size (4-8 for T4 GPU)
- Learning rate (2e-5 is a good default)
- Click "Start Training" and monitor the output
- The trained model will be pushed to HuggingFace Hub
Training Data
The Space uses the GITA dataset, which contains grade school math problems. The data is formatted according to PromptWizard specifications for optimal prompt optimization.
Model Output
After training, the model will be available at:
- HuggingFace Hub:
your-username/promptwizard-qwen-gsm8k - Local download: Available in the Space's output directory
Technical Details
- Base Model: Qwen2.5-7B (or your choice)
- Training Method: LoRA with rank 16
- Quantization: 8-bit for memory efficiency
- Mixed Precision: FP16 for faster training
- Gradient Checkpointing: Enabled for memory savings
Resource Requirements
- GPU: T4 or better recommended
- Memory: 16GB+ GPU memory
- Training Time: ~30-60 minutes on T4
Citation
If you use this training setup, please cite:
@misc{promptwizard2024,
title={PromptWizard: Task-Aware Prompt Optimization},
author={Microsoft Research},
year={2024}
}