01Automatic 4-bit quantization and gradient checkpointing
02Seamless export to GGUF, Ollama, and Hugging Face formats
031 GitHub stars
042x faster training speeds with 80% VRAM reduction
05Support for Llama 3.3, Mistral, Phi 3.5, and Gemma 2
06Hardware-specific performance optimization profiles