01Full QLoRA (4-bit quantization) support for memory-constrained environments
02Efficient parameter management reducing trainable weights to ~0.1%
03Multi-adapter management for task switching without reloading base models
04Best practices for rank (r) and alpha scaling optimization
0510 GitHub stars
06Pre-configured target modules for popular architectures like Llama, Mistral, and BERT