01Model selection guides for Qwen, DeepSeek, Llama, and Phi series
02Comparative analysis of alignment methods including DPO, ORPO, KTO, and SimPO
03Detailed configuration templates for LoRA, QLoRA, and DoRA optimization
040 GitHub stars
05Troubleshooting matrices for common training issues like overfitting and loss divergence
06Architecture deep-dives into Dense, Mixture of Experts (MoE), and MLA