About
This skill equips Claude with specialized knowledge for the Hugging Face Transformers ecosystem, providing standardized patterns for model loading, tokenization, and deployment. It covers advanced implementation details such as 4-bit/8-bit quantization, Parameter-Efficient Fine-Tuning (PEFT) with LoRA, and high-level Pipeline usage for tasks like text classification and question answering. It is ideal for developers building production-grade NLP applications who need to balance model performance with memory efficiency and inference speed.