01Advanced activation caching across all transformer layers and components
02Integration with SAELens for Sparse Autoencoder (SAE) research
03Support for 50+ model families including GPT-2, LLaMA, and Mistral
043,983 GitHub stars
05Direct logit attribution and circuit discovery patterns
06Activation patching and causal tracing for model behavior analysis