On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent Paper • 2410.04870 • Published Oct 7, 2024
AdaSPEC: Selective Knowledge Distillation for Efficient Speculative Decoders Paper • 2510.19779 • Published Oct 22 • 60
Efficient Hyperparameter Tuning via Trajectory Invariance Principle Paper • 2509.25049 • Published Sep 29 • 4
SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable Sparse-Linear Attention Paper • 2509.24006 • Published Sep 28 • 118
Efficient Hyperparameter Tuning via Trajectory Invariance Principle Paper • 2509.25049 • Published Sep 29 • 4
SageAttention3: Microscaling FP4 Attention for Inference and An Exploration of 8-Bit Training Paper • 2505.11594 • Published May 16 • 75