alphaXiv

Explore

State of the Art

Sign In

Labs

Feedback

Browser Extension

We're hiring
PaperBlogResources

On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention

BibTex
Copy
@misc{wang2025ontheflyadaptivedistillation,
      title={On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear  Attention},
      author={Zhangyang Wang and Zhenyu Zhang and Souvik Kundu and Aditya Akella and Yeonju Ro},
      year={2025},
      eprint={2506.09316},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2506.09316},
}
GitHub
DSLA-Serve
3
HTTPS
https://github.com/utnslab/DSLA-Serve
SSH
git@github.com:utnslab/DSLA-Serve.git
CLI
gh repo clone utnslab/DSLA-Serve
Transform this paper into an audio lecture
Get an engaging lecture and Q&A format to quickly understand the paper in minutes, perfect for learning on the go.
Audio lecture
Q&A format