EAGLE3-Qwen3-14B-YARN-64K
Model Description
This model extends AngelSlim/Qwen3-14B_eagle3 with YARN-based positional interpolation to support context lengths of up to 64K tokens.
It is designed to serve as the draft model in self-speculative decoding for long-context generation, as described in the SpecPV paper.
Citation
To cite the model, please use:
@article{tan2025specpv,
title={SpecPV: Improving Self-Speculative Decoding for Long-Context Generation via Partial Verification},
author={Tan, Zhendong and Zhang, Xingjun and Hu, Chaoyi and Peng, Junjie and Xia, Kun},
journal={arXiv preprint arXiv:2512.02337},
year={2025}
}
- Downloads last month
- 11
Model tree for TanBaby/EAGLE3-Qwen3-14B-YARN-64K
Base model
AngelSlim/Qwen3-14B_eagle3