Multimodal Speculative Decoding
Collection
9 items • Updated
• 1
This model repo is part of a multimodal speculative decoding benchmark suite.
We maintain a unified benchmark codebase that includes multiple methods (Baseline, EAGLE, EAGLE2, EAGLE3, Lookahead, MSD, ViSpec) so users can run training/evaluation more easily under one setup.
Qwen/Qwen2.5-VL-7B-Instructconfig.jsonpytorch_model.binThis checkpoint is intended to be loaded as the EAGLE3 draft/speculative model together with the base model above.
bash scripts/Qwen/eval_eagle3_mmspec.sh testmini Cloudriver/EAGLE3-Qwen2.5-VL-7B-Instruct
If you use this checkpoint and benchmark, please cite EAGLE3 and the baseline methods you compare against.
@inproceedings{li2024eagle,
author = {Yuhui Li and Fangyun Wei and Chao Zhang and Hongyang Zhang},
title = {{EAGLE}: Speculative Sampling Requires Rethinking Feature Uncertainty},
booktitle = {International Conference on Machine Learning},
year = {2024}
}
@inproceedings{li2024eagle2,
author = {Yuhui Li and Fangyun Wei and Chao Zhang and Hongyang Zhang},
title = {{EAGLE-2}: Faster Inference of Language Models with Dynamic Draft Trees},
booktitle = {Empirical Methods in Natural Language Processing},
year = {2024}
}
@inproceedings{li2025eagle3,
author = {Yuhui Li and Fangyun Wei and Chao Zhang and Hongyang Zhang},
title = {{EAGLE-3}: Scaling up Inference Acceleration of Large Language Models via Training-Time Test},
booktitle = {Annual Conference on Neural Information Processing Systems},
year = {2025}
}
Base model
Qwen/Qwen2.5-VL-7B-Instruct