BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-Distillation
Paper
•
2402.10631
•
Published
•
2
| PPL | arc_easy | arc_challenge | piqa | winogrande | hellaswag | mmlu | QA Avg |
|---|---|---|---|---|---|---|---|
| 22.39 | 36.11 ± 0.99 | 21.84 ± 1.21 | 60.72 ± 1.14 | 51.38 ± 1.40 | 32.73 ± 0.47 | - | 40.56 |
Training method based on BitDistiller Paper
Base model
TinyLlama/TinyLlama_v1.1