linzhao-amd commited on
Commit
bd3481d
·
verified ·
1 Parent(s): 469c869

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -19,12 +19,11 @@ base_model:
19
  - **Activation quantization:** OCP MXFP4
20
  - **Calibration Dataset:** [Pile](https://huggingface.co/datasets/mit-han-lab/pile-val-backup)
21
 
22
- The model is the quantized version of the [DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1) model, which is an auto-regressive language model that uses an optimized transformer architecture. The MXFP4 model is quantized with [AMD-Quark](https://quark.docs.amd.com/latest/index.html).
23
-
24
 
25
  # Model Quantization
26
 
27
- This model was obtained by quantizing [DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1)'s weights and activations to MXFP4, using AutoSmoothQuant algorithm in [AMD-Quark](https://quark.docs.amd.com/latest/index.html).
28
 
29
  **Quantization scripts:**
30
  ```
 
19
  - **Activation quantization:** OCP MXFP4
20
  - **Calibration Dataset:** [Pile](https://huggingface.co/datasets/mit-han-lab/pile-val-backup)
21
 
22
+ This model is a quantized version of [deepseek-ai/DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1)optimized using [AMD-Quark](https://quark.docs.amd.com/latest/index.html) framework with MXFP4 quantization.
 
23
 
24
  # Model Quantization
25
 
26
+ The model was quantized from [deepseek-ai/DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1) using [AMD-Quark](https://quark.docs.amd.com/latest/index.html). Weights and activations were quantized to MXFP4. The AutoSmoothQuant algorithm was applied to enhance accuracy during quantization.
27
 
28
  **Quantization scripts:**
29
  ```