Update README.md
Browse files
README.md
CHANGED
|
@@ -24,16 +24,11 @@ Trained on __20T+ tokens of high-quality data__, together with __supervised fine
|
|
| 24 |
### Powerful Complex Reasoning Abilities
|
| 25 |
|
| 26 |
We conducted a comprehensive evaluation of Ling-flash-2.0’s reasoning capabilities, reporting strong results on representative benchmarks:
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
● __Challenging code generation__: LiveCodeBench v6, CodeForces-Elo
|
| 33 |
-
|
| 34 |
-
● __Logical reasoning__: KOR-Bench, ARC-Prize
|
| 35 |
-
|
| 36 |
-
● __Key regulated industries (Finance, Healthcare)__: FinanceReasoning, HealthBench
|
| 37 |
|
| 38 |
Compared with __dense models under 40B__ (e.g., Qwen3-32B-Non-Thinking, Seed-OSS-36B-Instruct (think budget=0)) and __larger-activation/total-parameter MoE models__ (e.g., Hunyuan-A13B-Instruct, GPT-OSS-120B/low), __Ling-flash-2.0__ demonstrates stronger complex reasoning power. Moreover, it shows high competitiveness on __creative tasks__ (Creative Writing v3).
|
| 39 |
<p align="center">
|
|
@@ -244,7 +239,7 @@ More usage can be found [here](https://docs.sglang.ai/basic_usage/send_request.h
|
|
| 244 |
|
| 245 |
### Finetuning
|
| 246 |
|
| 247 |
-
We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ling](https://github.com/inclusionAI/Ling-V2/blob/main/docs/llamafactory_finetuning.md).
|
| 248 |
|
| 249 |
## License
|
| 250 |
|
|
|
|
| 24 |
### Powerful Complex Reasoning Abilities
|
| 25 |
|
| 26 |
We conducted a comprehensive evaluation of Ling-flash-2.0’s reasoning capabilities, reporting strong results on representative benchmarks:
|
| 27 |
+
* __Multi-disciplinary knowledge reasoning__: GPQA-Diamond, MMLU-Pro
|
| 28 |
+
* __Advanced mathematical reasoning__: AIME 2025, Omni-MATH, OptMATH (advanced mathematical optimization tasks)
|
| 29 |
+
* __Challenging code generation__: LiveCodeBench v6, CodeForces-Elo
|
| 30 |
+
* __Logical reasoning__: KOR-Bench, ARC-Prize
|
| 31 |
+
* __Key regulated industries (Finance, Healthcare)__: FinanceReasoning, HealthBench
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
|
| 33 |
Compared with __dense models under 40B__ (e.g., Qwen3-32B-Non-Thinking, Seed-OSS-36B-Instruct (think budget=0)) and __larger-activation/total-parameter MoE models__ (e.g., Hunyuan-A13B-Instruct, GPT-OSS-120B/low), __Ling-flash-2.0__ demonstrates stronger complex reasoning power. Moreover, it shows high competitiveness on __creative tasks__ (Creative Writing v3).
|
| 34 |
<p align="center">
|
|
|
|
| 239 |
|
| 240 |
### Finetuning
|
| 241 |
|
| 242 |
+
We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ling](https://github.com/inclusionAI/Ling-V2/blob/main/docs/llamafactory_finetuning.md).
|
| 243 |
|
| 244 |
## License
|
| 245 |
|