Update README.md
Browse files
README.md
CHANGED
|
@@ -12,7 +12,7 @@ language:
|
|
| 12 |
|
| 13 |
# KULLM-R
|
| 14 |
|
| 15 |
-
Introducing KULLM-R, a large language model specialized for high-level reasoning queries in Korean, with a particular focus on complex mathematical problems. The model is designed to provide both the correct reasoning paths and answers for such queries, offering enhanced reasoning efficiency and language transferability to Korean compared to general-purpose reasoning models.
|
| 16 |
|
| 17 |
|
| 18 |
## Model Details
|
|
@@ -34,7 +34,7 @@ KULLM-R is distinguished from standard reasoning LLMs based on Qwen3-8B by its f
|
|
| 34 |
- **Reasoning Efficiency Aware Reinforcement Learning**: Introduces RL techniques considering both reasoning path efficiency and answer correctness, reducing unnecessary steps while maintaining answer quality.
|
| 35 |
- **Reasoning Path Pruning**: Specialized for high-difficulty reasoning problems by pruning ineffective paths and emphasizing transparency and readability in generated answers.
|
| 36 |
- **Support High Readability in Korean System**: Enhanced both logical reasoning and natural Korean expression ability in answer.
|
| 37 |
-
- **Adaptive Length Penalty**: Adaptive penalties optimize the reasoning process according to the question’s complexity and
|
| 38 |
|
| 39 |
|
| 40 |
## Data & Training Process
|
|
@@ -105,7 +105,7 @@ print("content:", content)
|
|
| 105 |
```
|
| 106 |
|
| 107 |
> [!NOTE]
|
| 108 |
-
> As mentioned in Qwen3, use `Temperature=0.6`, `TopP=0.95`, `TopK=20`, and `MinP=0` (the default setting in `generation_config.json`). **DO NOT use greedy decoding**, as it can lead to performance degradation and endless repetitions.
|
| 109 |
|
| 110 |
|
| 111 |
## Evaluation
|
|
@@ -128,7 +128,6 @@ print("content:", content)
|
|
| 128 |
|
| 129 |
|
| 130 |
## Citation
|
| 131 |
-
|
| 132 |
```
|
| 133 |
@misc{KULLM-R2025,
|
| 134 |
title = {KULLM-R: Korea University Large Language Model for Reasoning},
|
|
|
|
| 12 |
|
| 13 |
# KULLM-R
|
| 14 |
|
| 15 |
+
Introducing KULLM-R, a large language model specialized for high-level reasoning queries in Korean, with a particular focus on complex mathematical problems. The model is designed to provide both the correct reasoning paths and answers for such queries, offering enhanced reasoning efficiency and language transferability to Korean compared to general-purpose reasoning models. Reinforcement learning strategy is employed for efficient reasoning path exploration and Korean-specific generation.
|
| 16 |
|
| 17 |
|
| 18 |
## Model Details
|
|
|
|
| 34 |
- **Reasoning Efficiency Aware Reinforcement Learning**: Introduces RL techniques considering both reasoning path efficiency and answer correctness, reducing unnecessary steps while maintaining answer quality.
|
| 35 |
- **Reasoning Path Pruning**: Specialized for high-difficulty reasoning problems by pruning ineffective paths and emphasizing transparency and readability in generated answers.
|
| 36 |
- **Support High Readability in Korean System**: Enhanced both logical reasoning and natural Korean expression ability in answer.
|
| 37 |
+
- **Adaptive Length Penalty**: Adaptive penalties optimize the reasoning process according to the question’s complexity and difficulty, ensuring efficient solutions for various math problems.
|
| 38 |
|
| 39 |
|
| 40 |
## Data & Training Process
|
|
|
|
| 105 |
```
|
| 106 |
|
| 107 |
> [!NOTE]
|
| 108 |
+
> As mentioned in Qwen3, use `Temperature=0.6`, `TopP=0.95`, `TopK=20`, and `MinP=0` (the default setting in `generation_config.json`). **DO NOT use greedy decoding**, as it can lead to performance degradation and endless repetitions.
|
| 109 |
|
| 110 |
|
| 111 |
## Evaluation
|
|
|
|
| 128 |
|
| 129 |
|
| 130 |
## Citation
|
|
|
|
| 131 |
```
|
| 132 |
@misc{KULLM-R2025,
|
| 133 |
title = {KULLM-R: Korea University Large Language Model for Reasoning},
|