nobrand commited on
Commit
1d23fa2
·
verified ·
1 Parent(s): 136f303

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -4
README.md CHANGED
@@ -12,7 +12,7 @@ language:
12
 
13
  # KULLM-R
14
 
15
- Introducing KULLM-R, a large language model specialized for high-level reasoning queries in Korean, with a particular focus on complex mathematical problems. The model is designed to provide both the correct reasoning paths and answers for such queries, offering enhanced reasoning efficiency and language transferability to Korean compared to general-purpose reasoning models. It leverages reinforcement learning strategies for effective path exploration and Korean-specific generation.
16
 
17
 
18
  ## Model Details
@@ -34,7 +34,7 @@ KULLM-R is distinguished from standard reasoning LLMs based on Qwen3-8B by its f
34
  - **Reasoning Efficiency Aware Reinforcement Learning**: Introduces RL techniques considering both reasoning path efficiency and answer correctness, reducing unnecessary steps while maintaining answer quality.
35
  - **Reasoning Path Pruning**: Specialized for high-difficulty reasoning problems by pruning ineffective paths and emphasizing transparency and readability in generated answers.
36
  - **Support High Readability in Korean System**: Enhanced both logical reasoning and natural Korean expression ability in answer.
37
- - **Adaptive Length Penalty**: Adaptive penalties optimize the reasoning process according to the question’s complexity and reasoning path length, ensuring efficient solutions for complex problems.
38
 
39
 
40
  ## Data & Training Process
@@ -105,7 +105,7 @@ print("content:", content)
105
  ```
106
 
107
  > [!NOTE]
108
- > As mentioned in Qwen3, use `Temperature=0.6`, `TopP=0.95`, `TopK=20`, and `MinP=0` (the default setting in `generation_config.json`). **DO NOT use greedy decoding**, as it can lead to performance degradation and endless repetitions. For more detailed guidance, please refer to the [Best Practices](#best-practices) section.
109
 
110
 
111
  ## Evaluation
@@ -128,7 +128,6 @@ print("content:", content)
128
 
129
 
130
  ## Citation
131
-
132
  ```
133
  @misc{KULLM-R2025,
134
  title = {KULLM-R: Korea University Large Language Model for Reasoning},
 
12
 
13
  # KULLM-R
14
 
15
+ Introducing KULLM-R, a large language model specialized for high-level reasoning queries in Korean, with a particular focus on complex mathematical problems. The model is designed to provide both the correct reasoning paths and answers for such queries, offering enhanced reasoning efficiency and language transferability to Korean compared to general-purpose reasoning models. Reinforcement learning strategy is employed for efficient reasoning path exploration and Korean-specific generation.
16
 
17
 
18
  ## Model Details
 
34
  - **Reasoning Efficiency Aware Reinforcement Learning**: Introduces RL techniques considering both reasoning path efficiency and answer correctness, reducing unnecessary steps while maintaining answer quality.
35
  - **Reasoning Path Pruning**: Specialized for high-difficulty reasoning problems by pruning ineffective paths and emphasizing transparency and readability in generated answers.
36
  - **Support High Readability in Korean System**: Enhanced both logical reasoning and natural Korean expression ability in answer.
37
+ - **Adaptive Length Penalty**: Adaptive penalties optimize the reasoning process according to the question’s complexity and difficulty, ensuring efficient solutions for various math problems.
38
 
39
 
40
  ## Data & Training Process
 
105
  ```
106
 
107
  > [!NOTE]
108
+ > As mentioned in Qwen3, use `Temperature=0.6`, `TopP=0.95`, `TopK=20`, and `MinP=0` (the default setting in `generation_config.json`). **DO NOT use greedy decoding**, as it can lead to performance degradation and endless repetitions.
109
 
110
 
111
  ## Evaluation
 
128
 
129
 
130
  ## Citation
 
131
  ```
132
  @misc{KULLM-R2025,
133
  title = {KULLM-R: Korea University Large Language Model for Reasoning},