DataSoul lbourdois commited on
Commit
8131921
·
verified ·
1 Parent(s): 9e08132

Improve language tag (#1)

Browse files

- Improve language tag (26f27307f20e7ecd1a1d83c44448738b440f1c54)


Co-authored-by: Loïck BOURDOIS <lbourdois@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +77 -64
README.md CHANGED
@@ -1,64 +1,77 @@
1
- ---
2
- base_model:
3
- - huihui-ai/QwQ-32B-abliterated
4
- - zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated
5
- - Qwen/Qwen2.5-32B
6
- - DataSoul/QAQ-32B-merge3
7
- library_name: transformers
8
- tags:
9
- - mergekit
10
- - merge
11
-
12
- ---
13
- Unstable "thinking" and "reasoning" models, which typically respond in four scenarios:
14
-
15
- 1 (occasionally), &lt;think&gt;...&lt;/think&gt; answer.
16
-
17
- 2 (occasionally), &lt;think&gt;... answer.
18
-
19
- 3 (occasionally), &lt;think&gt;... .
20
-
21
- 4 (rarely), answer.
22
-
23
- I don't know what to do next in order to get a stable, reasoning, completely uncensored model at the same time.
24
- If you have any innovative ideas, I warmly invite you to join the discussion or conduct your own experiments.
25
-
26
- More recommended [DataSoul/QAQ-32B-merge3](https://huggingface.co/DataSoul/QAQ-32B-merge3)But it is still not a 'thinking' model.
27
-
28
-
29
- # merge
30
-
31
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
32
-
33
- ## Merge Details
34
- ### Merge Method
35
-
36
- This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [Qwen/Qwen2.5-32B](https://huggingface.co/Qwen/Qwen2.5-32B) as a base.
37
-
38
- ### Models Merged
39
-
40
- The following models were included in the merge:
41
- * [huihui-ai/QwQ-32B-abliterated](https://huggingface.co/huihui-ai/QwQ-32B-abliterated)
42
- * [zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated](https://huggingface.co/zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated)
43
- * [DataSoul/QAQ-32B-merge3](https://huggingface.co/DataSoul/QAQ-32B-merge3)
44
-
45
- ### Configuration
46
-
47
- The following YAML configuration was used to produce this model:
48
-
49
- ```yaml
50
- models:
51
- # Pivot model
52
- - model: Qwen/Qwen2.5-32B
53
- # Target models
54
- - model: huihui-ai/QwQ-32B-abliterated
55
- - model: DataSoul/QAQ-32B-merge3
56
- - model: zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated
57
- merge_method: sce
58
- base_model: Qwen/Qwen2.5-32B
59
- tokenizer_source: zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated
60
- parameters:
61
- select_topk: 1.0
62
- dtype: bfloat16
63
-
64
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - huihui-ai/QwQ-32B-abliterated
4
+ - zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated
5
+ - Qwen/Qwen2.5-32B
6
+ - DataSoul/QAQ-32B-merge3
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ ---
26
+ Unstable "thinking" and "reasoning" models, which typically respond in four scenarios:
27
+
28
+ 1 (occasionally), &lt;think&gt;...&lt;/think&gt; answer.
29
+
30
+ 2 (occasionally), &lt;think&gt;... answer.
31
+
32
+ 3 (occasionally), &lt;think&gt;... .
33
+
34
+ 4 (rarely), answer.
35
+
36
+ I don't know what to do next in order to get a stable, reasoning, completely uncensored model at the same time.
37
+ If you have any innovative ideas, I warmly invite you to join the discussion or conduct your own experiments.
38
+
39
+ More recommended [DataSoul/QAQ-32B-merge3](https://huggingface.co/DataSoul/QAQ-32B-merge3)But it is still not a 'thinking' model.
40
+
41
+
42
+ # merge
43
+
44
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
45
+
46
+ ## Merge Details
47
+ ### Merge Method
48
+
49
+ This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [Qwen/Qwen2.5-32B](https://huggingface.co/Qwen/Qwen2.5-32B) as a base.
50
+
51
+ ### Models Merged
52
+
53
+ The following models were included in the merge:
54
+ * [huihui-ai/QwQ-32B-abliterated](https://huggingface.co/huihui-ai/QwQ-32B-abliterated)
55
+ * [zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated](https://huggingface.co/zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated)
56
+ * [DataSoul/QAQ-32B-merge3](https://huggingface.co/DataSoul/QAQ-32B-merge3)
57
+
58
+ ### Configuration
59
+
60
+ The following YAML configuration was used to produce this model:
61
+
62
+ ```yaml
63
+ models:
64
+ # Pivot model
65
+ - model: Qwen/Qwen2.5-32B
66
+ # Target models
67
+ - model: huihui-ai/QwQ-32B-abliterated
68
+ - model: DataSoul/QAQ-32B-merge3
69
+ - model: zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated
70
+ merge_method: sce
71
+ base_model: Qwen/Qwen2.5-32B
72
+ tokenizer_source: zetasepic/Rombo-LLM-V3.1-QWQ-32b-abliterated
73
+ parameters:
74
+ select_topk: 1.0
75
+ dtype: bfloat16
76
+
77
+ ```