Update README.md
Browse files
README.md
CHANGED
|
@@ -17,9 +17,13 @@ datasets:
|
|
| 17 |
- anthracite-org/kalo_misc_part2
|
| 18 |
tags:
|
| 19 |
- chat
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
---
|
| 21 |
|
| 22 |
-
|
| 23 |
|
| 24 |
|
| 25 |
# Quants
|
|
@@ -52,7 +56,6 @@ I would highly recommend using Sao10k's Euryale System prompt, But the "Roleplay
|
|
| 52 |
Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
|
| 53 |
|
| 54 |
<Guidelines>
|
| 55 |
-
• Write upto 200 words.
|
| 56 |
• Maintain the character persona but allow it to evolve with the story.
|
| 57 |
• Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
|
| 58 |
• All types of outputs are encouraged; respond accordingly to the narrative.
|
|
@@ -66,7 +69,6 @@ Currently, your role is {{char}}, described in detail below. As {{char}}, contin
|
|
| 66 |
</Guidelines>
|
| 67 |
|
| 68 |
<Forbidden>
|
| 69 |
-
• Writing more then 200 words.
|
| 70 |
• Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
|
| 71 |
• Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
|
| 72 |
• Repetitive and monotonous outputs.
|
|
@@ -102,7 +104,7 @@ load_in_4bit: false
|
|
| 102 |
strict: false
|
| 103 |
|
| 104 |
datasets:
|
| 105 |
-
- path:
|
| 106 |
type: sharegpt
|
| 107 |
conversation: chatml
|
| 108 |
- path: anthracite-org/kalo-opus-instruct-22k-no-refusal
|
|
@@ -201,21 +203,10 @@ special_tokens:
|
|
| 201 |
- [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
|
| 202 |
- [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
|
| 203 |
- [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
|
| 204 |
-
- [
|
| 205 |
|
| 206 |
|
| 207 |
## Training
|
| 208 |
-
The training was done for
|
| 209 |
-
|
| 210 |
-
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
| 211 |
-
|
| 212 |
-
## Safety
|
| 213 |
-
|
| 214 |
-
Avoid misusing this model, or you’ll need a ‘clicker’ to reset reality. ;)
|
| 215 |
-
|
| 216 |
-
## Musings
|
| 217 |
-
|
| 218 |
-
One of the members of Anthracite had quite an interesting idea, to finetune a smaller model for 4 epochs at a lower Learning rate as quote "Smaller models learn slower" - [Kalomaze](https://huggingface.co/kalomaze) provided access to 10 X A40s and We finetuned what now is [Tor-8B]() for 2.5 epochs (and it's 4 Epoch version released as [Darkens-8B]()) and the result was quite impressive and the same configuration being used to train [Magnum=9B-V4] & [Odin-9B]. We also finetuned the model at above the 8192 context length to see if the model could "heal" in a way to a context length of 16384 with Needle tests coming soon ;)
|
| 219 |
-
|
| 220 |
-
|
| 221 |
|
|
|
|
|
|
| 17 |
- anthracite-org/kalo_misc_part2
|
| 18 |
tags:
|
| 19 |
- chat
|
| 20 |
+
language:
|
| 21 |
+
- en
|
| 22 |
+
base_model:
|
| 23 |
+
- nvidia/Mistral-NeMo-Minitron-8B-Base
|
| 24 |
---
|
| 25 |
|
| 26 |
+
An earlier checkpoint of [Darkens-8B] using the same configuration, Finetuned ontop of the Prune/Distill NeMo 8B done by Nvidia, This model aims to have generally good prose and writing.
|
| 27 |
|
| 28 |
|
| 29 |
# Quants
|
|
|
|
| 56 |
Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
|
| 57 |
|
| 58 |
<Guidelines>
|
|
|
|
| 59 |
• Maintain the character persona but allow it to evolve with the story.
|
| 60 |
• Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
|
| 61 |
• All types of outputs are encouraged; respond accordingly to the narrative.
|
|
|
|
| 69 |
</Guidelines>
|
| 70 |
|
| 71 |
<Forbidden>
|
|
|
|
| 72 |
• Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
|
| 73 |
• Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
|
| 74 |
• Repetitive and monotonous outputs.
|
|
|
|
| 104 |
strict: false
|
| 105 |
|
| 106 |
datasets:
|
| 107 |
+
- path: PRIVATE CLAUDE LOG FILTER
|
| 108 |
type: sharegpt
|
| 109 |
conversation: chatml
|
| 110 |
- path: anthracite-org/kalo-opus-instruct-22k-no-refusal
|
|
|
|
| 203 |
- [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
|
| 204 |
- [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
|
| 205 |
- [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
|
| 206 |
+
- [Private Claude Log filter](https://google.com)
|
| 207 |
|
| 208 |
|
| 209 |
## Training
|
| 210 |
+
The training was done for 4 epochs. (This model is the 2 epoch checkpoint), I used 10 x [A40s](https://www.nvidia.com/en-us/data-center/a40/) GPUs graciously provided by [Kalomaze](https://huggingface.co/kalomaze) for the full-parameter fine-tuning of the model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 211 |
|
| 212 |
+
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|