Delete files recipe.yaml with huggingface_hub
Browse files- recipe.yaml +0 -46
recipe.yaml
DELETED
|
@@ -1,46 +0,0 @@
|
|
| 1 |
-
default_stage:
|
| 2 |
-
default_modifiers:
|
| 3 |
-
AWQModifier:
|
| 4 |
-
config_groups:
|
| 5 |
-
group_0:
|
| 6 |
-
targets: [Linear]
|
| 7 |
-
weights:
|
| 8 |
-
num_bits: 8
|
| 9 |
-
type: int
|
| 10 |
-
symmetric: true
|
| 11 |
-
group_size: 32
|
| 12 |
-
strategy: group
|
| 13 |
-
block_structure: null
|
| 14 |
-
dynamic: false
|
| 15 |
-
actorder: null
|
| 16 |
-
scale_dtype: null
|
| 17 |
-
zp_dtype: null
|
| 18 |
-
observer: mse
|
| 19 |
-
observer_kwargs: {}
|
| 20 |
-
input_activations: null
|
| 21 |
-
output_activations: null
|
| 22 |
-
format: null
|
| 23 |
-
targets: [Linear]
|
| 24 |
-
ignore: [lm_head, model.embed_tokens, 're:.*shared_experts.*', 're:model[.]layers[.]0[.].*',
|
| 25 |
-
're:.*mlp[.]gate$']
|
| 26 |
-
bypass_divisibility_checks: false
|
| 27 |
-
mappings:
|
| 28 |
-
- smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*input_layernorm$
|
| 29 |
-
balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*q_a_proj$', 're:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*kv_a_proj_with_mqa$']
|
| 30 |
-
activation_hook_target: null
|
| 31 |
-
- smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*q_a_layernorm$
|
| 32 |
-
balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*q_b_proj$']
|
| 33 |
-
activation_hook_target: null
|
| 34 |
-
- smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*kv_a_layernorm$
|
| 35 |
-
balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*kv_b_proj$']
|
| 36 |
-
activation_hook_target: null
|
| 37 |
-
- smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*post_attention_layernorm$
|
| 38 |
-
balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*gate_proj$', 're:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*up_proj$',
|
| 39 |
-
're:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*mlp[.]gate$']
|
| 40 |
-
activation_hook_target: null
|
| 41 |
-
- smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*mlp[.]experts.*up_proj$
|
| 42 |
-
balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*mlp[.]experts.*down_proj$']
|
| 43 |
-
activation_hook_target: null
|
| 44 |
-
offload_device: !!python/object/apply:torch.device [cuda]
|
| 45 |
-
duo_scaling: true
|
| 46 |
-
n_grid: 20
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|