cpatonn commited on
Commit
dda0c4f
·
verified ·
1 Parent(s): 548664e

Delete files recipe.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. recipe.yaml +0 -46
recipe.yaml DELETED
@@ -1,46 +0,0 @@
1
- default_stage:
2
- default_modifiers:
3
- AWQModifier:
4
- config_groups:
5
- group_0:
6
- targets: [Linear]
7
- weights:
8
- num_bits: 8
9
- type: int
10
- symmetric: true
11
- group_size: 32
12
- strategy: group
13
- block_structure: null
14
- dynamic: false
15
- actorder: null
16
- scale_dtype: null
17
- zp_dtype: null
18
- observer: mse
19
- observer_kwargs: {}
20
- input_activations: null
21
- output_activations: null
22
- format: null
23
- targets: [Linear]
24
- ignore: [lm_head, model.embed_tokens, 're:.*shared_experts.*', 're:model[.]layers[.]0[.].*',
25
- 're:.*mlp[.]gate$']
26
- bypass_divisibility_checks: false
27
- mappings:
28
- - smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*input_layernorm$
29
- balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*q_a_proj$', 're:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*kv_a_proj_with_mqa$']
30
- activation_hook_target: null
31
- - smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*q_a_layernorm$
32
- balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*q_b_proj$']
33
- activation_hook_target: null
34
- - smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*kv_a_layernorm$
35
- balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*kv_b_proj$']
36
- activation_hook_target: null
37
- - smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*post_attention_layernorm$
38
- balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*gate_proj$', 're:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*up_proj$',
39
- 're:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*mlp[.]gate$']
40
- activation_hook_target: null
41
- - smooth_layer: re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*mlp[.]experts.*up_proj$
42
- balance_layers: ['re:.*layers[.]((?:[1-9]|[1-3][0-9]|4[0-6]))[.].*mlp[.]experts.*down_proj$']
43
- activation_hook_target: null
44
- offload_device: !!python/object/apply:torch.device [cuda]
45
- duo_scaling: true
46
- n_grid: 20