HamzaYous89 commited on
Commit
8722387
·
verified ·
1 Parent(s): 6955034

Upload folder using huggingface_hub

Browse files
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
+ ' + message['content'] + '<|im_end|>' + '
3
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
+ ' }}{% endif %}
config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "FalconH1ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attention_in_multiplier": 1.0,
8
+ "attention_out_multiplier": 0.5,
9
+ "attn_layer_indices": null,
10
+ "bos_token_id": 1,
11
+ "embedding_multiplier": 5.656854249492381,
12
+ "eos_token_id": 11,
13
+ "head_dim": 128,
14
+ "hidden_act": "silu",
15
+ "hidden_size": 1280,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 3072,
18
+ "key_multiplier": 0.17677669529663687,
19
+ "lm_head_multiplier": 0.03125,
20
+ "mamba_chunk_size": 128,
21
+ "mamba_conv_bias": true,
22
+ "mamba_d_conv": 4,
23
+ "mamba_d_head": 64,
24
+ "mamba_d_ssm": 1536,
25
+ "mamba_d_state": 256,
26
+ "mamba_expand": 2,
27
+ "mamba_n_groups": 1,
28
+ "mamba_n_heads": 24,
29
+ "mamba_norm_before_gate": false,
30
+ "mamba_proj_bias": false,
31
+ "mamba_rms_norm": true,
32
+ "mamba_use_mlp": true,
33
+ "max_position_embeddings": 8192,
34
+ "mlp_bias": false,
35
+ "mlp_expansion_factor": 8,
36
+ "mlp_multipliers": [
37
+ 0.7071067811865476,
38
+ 0.3125
39
+ ],
40
+ "model_type": "falcon_h1",
41
+ "num_attention_heads": 6,
42
+ "num_hidden_layers": 66,
43
+ "num_key_value_heads": 2,
44
+ "num_logits_to_keep": 1,
45
+ "pad_token_id": 0,
46
+ "projectors_bias": false,
47
+ "quantization_config": {
48
+ "bits": 4,
49
+ "checkpoint_format": "gptq",
50
+ "desc_act": true,
51
+ "group_size": 128,
52
+ "lm_head": false,
53
+ "meta": {
54
+ "damp_auto_increment": 0.0025,
55
+ "damp_percent": 0.01,
56
+ "mse": 0.0,
57
+ "quantizer": [
58
+ "gptqmodel:2.2.0"
59
+ ],
60
+ "static_groups": false,
61
+ "true_sequential": true,
62
+ "uri": "https://github.com/modelcloud/gptqmodel"
63
+ },
64
+ "pack_dtype": "int32",
65
+ "quant_method": "gptq",
66
+ "sym": true
67
+ },
68
+ "rms_norm_eps": 1e-05,
69
+ "rope_scaling": null,
70
+ "rope_theta": 100000000000.0,
71
+ "ssm_in_multiplier": 1.0,
72
+ "ssm_multipliers": [
73
+ 0.3535533905932738,
74
+ 0.25,
75
+ 0.1767766952966369,
76
+ 0.5,
77
+ 0.3535533905932738
78
+ ],
79
+ "ssm_out_multiplier": 0.23570226039551587,
80
+ "tie_word_embeddings": false,
81
+ "torch_dtype": "bfloat16",
82
+ "transformers_version": "4.52.0.dev0",
83
+ "use_cache": true,
84
+ "vocab_size": 65537
85
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 11,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.52.0.dev0"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a6a12cee53fdffbc4dadb8d6a5bf0acedd9babbe32575aadbc0a18d292b7bcf
3
+ size 1958786144
quant_log.csv ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ layer,module,loss,samples,damp,time
2
+ 0,feed_forward.gate_proj,111.74320221,0.01000,3.982
3
+ 0,feed_forward.up_proj,236.62837219,0.01000,0.320
4
+ 0,feed_forward.down_proj,388.95077515,0.01000,0.773
5
+ 1,feed_forward.gate_proj,118.89686584,0.01000,0.319
6
+ 1,feed_forward.up_proj,210.03805542,0.01000,0.319
7
+ 1,feed_forward.down_proj,364.70776367,0.01000,0.768
8
+ 2,feed_forward.gate_proj,122.43881226,0.01000,0.323
9
+ 2,feed_forward.up_proj,197.41491699,0.01000,0.318
10
+ 2,feed_forward.down_proj,309.12982178,0.01000,0.758
11
+ 3,feed_forward.gate_proj,122.79499054,0.01000,0.315
12
+ 3,feed_forward.up_proj,184.00488281,0.01000,0.315
13
+ 3,feed_forward.down_proj,303.67919922,0.01000,0.757
14
+ 4,feed_forward.gate_proj,113.50039673,0.01000,0.315
15
+ 4,feed_forward.up_proj,169.97505188,0.01000,0.317
16
+ 4,feed_forward.down_proj,325.09625244,0.01000,0.760
17
+ 5,feed_forward.gate_proj,126.93116760,0.01000,0.319
18
+ 5,feed_forward.up_proj,188.28265381,0.01000,0.315
19
+ 5,feed_forward.down_proj,289.97735596,0.01000,0.760
20
+ 6,feed_forward.gate_proj,108.90504456,0.01000,0.316
21
+ 6,feed_forward.up_proj,167.51849365,0.01000,0.318
22
+ 6,feed_forward.down_proj,274.22109985,0.01000,0.762
23
+ 7,feed_forward.gate_proj,103.00424194,0.01000,0.315
24
+ 7,feed_forward.up_proj,160.86836243,0.01000,0.315
25
+ 7,feed_forward.down_proj,1013.91198730,0.01000,0.757
26
+ 8,feed_forward.gate_proj,81.34243011,0.01000,0.319
27
+ 8,feed_forward.up_proj,127.04229736,0.01000,0.316
28
+ 8,feed_forward.down_proj,208.30104065,0.01000,0.758
29
+ 9,feed_forward.gate_proj,113.06949615,0.01000,0.315
30
+ 9,feed_forward.up_proj,160.02661133,0.01000,0.315
31
+ 9,feed_forward.down_proj,243.04393005,0.01000,0.758
32
+ 10,feed_forward.gate_proj,99.50776672,0.01000,0.317
33
+ 10,feed_forward.up_proj,146.21742249,0.01000,0.318
34
+ 10,feed_forward.down_proj,194.84396362,0.01000,0.768
35
+ 11,feed_forward.gate_proj,91.98031616,0.01000,0.315
36
+ 11,feed_forward.up_proj,139.31994629,0.01000,0.315
37
+ 11,feed_forward.down_proj,189.97125244,0.01000,0.760
38
+ 12,feed_forward.gate_proj,81.66962433,0.01000,0.316
39
+ 12,feed_forward.up_proj,124.63871002,0.01000,0.315
40
+ 12,feed_forward.down_proj,176.43630981,0.01000,0.762
41
+ 13,feed_forward.gate_proj,81.46422577,0.01000,0.315
42
+ 13,feed_forward.up_proj,127.35617065,0.01000,0.315
43
+ 13,feed_forward.down_proj,173.96453857,0.01000,0.759
44
+ 14,feed_forward.gate_proj,72.19062805,0.01000,0.315
45
+ 14,feed_forward.up_proj,113.07981873,0.01000,0.315
46
+ 14,feed_forward.down_proj,152.21907043,0.01000,0.759
47
+ 15,feed_forward.gate_proj,69.28975677,0.01000,0.315
48
+ 15,feed_forward.up_proj,105.54737854,0.01000,0.315
49
+ 15,feed_forward.down_proj,140.14369202,0.01000,0.758
50
+ 16,feed_forward.gate_proj,72.53057098,0.01000,0.316
51
+ 16,feed_forward.up_proj,109.36140442,0.01000,0.315
52
+ 16,feed_forward.down_proj,142.06488037,0.01000,0.759
53
+ 17,feed_forward.gate_proj,73.08057404,0.01000,0.315
54
+ 17,feed_forward.up_proj,111.05429077,0.01000,0.315
55
+ 17,feed_forward.down_proj,199.73251343,0.01000,0.758
56
+ 18,feed_forward.gate_proj,76.59034729,0.01000,0.316
57
+ 18,feed_forward.up_proj,123.99762726,0.01000,0.316
58
+ 18,feed_forward.down_proj,171.76063538,0.01000,0.760
59
+ 19,feed_forward.gate_proj,83.81452942,0.01000,0.315
60
+ 19,feed_forward.up_proj,133.96391296,0.01000,0.315
61
+ 19,feed_forward.down_proj,161.21109009,0.01000,0.758
62
+ 20,feed_forward.gate_proj,82.09156799,0.01000,0.315
63
+ 20,feed_forward.up_proj,132.04408264,0.01000,0.315
64
+ 20,feed_forward.down_proj,170.59501648,0.01000,0.758
65
+ 21,feed_forward.gate_proj,73.64891052,0.01000,0.315
66
+ 21,feed_forward.up_proj,120.23171997,0.01000,0.315
67
+ 21,feed_forward.down_proj,196.58923340,0.01000,0.758
68
+ 22,feed_forward.gate_proj,96.88229370,0.01000,0.315
69
+ 22,feed_forward.up_proj,125.69157410,0.01000,0.315
70
+ 22,feed_forward.down_proj,167.71240234,0.01000,0.761
71
+ 23,feed_forward.gate_proj,99.65959167,0.01000,0.315
72
+ 23,feed_forward.up_proj,132.40330505,0.01000,0.319
73
+ 23,feed_forward.down_proj,162.09573364,0.01000,0.758
74
+ 24,feed_forward.gate_proj,96.29583740,0.01000,0.315
75
+ 24,feed_forward.up_proj,138.25765991,0.01000,0.314
76
+ 24,feed_forward.down_proj,183.68345642,0.01000,0.758
77
+ 25,feed_forward.gate_proj,98.48146820,0.01000,0.315
78
+ 25,feed_forward.up_proj,143.75862122,0.01000,0.315
79
+ 25,feed_forward.down_proj,172.26818848,0.01000,0.758
80
+ 26,feed_forward.gate_proj,102.55484772,0.01000,0.316
81
+ 26,feed_forward.up_proj,151.94816589,0.01000,0.315
82
+ 26,feed_forward.down_proj,200.07752991,0.01000,0.759
83
+ 27,feed_forward.gate_proj,101.92828369,0.01000,0.316
84
+ 27,feed_forward.up_proj,153.61186218,0.01000,0.316
85
+ 27,feed_forward.down_proj,213.37088013,0.01000,0.759
86
+ 28,feed_forward.gate_proj,107.33745575,0.01000,0.316
87
+ 28,feed_forward.up_proj,160.71836853,0.01000,0.315
88
+ 28,feed_forward.down_proj,225.36859131,0.01000,0.758
89
+ 29,feed_forward.gate_proj,106.04288483,0.01000,0.315
90
+ 29,feed_forward.up_proj,160.50518799,0.01000,0.315
91
+ 29,feed_forward.down_proj,233.17849731,0.01000,0.759
92
+ 30,feed_forward.gate_proj,103.57704926,0.01000,0.316
93
+ 30,feed_forward.up_proj,156.03421021,0.01000,0.314
94
+ 30,feed_forward.down_proj,257.80474854,0.01000,0.762
95
+ 31,feed_forward.gate_proj,112.84368896,0.01000,0.315
96
+ 31,feed_forward.up_proj,172.18254089,0.01000,0.314
97
+ 31,feed_forward.down_proj,337.24584961,0.01000,0.758
98
+ 32,feed_forward.gate_proj,133.08917236,0.01000,0.316
99
+ 32,feed_forward.up_proj,192.55874634,0.01000,0.315
100
+ 32,feed_forward.down_proj,352.67297363,0.01000,0.757
101
+ 33,feed_forward.gate_proj,142.59655762,0.01000,0.314
102
+ 33,feed_forward.up_proj,205.99560547,0.01000,0.316
103
+ 33,feed_forward.down_proj,399.31158447,0.01000,0.755
104
+ 34,feed_forward.gate_proj,143.29968262,0.01000,0.316
105
+ 34,feed_forward.up_proj,211.95402527,0.01000,0.314
106
+ 34,feed_forward.down_proj,496.21578979,0.01000,0.760
107
+ 35,feed_forward.gate_proj,171.93203735,0.01000,0.314
108
+ 35,feed_forward.up_proj,239.84025574,0.01000,0.314
109
+ 35,feed_forward.down_proj,588.40148926,0.01000,0.758
110
+ 36,feed_forward.gate_proj,191.02435303,0.01000,0.314
111
+ 36,feed_forward.up_proj,257.85241699,0.01000,0.315
112
+ 36,feed_forward.down_proj,596.66259766,0.01000,0.763
113
+ 37,feed_forward.gate_proj,212.65242004,0.01000,0.315
114
+ 37,feed_forward.up_proj,282.30514526,0.01000,0.314
115
+ 37,feed_forward.down_proj,709.05725098,0.01000,0.758
116
+ 38,feed_forward.gate_proj,212.11619568,0.01000,0.318
117
+ 38,feed_forward.up_proj,289.99993896,0.01000,0.315
118
+ 38,feed_forward.down_proj,756.64733887,0.01000,0.757
119
+ 39,feed_forward.gate_proj,240.64978027,0.01000,0.318
120
+ 39,feed_forward.up_proj,324.07727051,0.01000,0.314
121
+ 39,feed_forward.down_proj,846.79614258,0.01000,0.756
122
+ 40,feed_forward.gate_proj,248.88653564,0.01000,0.316
123
+ 40,feed_forward.up_proj,337.91979980,0.01000,0.316
124
+ 40,feed_forward.down_proj,944.45239258,0.01000,0.759
125
+ 41,feed_forward.gate_proj,271.32739258,0.01000,0.314
126
+ 41,feed_forward.up_proj,372.99176025,0.01000,0.315
127
+ 41,feed_forward.down_proj,1064.48120117,0.01000,0.759
128
+ 42,feed_forward.gate_proj,279.37274170,0.01000,0.319
129
+ 42,feed_forward.up_proj,391.11901855,0.01000,0.315
130
+ 42,feed_forward.down_proj,1176.21972656,0.01000,0.761
131
+ 43,feed_forward.gate_proj,295.24725342,0.01000,0.316
132
+ 43,feed_forward.up_proj,418.80706787,0.01000,0.315
133
+ 43,feed_forward.down_proj,1368.23937988,0.01000,0.757
134
+ 44,feed_forward.gate_proj,306.15591431,0.01000,0.315
135
+ 44,feed_forward.up_proj,428.31079102,0.01000,0.317
136
+ 44,feed_forward.down_proj,1447.78320312,0.01000,0.758
137
+ 45,feed_forward.gate_proj,333.10876465,0.01000,0.315
138
+ 45,feed_forward.up_proj,454.30328369,0.01000,0.315
139
+ 45,feed_forward.down_proj,1482.55554199,0.01000,0.758
140
+ 46,feed_forward.gate_proj,377.98596191,0.01000,0.315
141
+ 46,feed_forward.up_proj,513.86663818,0.01000,0.316
142
+ 46,feed_forward.down_proj,1538.15014648,0.01000,0.758
143
+ 47,feed_forward.gate_proj,420.98831177,0.01000,0.316
144
+ 47,feed_forward.up_proj,573.67767334,0.01000,0.315
145
+ 47,feed_forward.down_proj,1815.25012207,0.01000,0.759
146
+ 48,feed_forward.gate_proj,468.18768311,0.01000,0.315
147
+ 48,feed_forward.up_proj,643.76281738,0.01000,0.315
148
+ 48,feed_forward.down_proj,1891.54699707,0.01000,0.759
149
+ 49,feed_forward.gate_proj,482.29534912,0.01000,0.314
150
+ 49,feed_forward.up_proj,672.92626953,0.01000,0.317
151
+ 49,feed_forward.down_proj,2082.15966797,0.01000,0.758
152
+ 50,feed_forward.gate_proj,474.23300171,0.01000,0.315
153
+ 50,feed_forward.up_proj,681.62872314,0.01000,0.314
154
+ 50,feed_forward.down_proj,2451.87109375,0.01000,0.757
155
+ 51,feed_forward.gate_proj,516.93835449,0.01000,0.314
156
+ 51,feed_forward.up_proj,758.75915527,0.01000,0.314
157
+ 51,feed_forward.down_proj,2591.85986328,0.01000,0.764
158
+ 52,feed_forward.gate_proj,557.41088867,0.01000,0.314
159
+ 52,feed_forward.up_proj,810.66589355,0.01000,0.317
160
+ 52,feed_forward.down_proj,2903.74438477,0.01000,0.758
161
+ 53,feed_forward.gate_proj,621.95080566,0.01000,0.316
162
+ 53,feed_forward.up_proj,896.72338867,0.01000,0.318
163
+ 53,feed_forward.down_proj,2944.29272461,0.01000,0.759
164
+ 54,feed_forward.gate_proj,623.52075195,0.01000,0.316
165
+ 54,feed_forward.up_proj,915.17077637,0.01000,0.316
166
+ 54,feed_forward.down_proj,3341.56909180,0.01000,0.758
167
+ 55,feed_forward.gate_proj,636.33288574,0.01000,0.315
168
+ 55,feed_forward.up_proj,934.63183594,0.01000,0.315
169
+ 55,feed_forward.down_proj,3537.10156250,0.01000,0.758
170
+ 56,feed_forward.gate_proj,662.20495605,0.01000,0.316
171
+ 56,feed_forward.up_proj,987.23132324,0.01000,0.316
172
+ 56,feed_forward.down_proj,4167.27050781,0.01000,0.760
173
+ 57,feed_forward.gate_proj,665.96362305,0.01000,0.314
174
+ 57,feed_forward.up_proj,1004.44226074,0.01000,0.313
175
+ 57,feed_forward.down_proj,4681.87792969,0.01000,0.757
176
+ 58,feed_forward.gate_proj,683.81127930,0.01000,0.319
177
+ 58,feed_forward.up_proj,1038.61291504,0.01000,0.315
178
+ 58,feed_forward.down_proj,5349.24853516,0.01000,0.758
179
+ 59,feed_forward.gate_proj,719.57690430,0.01000,0.316
180
+ 59,feed_forward.up_proj,1079.96667480,0.01000,0.314
181
+ 59,feed_forward.down_proj,5659.36328125,0.01000,0.758
182
+ 60,feed_forward.gate_proj,769.08300781,0.01000,0.318
183
+ 60,feed_forward.up_proj,1138.22961426,0.01000,0.315
184
+ 60,feed_forward.down_proj,5963.39794922,0.01000,0.759
185
+ 61,feed_forward.gate_proj,820.04296875,0.01000,0.318
186
+ 61,feed_forward.up_proj,1152.85827637,0.01000,0.318
187
+ 61,feed_forward.down_proj,6901.42626953,0.01000,0.759
188
+ 62,feed_forward.gate_proj,842.18273926,0.01000,0.319
189
+ 62,feed_forward.up_proj,1146.75561523,0.01000,0.315
190
+ 62,feed_forward.down_proj,7562.22949219,0.01000,0.758
191
+ 63,feed_forward.gate_proj,902.06835938,0.01000,0.316
192
+ 63,feed_forward.up_proj,1160.74414062,0.01000,0.315
193
+ 63,feed_forward.down_proj,12306.59570312,0.01000,0.761
194
+ 64,feed_forward.gate_proj,1086.16149902,0.01000,0.318
195
+ 64,feed_forward.up_proj,1300.27490234,0.01000,0.318
196
+ 64,feed_forward.down_proj,12706.41406250,0.01000,0.759
197
+ 65,feed_forward.gate_proj,2097.17065430,0.01000,0.316
198
+ 65,feed_forward.up_proj,2490.09667969,0.01000,0.319
199
+ 65,feed_forward.down_proj,34298.35156250,0.01000,0.760
quantize_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "desc_act": true,
5
+ "sym": true,
6
+ "lm_head": false,
7
+ "quant_method": "gptq",
8
+ "checkpoint_format": "gptq",
9
+ "pack_dtype": "int32",
10
+ "meta": {
11
+ "quantizer": [
12
+ "gptqmodel:2.2.0"
13
+ ],
14
+ "uri": "https://github.com/modelcloud/gptqmodel",
15
+ "damp_percent": 0.01,
16
+ "damp_auto_increment": 0.0025,
17
+ "static_groups": false,
18
+ "true_sequential": true,
19
+ "mse": 0.0
20
+ }
21
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,356 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|pad|>",
4
+ ">>ABSTRACT<<",
5
+ ">>INTRODUCTION<<",
6
+ ">>SUMMARY<<",
7
+ ">>COMMENT<<",
8
+ ">>ANSWER<<",
9
+ ">>QUESTION<<",
10
+ ">>DOMAIN<<",
11
+ ">>PREFIX<<",
12
+ ">>SUFFIX<<",
13
+ ">>MIDDLE<<",
14
+ "<|finetune_right_pad_id|>",
15
+ "<|start_header_id|>",
16
+ "<|end_header_id|>",
17
+ "<|eom_id|>",
18
+ "<|eot_id|>",
19
+ "<|begin_of_text|>",
20
+ ">>TITLE<<",
21
+ "<tool_response>",
22
+ "</tool_response>",
23
+ "<tool_call>",
24
+ "</tool_call>",
25
+ "<schema>",
26
+ "</schema>",
27
+ "<scratch_pad>",
28
+ "</scratch_pad>",
29
+ "<thinking>",
30
+ "</thinking>",
31
+ "<explanation>",
32
+ "</explanation>",
33
+ "<file_sep>",
34
+ "<repo_name>",
35
+ "<|im_end|>",
36
+ "<|im_start|>",
37
+ ">>UNUSED_119<<",
38
+ ">>UNUSED_120<<",
39
+ "<|system|>",
40
+ ">>UNUSED_218<<",
41
+ ">>UNUSED_219<<",
42
+ ">>UNUSED_220<<",
43
+ ">>UNUSED_221<<",
44
+ ">>UNUSED_222<<",
45
+ ">>UNUSED_223<<",
46
+ ">>UNUSED_224<<",
47
+ ">>UNUSED_225<<",
48
+ ">>UNUSED_226<<",
49
+ ">>UNUSED_227<<",
50
+ ">>UNUSED_228<<",
51
+ ">>UNUSED_229<<",
52
+ ">>UNUSED_230<<",
53
+ ">>UNUSED_231<<",
54
+ ">>UNUSED_232<<",
55
+ ">>UNUSED_233<<",
56
+ ">>UNUSED_234<<",
57
+ ">>UNUSED_235<<",
58
+ ">>UNUSED_236<<",
59
+ ">>UNUSED_237<<",
60
+ ">>UNUSED_238<<",
61
+ ">>UNUSED_239<<",
62
+ ">>UNUSED_240<<",
63
+ ">>UNUSED_241<<",
64
+ ">>UNUSED_242<<",
65
+ ">>UNUSED_243<<",
66
+ ">>UNUSED_244<<",
67
+ ">>UNUSED_245<<",
68
+ ">>UNUSED_246<<",
69
+ ">>UNUSED_247<<",
70
+ ">>UNUSED_248<<",
71
+ ">>UNUSED_249<<",
72
+ ">>UNUSED_250<<",
73
+ ">>UNUSED_251<<",
74
+ ">>UNUSED_252<<",
75
+ ">>UNUSED_253<<",
76
+ ">>UNUSED_254<<",
77
+ ">>UNUSED_255<<",
78
+ ">>UNUSED_256<<",
79
+ ">>UNUSED_257<<",
80
+ ">>UNUSED_258<<",
81
+ ">>UNUSED_259<<",
82
+ ">>UNUSED_260<<",
83
+ ">>UNUSED_261<<",
84
+ ">>UNUSED_262<<",
85
+ ">>UNUSED_263<<",
86
+ ">>UNUSED_264<<",
87
+ ">>UNUSED_265<<",
88
+ ">>UNUSED_266<<",
89
+ ">>UNUSED_267<<",
90
+ ">>UNUSED_268<<",
91
+ ">>UNUSED_269<<",
92
+ ">>UNUSED_270<<",
93
+ ">>UNUSED_271<<",
94
+ ">>UNUSED_272<<",
95
+ ">>UNUSED_273<<",
96
+ ">>UNUSED_274<<",
97
+ ">>UNUSED_275<<",
98
+ ">>UNUSED_276<<",
99
+ ">>UNUSED_277<<",
100
+ ">>UNUSED_278<<",
101
+ ">>UNUSED_279<<",
102
+ ">>UNUSED_280<<",
103
+ ">>UNUSED_281<<",
104
+ ">>UNUSED_282<<",
105
+ ">>UNUSED_283<<",
106
+ ">>UNUSED_284<<",
107
+ ">>UNUSED_285<<",
108
+ ">>UNUSED_286<<",
109
+ ">>UNUSED_287<<",
110
+ ">>UNUSED_288<<",
111
+ ">>UNUSED_289<<",
112
+ ">>UNUSED_290<<",
113
+ ">>UNUSED_291<<",
114
+ ">>UNUSED_292<<",
115
+ ">>UNUSED_293<<",
116
+ ">>UNUSED_294<<",
117
+ ">>UNUSED_295<<",
118
+ ">>UNUSED_296<<",
119
+ ">>UNUSED_297<<",
120
+ ">>UNUSED_298<<",
121
+ ">>UNUSED_299<<",
122
+ ">>UNUSED_300<<",
123
+ ">>UNUSED_301<<",
124
+ ">>UNUSED_302<<",
125
+ ">>UNUSED_303<<",
126
+ ">>UNUSED_304<<",
127
+ ">>UNUSED_305<<",
128
+ ">>UNUSED_306<<",
129
+ ">>UNUSED_307<<",
130
+ ">>UNUSED_308<<",
131
+ ">>UNUSED_309<<",
132
+ ">>UNUSED_310<<",
133
+ ">>UNUSED_311<<",
134
+ ">>UNUSED_312<<",
135
+ ">>UNUSED_313<<",
136
+ ">>UNUSED_314<<",
137
+ ">>UNUSED_315<<",
138
+ ">>UNUSED_316<<",
139
+ ">>UNUSED_317<<",
140
+ ">>UNUSED_318<<",
141
+ ">>UNUSED_319<<",
142
+ ">>UNUSED_320<<",
143
+ ">>UNUSED_321<<",
144
+ ">>UNUSED_322<<",
145
+ ">>UNUSED_323<<",
146
+ ">>UNUSED_324<<",
147
+ ">>UNUSED_325<<",
148
+ ">>UNUSED_326<<",
149
+ ">>UNUSED_327<<",
150
+ ">>UNUSED_328<<",
151
+ ">>UNUSED_329<<",
152
+ ">>UNUSED_330<<",
153
+ ">>UNUSED_331<<",
154
+ ">>UNUSED_332<<",
155
+ ">>UNUSED_333<<",
156
+ ">>UNUSED_334<<",
157
+ ">>UNUSED_335<<",
158
+ ">>UNUSED_336<<",
159
+ ">>UNUSED_337<<",
160
+ ">>UNUSED_338<<",
161
+ ">>UNUSED_339<<",
162
+ ">>UNUSED_340<<",
163
+ ">>UNUSED_341<<",
164
+ ">>UNUSED_342<<",
165
+ ">>UNUSED_343<<",
166
+ ">>UNUSED_344<<",
167
+ ">>UNUSED_345<<",
168
+ ">>UNUSED_346<<",
169
+ ">>UNUSED_347<<",
170
+ ">>UNUSED_348<<",
171
+ ">>UNUSED_349<<",
172
+ ">>UNUSED_350<<",
173
+ ">>UNUSED_351<<",
174
+ ">>UNUSED_352<<",
175
+ ">>UNUSED_353<<",
176
+ ">>UNUSED_354<<",
177
+ ">>UNUSED_355<<",
178
+ ">>UNUSED_356<<",
179
+ ">>UNUSED_357<<",
180
+ ">>UNUSED_358<<",
181
+ ">>UNUSED_359<<",
182
+ ">>UNUSED_360<<",
183
+ ">>UNUSED_361<<",
184
+ ">>UNUSED_362<<",
185
+ ">>UNUSED_363<<",
186
+ ">>UNUSED_364<<",
187
+ ">>UNUSED_365<<",
188
+ ">>UNUSED_366<<",
189
+ ">>UNUSED_367<<",
190
+ ">>UNUSED_368<<",
191
+ ">>UNUSED_369<<",
192
+ ">>UNUSED_370<<",
193
+ ">>UNUSED_371<<",
194
+ ">>UNUSED_372<<",
195
+ ">>UNUSED_373<<",
196
+ ">>UNUSED_374<<",
197
+ ">>UNUSED_375<<",
198
+ ">>UNUSED_376<<",
199
+ ">>UNUSED_377<<",
200
+ ">>UNUSED_378<<",
201
+ ">>UNUSED_379<<",
202
+ ">>UNUSED_380<<",
203
+ ">>UNUSED_381<<",
204
+ ">>UNUSED_382<<",
205
+ ">>UNUSED_383<<",
206
+ ">>UNUSED_384<<",
207
+ ">>UNUSED_385<<",
208
+ ">>UNUSED_386<<",
209
+ ">>UNUSED_387<<",
210
+ ">>UNUSED_388<<",
211
+ ">>UNUSED_389<<",
212
+ ">>UNUSED_390<<",
213
+ ">>UNUSED_391<<",
214
+ ">>UNUSED_392<<",
215
+ ">>UNUSED_393<<",
216
+ ">>UNUSED_394<<",
217
+ ">>UNUSED_395<<",
218
+ ">>UNUSED_396<<",
219
+ ">>UNUSED_397<<",
220
+ ">>UNUSED_398<<",
221
+ ">>UNUSED_399<<",
222
+ ">>UNUSED_400<<",
223
+ ">>UNUSED_401<<",
224
+ ">>UNUSED_402<<",
225
+ ">>UNUSED_403<<",
226
+ ">>UNUSED_404<<",
227
+ ">>UNUSED_405<<",
228
+ ">>UNUSED_406<<",
229
+ ">>UNUSED_407<<",
230
+ ">>UNUSED_408<<",
231
+ ">>UNUSED_409<<",
232
+ ">>UNUSED_410<<",
233
+ ">>UNUSED_411<<",
234
+ ">>UNUSED_412<<",
235
+ ">>UNUSED_413<<",
236
+ ">>UNUSED_414<<",
237
+ ">>UNUSED_415<<",
238
+ ">>UNUSED_416<<",
239
+ ">>UNUSED_417<<",
240
+ ">>UNUSED_418<<",
241
+ ">>UNUSED_419<<",
242
+ ">>UNUSED_420<<",
243
+ ">>UNUSED_421<<",
244
+ ">>UNUSED_422<<",
245
+ ">>UNUSED_423<<",
246
+ ">>UNUSED_424<<",
247
+ ">>UNUSED_425<<",
248
+ ">>UNUSED_426<<",
249
+ ">>UNUSED_427<<",
250
+ ">>UNUSED_428<<",
251
+ ">>UNUSED_429<<",
252
+ ">>UNUSED_430<<",
253
+ ">>UNUSED_431<<",
254
+ ">>UNUSED_432<<",
255
+ ">>UNUSED_433<<",
256
+ ">>UNUSED_434<<",
257
+ ">>UNUSED_435<<",
258
+ ">>UNUSED_436<<",
259
+ ">>UNUSED_437<<",
260
+ ">>UNUSED_438<<",
261
+ ">>UNUSED_439<<",
262
+ ">>UNUSED_440<<",
263
+ ">>UNUSED_441<<",
264
+ ">>UNUSED_442<<",
265
+ ">>UNUSED_443<<",
266
+ ">>UNUSED_444<<",
267
+ ">>UNUSED_445<<",
268
+ ">>UNUSED_446<<",
269
+ ">>UNUSED_447<<",
270
+ ">>UNUSED_448<<",
271
+ ">>UNUSED_449<<",
272
+ ">>UNUSED_450<<",
273
+ ">>UNUSED_451<<",
274
+ ">>UNUSED_452<<",
275
+ ">>UNUSED_453<<",
276
+ ">>UNUSED_454<<",
277
+ ">>UNUSED_455<<",
278
+ ">>UNUSED_456<<",
279
+ ">>UNUSED_457<<",
280
+ ">>UNUSED_458<<",
281
+ ">>UNUSED_459<<",
282
+ ">>UNUSED_460<<",
283
+ ">>UNUSED_461<<",
284
+ ">>UNUSED_462<<",
285
+ ">>UNUSED_463<<",
286
+ ">>UNUSED_464<<",
287
+ ">>UNUSED_465<<",
288
+ ">>UNUSED_466<<",
289
+ ">>UNUSED_467<<",
290
+ ">>UNUSED_468<<",
291
+ ">>UNUSED_469<<",
292
+ ">>UNUSED_470<<",
293
+ ">>UNUSED_471<<",
294
+ ">>UNUSED_472<<",
295
+ ">>UNUSED_473<<",
296
+ ">>UNUSED_474<<",
297
+ ">>UNUSED_475<<",
298
+ ">>UNUSED_476<<",
299
+ ">>UNUSED_477<<",
300
+ ">>UNUSED_478<<",
301
+ ">>UNUSED_479<<",
302
+ ">>UNUSED_480<<",
303
+ ">>UNUSED_481<<",
304
+ ">>UNUSED_482<<",
305
+ ">>UNUSED_483<<",
306
+ ">>UNUSED_484<<",
307
+ ">>UNUSED_485<<",
308
+ ">>UNUSED_486<<",
309
+ ">>UNUSED_487<<",
310
+ ">>UNUSED_488<<",
311
+ ">>UNUSED_489<<",
312
+ ">>UNUSED_490<<",
313
+ ">>UNUSED_491<<",
314
+ ">>UNUSED_492<<",
315
+ ">>UNUSED_493<<",
316
+ ">>UNUSED_494<<",
317
+ ">>UNUSED_495<<",
318
+ ">>UNUSED_496<<",
319
+ ">>UNUSED_497<<",
320
+ ">>UNUSED_498<<",
321
+ ">>UNUSED_499<<",
322
+ ">>UNUSED_500<<",
323
+ ">>UNUSED_501<<",
324
+ ">>UNUSED_502<<",
325
+ ">>UNUSED_503<<",
326
+ ">>UNUSED_504<<",
327
+ ">>UNUSED_505<<",
328
+ ">>UNUSED_506<<",
329
+ ">>UNUSED_507<<",
330
+ ">>UNUSED_508<<",
331
+ ">>UNUSED_509<<",
332
+ ">>UNUSED_510<<",
333
+ ">>UNUSED_511<<"
334
+ ],
335
+ "bos_token": {
336
+ "content": "<|begin_of_text|>",
337
+ "lstrip": false,
338
+ "normalized": false,
339
+ "rstrip": false,
340
+ "single_word": false
341
+ },
342
+ "eos_token": {
343
+ "content": "<|end_of_text|>",
344
+ "lstrip": false,
345
+ "normalized": false,
346
+ "rstrip": false,
347
+ "single_word": false
348
+ },
349
+ "pad_token": {
350
+ "content": "<pad>",
351
+ "lstrip": false,
352
+ "normalized": false,
353
+ "rstrip": false,
354
+ "single_word": false
355
+ }
356
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff