balhafni commited on
Commit
211ef03
·
verified ·
1 Parent(s): 6ce29df

CODA model upload

Browse files
config.json ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/scratch/ba63/BERT_models/bert-base-arabertv02",
3
+ "architectures": [
4
+ "BertForTokenClassificationSingleLabel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "D*",
13
+ "1": "DK*",
14
+ "2": "D*KI_[\u0630\u0627]",
15
+ "3": "D*KI_[\u0646]K",
16
+ "4": "I_[ ]K*",
17
+ "5": "I_[ \u0627]K*",
18
+ "6": "I_[\u0627]K*",
19
+ "7": "I_[\u0627]KI_[\u0627]K*",
20
+ "8": "I_[\u0627]KI_[\u0644]K*",
21
+ "9": "I_[ \u0627\u0644]K*",
22
+ "10": "I_[\u0627\u0644]K*",
23
+ "11": "I_[ \u0644]K*",
24
+ "12": "I_[\u0647]K*",
25
+ "13": "I_[\u064a]K*",
26
+ "14": "K*",
27
+ "15": "K*D",
28
+ "16": "K*DK",
29
+ "17": "KDK*",
30
+ "18": "K*DKKK",
31
+ "19": "K*I_[ ]K",
32
+ "20": "KI_[ ]K*",
33
+ "21": "K*I_[\u0627]",
34
+ "22": "K*I_[\u0627]K",
35
+ "23": "KI_[\u0627 ]K*",
36
+ "24": "KI_[\u0627]K*",
37
+ "25": "K*I_[\u0627]KI_[\u0644]K",
38
+ "26": "KI_[\u0627\u0644]K*",
39
+ "27": "K*I_[\u0644]K",
40
+ "28": "KI_[\u0644]K*",
41
+ "29": "K*I_[\u0646]K",
42
+ "30": "KI_[\u0646]K*",
43
+ "31": "K*I_[\u0647]",
44
+ "32": "K*I_[\u0647]K",
45
+ "33": "K*I_[\u0648]K",
46
+ "34": "K*I_[\u064a]K",
47
+ "35": "KI_[\u064a]K*",
48
+ "36": "KKI_[ ]K*",
49
+ "37": "KKI_[ \u0627]K*",
50
+ "38": "KKI_[\u0627]K*",
51
+ "39": "KKKR_[\u0627]K*",
52
+ "40": "KKR_[\u0627]K*",
53
+ "41": "K*R_[ ]",
54
+ "42": "K*R_[\u0622]K",
55
+ "43": "K*R_[\u0627]",
56
+ "44": "KR_[\u0627]K*",
57
+ "45": "K*R_[\u0629]",
58
+ "46": "K*R_[\u062b]",
59
+ "47": "K*R_[\u062b]K",
60
+ "48": "K*R_[\u062b]KK",
61
+ "49": "K*R_[\u0630]",
62
+ "50": "K*R_[\u0638]",
63
+ "51": "K*R_[\u0639]",
64
+ "52": "K*R_[\u0642]",
65
+ "53": "K*R_[\u0642]K",
66
+ "54": "K*R_[\u0647]",
67
+ "55": "K*R_[\u0649]",
68
+ "56": "K*R_[\u064a]",
69
+ "57": "MI_[\u0627]K*",
70
+ "58": "MK*",
71
+ "59": "MK*R_[\u0629]",
72
+ "60": "MR_[\u0627]K*",
73
+ "61": "MR_[\u062b]K*",
74
+ "62": "R_[\u0622]K*",
75
+ "63": "R_[\u0627]",
76
+ "64": "R_[\u0627]K*",
77
+ "65": "R_[\u0627]K*I_[\u0644]K",
78
+ "66": "R_[\u0627]K*R_[\u062b]",
79
+ "67": "R_[\u0629]",
80
+ "68": "R_[\u062a]I_[\u0634]K*",
81
+ "69": "R_[\u062a]K*",
82
+ "70": "R_[\u062a]R_[\u0639]K*",
83
+ "71": "R_[\u062b]",
84
+ "72": "R_[\u062b]K*",
85
+ "73": "R_[\u062b]R_[\u0629]",
86
+ "74": "R_[\u0630]",
87
+ "75": "R_[\u0630]K*",
88
+ "76": "R_[\u0638]K*",
89
+ "77": "R_[\u063a]K*",
90
+ "78": "R_[\u0642]",
91
+ "79": "R_[\u0642]K*",
92
+ "80": "R_[\u0643]K*",
93
+ "81": "R_[\u0647]",
94
+ "82": "R_[\u0649]",
95
+ "83": "R_[\u064a]"
96
+ },
97
+ "initializer_range": 0.02,
98
+ "intermediate_size": 3072,
99
+ "label2id": {
100
+ "D*": 0,
101
+ "D*KI_[\u0630\u0627]": 2,
102
+ "D*KI_[\u0646]K": 3,
103
+ "DK*": 1,
104
+ "I_[ ]K*": 4,
105
+ "I_[ \u0627]K*": 5,
106
+ "I_[ \u0627\u0644]K*": 9,
107
+ "I_[ \u0644]K*": 11,
108
+ "I_[\u0627]K*": 6,
109
+ "I_[\u0627]KI_[\u0627]K*": 7,
110
+ "I_[\u0627]KI_[\u0644]K*": 8,
111
+ "I_[\u0627\u0644]K*": 10,
112
+ "I_[\u0647]K*": 12,
113
+ "I_[\u064a]K*": 13,
114
+ "K*": 14,
115
+ "K*D": 15,
116
+ "K*DK": 16,
117
+ "K*DKKK": 18,
118
+ "K*I_[ ]K": 19,
119
+ "K*I_[\u0627]": 21,
120
+ "K*I_[\u0627]K": 22,
121
+ "K*I_[\u0627]KI_[\u0644]K": 25,
122
+ "K*I_[\u0644]K": 27,
123
+ "K*I_[\u0646]K": 29,
124
+ "K*I_[\u0647]": 31,
125
+ "K*I_[\u0647]K": 32,
126
+ "K*I_[\u0648]K": 33,
127
+ "K*I_[\u064a]K": 34,
128
+ "K*R_[ ]": 41,
129
+ "K*R_[\u0622]K": 42,
130
+ "K*R_[\u0627]": 43,
131
+ "K*R_[\u0629]": 45,
132
+ "K*R_[\u062b]": 46,
133
+ "K*R_[\u062b]K": 47,
134
+ "K*R_[\u062b]KK": 48,
135
+ "K*R_[\u0630]": 49,
136
+ "K*R_[\u0638]": 50,
137
+ "K*R_[\u0639]": 51,
138
+ "K*R_[\u0642]": 52,
139
+ "K*R_[\u0642]K": 53,
140
+ "K*R_[\u0647]": 54,
141
+ "K*R_[\u0649]": 55,
142
+ "K*R_[\u064a]": 56,
143
+ "KDK*": 17,
144
+ "KI_[ ]K*": 20,
145
+ "KI_[\u0627 ]K*": 23,
146
+ "KI_[\u0627]K*": 24,
147
+ "KI_[\u0627\u0644]K*": 26,
148
+ "KI_[\u0644]K*": 28,
149
+ "KI_[\u0646]K*": 30,
150
+ "KI_[\u064a]K*": 35,
151
+ "KKI_[ ]K*": 36,
152
+ "KKI_[ \u0627]K*": 37,
153
+ "KKI_[\u0627]K*": 38,
154
+ "KKKR_[\u0627]K*": 39,
155
+ "KKR_[\u0627]K*": 40,
156
+ "KR_[\u0627]K*": 44,
157
+ "MI_[\u0627]K*": 57,
158
+ "MK*": 58,
159
+ "MK*R_[\u0629]": 59,
160
+ "MR_[\u0627]K*": 60,
161
+ "MR_[\u062b]K*": 61,
162
+ "R_[\u0622]K*": 62,
163
+ "R_[\u0627]": 63,
164
+ "R_[\u0627]K*": 64,
165
+ "R_[\u0627]K*I_[\u0644]K": 65,
166
+ "R_[\u0627]K*R_[\u062b]": 66,
167
+ "R_[\u0629]": 67,
168
+ "R_[\u062a]I_[\u0634]K*": 68,
169
+ "R_[\u062a]K*": 69,
170
+ "R_[\u062a]R_[\u0639]K*": 70,
171
+ "R_[\u062b]": 71,
172
+ "R_[\u062b]K*": 72,
173
+ "R_[\u062b]R_[\u0629]": 73,
174
+ "R_[\u0630]": 74,
175
+ "R_[\u0630]K*": 75,
176
+ "R_[\u0638]K*": 76,
177
+ "R_[\u063a]K*": 77,
178
+ "R_[\u0642]": 78,
179
+ "R_[\u0642]K*": 79,
180
+ "R_[\u0643]K*": 80,
181
+ "R_[\u0647]": 81,
182
+ "R_[\u0649]": 82,
183
+ "R_[\u064a]": 83
184
+ },
185
+ "layer_norm_eps": 1e-12,
186
+ "max_position_embeddings": 512,
187
+ "model_type": "bert",
188
+ "num_attention_heads": 12,
189
+ "num_hidden_layers": 12,
190
+ "pad_token_id": 0,
191
+ "position_embedding_type": "absolute",
192
+ "torch_dtype": "float32",
193
+ "transformers_version": "4.30.0",
194
+ "type_vocab_size": 2,
195
+ "use_cache": true,
196
+ "vocab_size": 64000
197
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feec43453230e66980d940571d564d6689400d4eeded05468d5194a51e10bbdf
3
+ size 538739825
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_basic_tokenize": true,
5
+ "do_lower_case": false,
6
+ "mask_token": "[MASK]",
7
+ "max_len": 512,
8
+ "model_max_length": 512,
9
+ "never_split": [
10
+ "[بريد]",
11
+ "[مستخدم]",
12
+ "[رابط]"
13
+ ],
14
+ "pad_token": "[PAD]",
15
+ "sep_token": "[SEP]",
16
+ "strip_accents": null,
17
+ "tokenize_chinese_chars": true,
18
+ "tokenizer_class": "BertTokenizer",
19
+ "unk_token": "[UNK]"
20
+ }
trainer_state.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 6.8493150684931505,
5
+ "global_step": 1500,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 2.28,
12
+ "learning_rate": 3.8584474885844754e-05,
13
+ "loss": 0.3009,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 4.57,
18
+ "learning_rate": 2.71689497716895e-05,
19
+ "loss": 0.0796,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 6.85,
24
+ "learning_rate": 1.5753424657534248e-05,
25
+ "loss": 0.0317,
26
+ "step": 1500
27
+ }
28
+ ],
29
+ "max_steps": 2190,
30
+ "num_train_epochs": 10,
31
+ "total_flos": 660892650498432.0,
32
+ "trial_name": null,
33
+ "trial_params": null
34
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c7c030beb1dfd05a088aa19f2c3f27f4b457f8f7979298c038a527af1e9753c
3
+ size 4015
vocab.txt ADDED
The diff for this file is too large to render. See raw diff