balhafni commited on
Commit
dc08fc9
·
verified ·
1 Parent(s): 13ca5ca

qalb14 pnx pretrained model

Browse files
config.json ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/scratch/ba63/BERT_models/bert-base-arabertv02",
3
+ "architectures": [
4
+ "BertForTokenClassificationSingleLabel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "A_[\"]K*",
13
+ "1": "A_[-]K*",
14
+ "2": "I_[ . ]K*",
15
+ "3": "I_[ \u060c ]K*",
16
+ "4": "I_[ \u061f ]K*",
17
+ "5": "K*",
18
+ "6": "K*A_[!]",
19
+ "7": "K*A_[\" .]",
20
+ "8": "K*A_[\" \u060c]",
21
+ "9": "K*A_[\"]",
22
+ "10": "K*A_[(]",
23
+ "11": "K*A_[) .]",
24
+ "12": "K*A_[)]",
25
+ "13": "K*A_[-]",
26
+ "14": "K*A_[.]",
27
+ "15": "K*A_[: \"]",
28
+ "16": "K*A_[:]",
29
+ "17": "K*A_[\u060c]",
30
+ "18": "K*A_[\u061b]",
31
+ "19": "K*A_[\u061f]",
32
+ "20": "K*I_[ \u060c ]K",
33
+ "21": "R_[!]",
34
+ "22": "R_[\"]",
35
+ "23": "R_[(]",
36
+ "24": "R_[)]",
37
+ "25": "R_[,]",
38
+ "26": "R_[-]",
39
+ "27": "R_[.]",
40
+ "28": "R_[:]",
41
+ "29": "R_[\u060c]",
42
+ "30": "R_[\u061b]",
43
+ "31": "R_[\u061f]",
44
+ "32": "R_[\"]A_[.]",
45
+ "33": "R_[\"]A_[\u060c]",
46
+ "34": "R_[:]A_[\"]"
47
+ },
48
+ "initializer_range": 0.02,
49
+ "intermediate_size": 3072,
50
+ "label2id": {
51
+ "A_[\"]K*": 0,
52
+ "A_[-]K*": 1,
53
+ "I_[ . ]K*": 2,
54
+ "I_[ \u060c ]K*": 3,
55
+ "I_[ \u061f ]K*": 4,
56
+ "K*": 5,
57
+ "K*A_[!]": 6,
58
+ "K*A_[\" .]": 7,
59
+ "K*A_[\" \u060c]": 8,
60
+ "K*A_[\"]": 9,
61
+ "K*A_[(]": 10,
62
+ "K*A_[) .]": 11,
63
+ "K*A_[)]": 12,
64
+ "K*A_[-]": 13,
65
+ "K*A_[.]": 14,
66
+ "K*A_[: \"]": 15,
67
+ "K*A_[:]": 16,
68
+ "K*A_[\u060c]": 17,
69
+ "K*A_[\u061b]": 18,
70
+ "K*A_[\u061f]": 19,
71
+ "K*I_[ \u060c ]K": 20,
72
+ "R_[!]": 21,
73
+ "R_[\"]": 22,
74
+ "R_[\"]A_[.]": 32,
75
+ "R_[\"]A_[\u060c]": 33,
76
+ "R_[(]": 23,
77
+ "R_[)]": 24,
78
+ "R_[,]": 25,
79
+ "R_[-]": 26,
80
+ "R_[.]": 27,
81
+ "R_[:]": 28,
82
+ "R_[:]A_[\"]": 34,
83
+ "R_[\u060c]": 29,
84
+ "R_[\u061b]": 30,
85
+ "R_[\u061f]": 31
86
+ },
87
+ "layer_norm_eps": 1e-12,
88
+ "max_position_embeddings": 512,
89
+ "model_type": "bert",
90
+ "num_attention_heads": 12,
91
+ "num_hidden_layers": 12,
92
+ "pad_token_id": 0,
93
+ "position_embedding_type": "absolute",
94
+ "torch_dtype": "float32",
95
+ "transformers_version": "4.30.0",
96
+ "type_vocab_size": 2,
97
+ "use_cache": true,
98
+ "vocab_size": 64000
99
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d98d683f9ef1738c27f5031c99483d93e9f73c4116c158a97a678270c84fd262
3
+ size 538589105
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_basic_tokenize": true,
5
+ "do_lower_case": false,
6
+ "mask_token": "[MASK]",
7
+ "max_len": 512,
8
+ "model_max_length": 512,
9
+ "never_split": [
10
+ "[بريد]",
11
+ "[مستخدم]",
12
+ "[رابط]"
13
+ ],
14
+ "pad_token": "[PAD]",
15
+ "sep_token": "[SEP]",
16
+ "strip_accents": null,
17
+ "tokenize_chinese_chars": true,
18
+ "tokenizer_class": "BertTokenizer",
19
+ "unk_token": "[UNK]"
20
+ }
trainer_state.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.8237232289950577,
5
+ "global_step": 500,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.82,
12
+ "learning_rate": 4.5881383855024714e-05,
13
+ "loss": 0.1893,
14
+ "step": 500
15
+ }
16
+ ],
17
+ "max_steps": 6070,
18
+ "num_train_epochs": 10,
19
+ "total_flos": 725004451318080.0,
20
+ "trial_name": null,
21
+ "trial_params": null
22
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1b5d9df9dd904398d49aa31f53414ca0758a9b95775f15f7fb501e04192b5b9
3
+ size 4079
vocab.txt ADDED
The diff for this file is too large to render. See raw diff