davanstrien HF Staff commited on
Commit
afc3cae
·
verified ·
1 Parent(s): d9f025a

Upload folder using huggingface_hub (#3)

Browse files

- Upload folder using huggingface_hub (da2d36ecd42472c355d88cbc9d274b8a2d0f0fa8)
- Update README.md (e99398594b9dd54097e7312db24a907ce9fbcefb)

added_tokens.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
- "</CARD>": 49153,
3
- "</CARD_SUMMARY>": 49155,
4
- "<CARD>": 49152,
5
- "<CARD_SUMMARY>": 49154,
6
- "<DATASET_CARD>": 49156,
7
- "<MODEL_CARD>": 49157
 
8
  }
 
1
  {
2
+ "</CARD>": 49154,
3
+ "</CARD_SUMMARY>": 49156,
4
+ "<CARD>": 49153,
5
+ "<CARD_SUMMARY>": 49155,
6
+ "<DATASET_CARD>": 49157,
7
+ "<MODEL_CARD>": 49158,
8
+ "[PAD]": 49152
9
  }
chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {% for message in messages %}{% if message['role'] == 'user' %}<CARD>{{ message['content'] }}</CARD>{% elif message['role'] == 'assistant' %}<CARD_SUMMARY>{{ message['content'] }}</CARD_SUMMARY>{% endif %}{% endfor %}{% if not add_generation_prompt %}<|endoftext|>{% endif %}{% if add_generation_prompt %}<CARD_SUMMARY>{% endif %}
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "HuggingFaceTB/SmolLM2-360M",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -26,7 +25,7 @@
26
  "rope_theta": 100000,
27
  "tie_word_embeddings": true,
28
  "torch_dtype": "bfloat16",
29
- "transformers_version": "4.48.3",
30
  "use_cache": true,
31
- "vocab_size": 49158
32
  }
 
1
  {
 
2
  "architectures": [
3
  "LlamaForCausalLM"
4
  ],
 
25
  "rope_theta": 100000,
26
  "tie_word_embeddings": true,
27
  "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.52.3",
29
  "use_cache": true,
30
+ "vocab_size": 49159
31
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
- "transformers_version": "4.48.3"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
+ "transformers_version": "4.52.3"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:458b1458e6d5eb6c17779eff4a8f82e6df895ad0ce8fc06a8f04f0dc7a3d4377
3
- size 723686432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5ad1d90198dfce88e6ee1f90539332d12d087f3a59da4f6abdf436813ff622c
3
+ size 723688352
runs/May28_17-48-55_r-davanstrien-jupyterlab-svh3scip-91c05-sx22q/events.out.tfevents.1748447366.r-davanstrien-jupyterlab-svh3scip-91c05-sx22q.1156.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e473849f0b893c6dad386796492a653251310f5d85bb1a862d98be3b90c16e7
3
+ size 18148
special_tokens_map.json CHANGED
@@ -57,7 +57,13 @@
57
  "rstrip": false,
58
  "single_word": false
59
  },
60
- "pad_token": "<|endoftext|>",
 
 
 
 
 
 
61
  "unk_token": {
62
  "content": "<|endoftext|>",
63
  "lstrip": false,
 
57
  "rstrip": false,
58
  "single_word": false
59
  },
60
+ "pad_token": {
61
+ "content": "[PAD]",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false
66
+ },
67
  "unk_token": {
68
  "content": "<|endoftext|>",
69
  "lstrip": false,
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 8192,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
@@ -163,7 +158,7 @@
163
  },
164
  {
165
  "id": 49152,
166
- "content": "<CARD>",
167
  "single_word": false,
168
  "lstrip": false,
169
  "rstrip": false,
@@ -172,7 +167,7 @@
172
  },
173
  {
174
  "id": 49153,
175
- "content": "</CARD>",
176
  "single_word": false,
177
  "lstrip": false,
178
  "rstrip": false,
@@ -181,7 +176,7 @@
181
  },
182
  {
183
  "id": 49154,
184
- "content": "<CARD_SUMMARY>",
185
  "single_word": false,
186
  "lstrip": false,
187
  "rstrip": false,
@@ -190,7 +185,7 @@
190
  },
191
  {
192
  "id": 49155,
193
- "content": "</CARD_SUMMARY>",
194
  "single_word": false,
195
  "lstrip": false,
196
  "rstrip": false,
@@ -199,7 +194,7 @@
199
  },
200
  {
201
  "id": 49156,
202
- "content": "<DATASET_CARD>",
203
  "single_word": false,
204
  "lstrip": false,
205
  "rstrip": false,
@@ -208,6 +203,15 @@
208
  },
209
  {
210
  "id": 49157,
 
 
 
 
 
 
 
 
 
211
  "content": "<MODEL_CARD>",
212
  "single_word": false,
213
  "lstrip": false,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
158
  },
159
  {
160
  "id": 49152,
161
+ "content": "[PAD]",
162
  "single_word": false,
163
  "lstrip": false,
164
  "rstrip": false,
 
167
  },
168
  {
169
  "id": 49153,
170
+ "content": "<CARD>",
171
  "single_word": false,
172
  "lstrip": false,
173
  "rstrip": false,
 
176
  },
177
  {
178
  "id": 49154,
179
+ "content": "</CARD>",
180
  "single_word": false,
181
  "lstrip": false,
182
  "rstrip": false,
 
185
  },
186
  {
187
  "id": 49155,
188
+ "content": "<CARD_SUMMARY>",
189
  "single_word": false,
190
  "lstrip": false,
191
  "rstrip": false,
 
194
  },
195
  {
196
  "id": 49156,
197
+ "content": "</CARD_SUMMARY>",
198
  "single_word": false,
199
  "lstrip": false,
200
  "rstrip": false,
 
203
  },
204
  {
205
  "id": 49157,
206
+ "content": "<DATASET_CARD>",
207
+ "single_word": false,
208
+ "lstrip": false,
209
+ "rstrip": false,
210
+ "normalized": false,
211
+ "special": true
212
+ },
213
+ {
214
+ "id": 49158,
215
  "content": "<MODEL_CARD>",
216
  "single_word": false,
217
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -138,7 +138,7 @@
138
  "special": true
139
  },
140
  "49152": {
141
- "content": "<CARD>",
142
  "lstrip": false,
143
  "normalized": false,
144
  "rstrip": false,
@@ -146,7 +146,7 @@
146
  "special": true
147
  },
148
  "49153": {
149
- "content": "</CARD>",
150
  "lstrip": false,
151
  "normalized": false,
152
  "rstrip": false,
@@ -154,7 +154,7 @@
154
  "special": true
155
  },
156
  "49154": {
157
- "content": "<CARD_SUMMARY>",
158
  "lstrip": false,
159
  "normalized": false,
160
  "rstrip": false,
@@ -162,7 +162,7 @@
162
  "special": true
163
  },
164
  "49155": {
165
- "content": "</CARD_SUMMARY>",
166
  "lstrip": false,
167
  "normalized": false,
168
  "rstrip": false,
@@ -170,7 +170,7 @@
170
  "special": true
171
  },
172
  "49156": {
173
- "content": "<DATASET_CARD>",
174
  "lstrip": false,
175
  "normalized": false,
176
  "rstrip": false,
@@ -178,6 +178,14 @@
178
  "special": true
179
  },
180
  "49157": {
 
 
 
 
 
 
 
 
181
  "content": "<MODEL_CARD>",
182
  "lstrip": false,
183
  "normalized": false,
@@ -195,12 +203,11 @@
195
  "<MODEL_CARD>"
196
  ],
197
  "bos_token": "<|endoftext|>",
198
- "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}<CARD>{{ message['content'] }}</CARD>{% elif message['role'] == 'assistant' %}<CARD_SUMMARY>{{ message['content'] }}</CARD_SUMMARY>{% endif %}{% endfor %}{% if not add_generation_prompt %}<|endoftext|>{% endif %}{% if add_generation_prompt %}<CARD_SUMMARY>{% endif %}",
199
  "clean_up_tokenization_spaces": false,
200
  "eos_token": "<|endoftext|>",
201
  "extra_special_tokens": {},
202
  "model_max_length": 8192,
203
- "pad_token": "<|endoftext|>",
204
  "tokenizer_class": "GPT2Tokenizer",
205
  "unk_token": "<|endoftext|>",
206
  "vocab_size": 49152
 
138
  "special": true
139
  },
140
  "49152": {
141
+ "content": "[PAD]",
142
  "lstrip": false,
143
  "normalized": false,
144
  "rstrip": false,
 
146
  "special": true
147
  },
148
  "49153": {
149
+ "content": "<CARD>",
150
  "lstrip": false,
151
  "normalized": false,
152
  "rstrip": false,
 
154
  "special": true
155
  },
156
  "49154": {
157
+ "content": "</CARD>",
158
  "lstrip": false,
159
  "normalized": false,
160
  "rstrip": false,
 
162
  "special": true
163
  },
164
  "49155": {
165
+ "content": "<CARD_SUMMARY>",
166
  "lstrip": false,
167
  "normalized": false,
168
  "rstrip": false,
 
170
  "special": true
171
  },
172
  "49156": {
173
+ "content": "</CARD_SUMMARY>",
174
  "lstrip": false,
175
  "normalized": false,
176
  "rstrip": false,
 
178
  "special": true
179
  },
180
  "49157": {
181
+ "content": "<DATASET_CARD>",
182
+ "lstrip": false,
183
+ "normalized": false,
184
+ "rstrip": false,
185
+ "single_word": false,
186
+ "special": true
187
+ },
188
+ "49158": {
189
  "content": "<MODEL_CARD>",
190
  "lstrip": false,
191
  "normalized": false,
 
203
  "<MODEL_CARD>"
204
  ],
205
  "bos_token": "<|endoftext|>",
 
206
  "clean_up_tokenization_spaces": false,
207
  "eos_token": "<|endoftext|>",
208
  "extra_special_tokens": {},
209
  "model_max_length": 8192,
210
+ "pad_token": "[PAD]",
211
  "tokenizer_class": "GPT2Tokenizer",
212
  "unk_token": "<|endoftext|>",
213
  "vocab_size": 49152
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc065d067b0b20c9ee90c61037b138a771ed147cc184d5523c1e1e87c5446736
3
- size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75bc5c4c8f5f8eed5a885964a0e2fea3ca3737607c6f9890a35ef46b516d96d9
3
+ size 6097