Shane commited on
Commit
6d70d33
·
verified ·
1 Parent(s): c4c268b

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: microsoft/Phi-3-mini-4k-instruct
3
  library_name: transformers
4
  model_name: phi3-mini-SAMbot
5
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for phi3-mini-SAMbot
13
 
14
- This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -34,10 +34,10 @@ This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.16.0
38
- - Transformers: 4.50.0
39
  - Pytorch: 2.6.0+cu124
40
- - Datasets: 3.5.0
41
  - Tokenizers: 0.21.1
42
 
43
  ## Citations
 
1
  ---
2
+ base_model: google/gemma-3-1b-pt
3
  library_name: transformers
4
  model_name: phi3-mini-SAMbot
5
  tags:
 
11
 
12
  # Model Card for phi3-mini-SAMbot
13
 
14
+ This model is a fine-tuned version of [google/gemma-3-1b-pt](https://huggingface.co/google/gemma-3-1b-pt).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.15.2
38
+ - Transformers: 4.50.0.dev0
39
  - Pytorch: 2.6.0+cu124
40
+ - Datasets: 3.3.2
41
  - Tokenizers: 0.21.1
42
 
43
  ## Citations
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "microsoft/Phi-3-mini-4k-instruct",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
@@ -23,10 +23,10 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "down_proj",
27
  "o_proj",
28
- "qkv_proj",
29
- "gate_up_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "google/gemma-3-1b-pt",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "o_proj",
27
+ "gate_up_proj",
28
+ "down_proj",
29
+ "qkv_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c37ef470254545070a6557781333d10e0a35469507b5c5a48a577411b20234d
3
- size 62948960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75e42c5c84fde95588a513f3dac097e7dc01a8d4bcd2f0e794db60b68fe36fc5
3
+ size 10663272
added_tokens.json CHANGED
@@ -1,13 +1,3 @@
1
  {
2
- "<|assistant|>": 32001,
3
- "<|endoftext|>": 32000,
4
- "<|end|>": 32007,
5
- "<|placeholder1|>": 32002,
6
- "<|placeholder2|>": 32003,
7
- "<|placeholder3|>": 32004,
8
- "<|placeholder4|>": 32005,
9
- "<|placeholder5|>": 32008,
10
- "<|placeholder6|>": 32009,
11
- "<|system|>": 32006,
12
- "<|user|>": 32010
13
  }
 
1
  {
2
+ "<image_soft_token>": 262144
 
 
 
 
 
 
 
 
 
 
3
  }
special_tokens_map.json CHANGED
@@ -1,20 +1,23 @@
1
  {
 
2
  "bos_token": {
3
- "content": "<s>",
4
  "lstrip": false,
5
  "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
 
9
  "eos_token": {
10
- "content": "<|endoftext|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
16
  "pad_token": {
17
- "content": "<|endoftext|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
1
  {
2
+ "boi_token": "<start_of_image>",
3
  "bos_token": {
4
+ "content": "<bos>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false
9
  },
10
+ "eoi_token": "<end_of_image>",
11
  "eos_token": {
12
+ "content": "<eos>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false
17
  },
18
+ "image_token": "<image_soft_token>",
19
  "pad_token": {
20
+ "content": "<pad>",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1299c11d7cf632ef3b4e11937501358ada021bbdf7c47638d13c0ee982f2e79c
3
+ size 4689074
tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6064f9beb06153ff16b06b7e93115092fbd119e1b8e789e20a80cfb89e8fff4
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f596bfdfb6093cb53d1a5deffcc6063f36e1172b2de027ca0b706e06b77fc92f
3
  size 5560