Qingyun commited on
Commit
8d61c78
·
verified ·
1 Parent(s): c5b75d2

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -32,4 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.xz filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
32
  *.xz filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ pipeline_tag: image-text-to-text
4
+ library_name: transformers
5
+ base_model:
6
+ - lmsys/vicuna-13b-v1.5
7
+ base_model_relation: merge
8
+ tags:
9
+ - llava
10
+ - vision
11
+ - ocr
12
+ - custom_code
13
+ ---
14
+
15
+ This repository contains the PIIP-LLaVA_CLIP-BL_512-448_13B model, based on vicuna-13b-v1.5.
16
+
17
+ Please refer to our [**paper**](https://huggingface.co/papers/2501.07783) and [**GitHub repository**](https://github.com/OpenGVLab/PIIP/tree/main/llava) for introduction and usage.
18
+
19
+
20
+
21
+ ## Citation
22
+
23
+ If you find this project useful in your research, please consider citing:
24
+
25
+ ```BibTeX
26
+ @article{piip,
27
+ title={Parameter-Inverted Image Pyramid Networks},
28
+ author={Zhu, Xizhou and Yang, Xue and Wang, Zhaokai and Li, Hao and Dou, Wenhan and Ge, Junqi and Lu, Lewei and Qiao, Yu and Dai, Jifeng},
29
+ journal={arXiv preprint arXiv:2406.04330},
30
+ year={2024}
31
+ }
32
+
33
+ @article{piip_v2,
34
+ title={Parameter-Inverted Image Pyramid Networks for Visual Perception and Multimodal Understanding},
35
+ author={Wang, Zhaokai and Zhu, Xizhou and Yang, Xue and Luo, Gen and Li, Hao and Tian, Changyao and Dou, Wenhan and Ge, Junqi and Lu, Lewei and Qiao, Yu and Dai, Jifeng},
36
+ journal={arXiv preprint arXiv:2501.07783},
37
+ year={2025}
38
+ }
39
+ ```
40
+
config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "PIIP-LLaVA_CLIP-BL_512-448_13B",
3
+ "architectures": [
4
+ "LlavaLlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "freeze_mm_mlp_adapter": false,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 5120,
13
+ "image_aspect_ratio": "pad",
14
+ "image_feature_multiply": 1.0,
15
+ "image_grid_pinpoints": null,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 13824,
18
+ "max_length": 4096,
19
+ "max_position_embeddings": 4096,
20
+ "mm_hidden_size": 1024,
21
+ "mm_patch_merge_type": "flat",
22
+ "mm_projector_lr": null,
23
+ "mm_projector_type": "mlp2x_gelu",
24
+ "mm_use_im_patch_token": false,
25
+ "mm_use_im_start_end": false,
26
+ "mm_vision_select_feature": "patch",
27
+ "mm_vision_select_layer": -2,
28
+ "mm_vision_tower": "configs/piip-llava_clip-bl_512-448_13B.py",
29
+ "model_type": "llava_llama",
30
+ "num_attention_heads": 40,
31
+ "num_hidden_layers": 40,
32
+ "num_key_value_heads": 40,
33
+ "pad_token_id": 0,
34
+ "pretraining_tp": 1,
35
+ "rms_norm_eps": 1e-05,
36
+ "rope_scaling": null,
37
+ "rope_theta": 10000.0,
38
+ "tie_word_embeddings": false,
39
+ "tokenizer_model_max_length": 2048,
40
+ "tokenizer_padding_side": "right",
41
+ "torch_dtype": "bfloat16",
42
+ "transformers_version": "4.37.2",
43
+ "tune_mm_mlp_adapter": false,
44
+ "tune_vision_tower": true,
45
+ "tune_vit_pos_embedding": false,
46
+ "use_cache": true,
47
+ "use_mm_proj": true,
48
+ "vision_tower_name": "configs/piip-llava_clip-bl_512-448_13B.py",
49
+ "vocab_size": 32000
50
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "max_length": 4096,
7
+ "pad_token_id": 0,
8
+ "temperature": 0.9,
9
+ "top_p": 0.6,
10
+ "transformers_version": "4.37.2"
11
+ }
model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e71df53b9449a938603e98e6f57109b1566550e02e6315ba43de2bf279d299c0
3
+ size 4978265800
model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14a25ba1e6d232576cecb17781214346bc0a6910774a672623ea72b15847a9e6
3
+ size 4970422232
model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:137eb7c8484295d7bee319d72500d411a52b860b30bdcf0ab1c5842335d13f2a
3
+ size 4970422256
model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c9569c67465cfc74c58afe7dac0420f7e66a6a04a237e0c1bf641046a5a09ca
3
+ size 4933701504
model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c3be1034b57fa13ab89cc31c4820014b69602671cfe6c0b659b80a093798844
3
+ size 4933722216
model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd5d7923dcaaf67d71d520eddb6ed5c27adada95d3b5c9f75484038995fe20d3
3
+ size 2253043888
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<unk>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "bos_token": "<s>",
31
+ "clean_up_tokenization_spaces": false,
32
+ "eos_token": "</s>",
33
+ "legacy": false,
34
+ "model_max_length": 2048,
35
+ "pad_token": "<unk>",
36
+ "padding_side": "right",
37
+ "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
+ "tokenizer_class": "LlamaTokenizer",
40
+ "unk_token": "<unk>",
41
+ "use_default_system_prompt": false
42
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea53092687b7861426558fe261ff28c0850786acef5b45d44ef64818de64247b
3
+ size 6779