Upload folder using huggingface_hub
Browse files- .gitattributes +1 -1
- README.md +40 -0
- config.json +50 -0
- generation_config.json +11 -0
- model-00001-of-00006.safetensors +3 -0
- model-00002-of-00006.safetensors +3 -0
- model-00003-of-00006.safetensors +3 -0
- model-00004-of-00006.safetensors +3 -0
- model-00005-of-00006.safetensors +3 -0
- model-00006-of-00006.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +24 -0
- tokenizer.model +3 -0
- tokenizer_config.json +42 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
.gitattributes
CHANGED
@@ -32,4 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
32 |
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
pipeline_tag: image-text-to-text
|
4 |
+
library_name: transformers
|
5 |
+
base_model:
|
6 |
+
- lmsys/vicuna-13b-v1.5
|
7 |
+
base_model_relation: merge
|
8 |
+
tags:
|
9 |
+
- llava
|
10 |
+
- vision
|
11 |
+
- ocr
|
12 |
+
- custom_code
|
13 |
+
---
|
14 |
+
|
15 |
+
This repository contains the PIIP-LLaVA_CLIP-BL_512-448_13B model, based on vicuna-13b-v1.5.
|
16 |
+
|
17 |
+
Please refer to our [**paper**](https://huggingface.co/papers/2501.07783) and [**GitHub repository**](https://github.com/OpenGVLab/PIIP/tree/main/llava) for introduction and usage.
|
18 |
+
|
19 |
+
|
20 |
+
|
21 |
+
## Citation
|
22 |
+
|
23 |
+
If you find this project useful in your research, please consider citing:
|
24 |
+
|
25 |
+
```BibTeX
|
26 |
+
@article{piip,
|
27 |
+
title={Parameter-Inverted Image Pyramid Networks},
|
28 |
+
author={Zhu, Xizhou and Yang, Xue and Wang, Zhaokai and Li, Hao and Dou, Wenhan and Ge, Junqi and Lu, Lewei and Qiao, Yu and Dai, Jifeng},
|
29 |
+
journal={arXiv preprint arXiv:2406.04330},
|
30 |
+
year={2024}
|
31 |
+
}
|
32 |
+
|
33 |
+
@article{piip_v2,
|
34 |
+
title={Parameter-Inverted Image Pyramid Networks for Visual Perception and Multimodal Understanding},
|
35 |
+
author={Wang, Zhaokai and Zhu, Xizhou and Yang, Xue and Luo, Gen and Li, Hao and Tian, Changyao and Dou, Wenhan and Ge, Junqi and Lu, Lewei and Qiao, Yu and Dai, Jifeng},
|
36 |
+
journal={arXiv preprint arXiv:2501.07783},
|
37 |
+
year={2025}
|
38 |
+
}
|
39 |
+
```
|
40 |
+
|
config.json
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "PIIP-LLaVA_CLIP-BL_512-448_13B",
|
3 |
+
"architectures": [
|
4 |
+
"LlavaLlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 1,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"freeze_mm_mlp_adapter": false,
|
11 |
+
"hidden_act": "silu",
|
12 |
+
"hidden_size": 5120,
|
13 |
+
"image_aspect_ratio": "pad",
|
14 |
+
"image_feature_multiply": 1.0,
|
15 |
+
"image_grid_pinpoints": null,
|
16 |
+
"initializer_range": 0.02,
|
17 |
+
"intermediate_size": 13824,
|
18 |
+
"max_length": 4096,
|
19 |
+
"max_position_embeddings": 4096,
|
20 |
+
"mm_hidden_size": 1024,
|
21 |
+
"mm_patch_merge_type": "flat",
|
22 |
+
"mm_projector_lr": null,
|
23 |
+
"mm_projector_type": "mlp2x_gelu",
|
24 |
+
"mm_use_im_patch_token": false,
|
25 |
+
"mm_use_im_start_end": false,
|
26 |
+
"mm_vision_select_feature": "patch",
|
27 |
+
"mm_vision_select_layer": -2,
|
28 |
+
"mm_vision_tower": "configs/piip-llava_clip-bl_512-448_13B.py",
|
29 |
+
"model_type": "llava_llama",
|
30 |
+
"num_attention_heads": 40,
|
31 |
+
"num_hidden_layers": 40,
|
32 |
+
"num_key_value_heads": 40,
|
33 |
+
"pad_token_id": 0,
|
34 |
+
"pretraining_tp": 1,
|
35 |
+
"rms_norm_eps": 1e-05,
|
36 |
+
"rope_scaling": null,
|
37 |
+
"rope_theta": 10000.0,
|
38 |
+
"tie_word_embeddings": false,
|
39 |
+
"tokenizer_model_max_length": 2048,
|
40 |
+
"tokenizer_padding_side": "right",
|
41 |
+
"torch_dtype": "bfloat16",
|
42 |
+
"transformers_version": "4.37.2",
|
43 |
+
"tune_mm_mlp_adapter": false,
|
44 |
+
"tune_vision_tower": true,
|
45 |
+
"tune_vit_pos_embedding": false,
|
46 |
+
"use_cache": true,
|
47 |
+
"use_mm_proj": true,
|
48 |
+
"vision_tower_name": "configs/piip-llava_clip-bl_512-448_13B.py",
|
49 |
+
"vocab_size": 32000
|
50 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"do_sample": true,
|
5 |
+
"eos_token_id": 2,
|
6 |
+
"max_length": 4096,
|
7 |
+
"pad_token_id": 0,
|
8 |
+
"temperature": 0.9,
|
9 |
+
"top_p": 0.6,
|
10 |
+
"transformers_version": "4.37.2"
|
11 |
+
}
|
model-00001-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e71df53b9449a938603e98e6f57109b1566550e02e6315ba43de2bf279d299c0
|
3 |
+
size 4978265800
|
model-00002-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14a25ba1e6d232576cecb17781214346bc0a6910774a672623ea72b15847a9e6
|
3 |
+
size 4970422232
|
model-00003-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:137eb7c8484295d7bee319d72500d411a52b860b30bdcf0ab1c5842335d13f2a
|
3 |
+
size 4970422256
|
model-00004-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c9569c67465cfc74c58afe7dac0420f7e66a6a04a237e0c1bf641046a5a09ca
|
3 |
+
size 4933701504
|
model-00005-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c3be1034b57fa13ab89cc31c4820014b69602671cfe6c0b659b80a093798844
|
3 |
+
size 4933722216
|
model-00006-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cd5d7923dcaaf67d71d520eddb6ed5c27adada95d3b5c9f75484038995fe20d3
|
3 |
+
size 2253043888
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "<unk>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<unk>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
3 |
+
size 499723
|
tokenizer_config.json
ADDED
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"0": {
|
6 |
+
"content": "<unk>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"1": {
|
14 |
+
"content": "<s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"2": {
|
22 |
+
"content": "</s>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
}
|
29 |
+
},
|
30 |
+
"bos_token": "<s>",
|
31 |
+
"clean_up_tokenization_spaces": false,
|
32 |
+
"eos_token": "</s>",
|
33 |
+
"legacy": false,
|
34 |
+
"model_max_length": 2048,
|
35 |
+
"pad_token": "<unk>",
|
36 |
+
"padding_side": "right",
|
37 |
+
"sp_model_kwargs": {},
|
38 |
+
"spaces_between_special_tokens": false,
|
39 |
+
"tokenizer_class": "LlamaTokenizer",
|
40 |
+
"unk_token": "<unk>",
|
41 |
+
"use_default_system_prompt": false
|
42 |
+
}
|
trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ea53092687b7861426558fe261ff28c0850786acef5b45d44ef64818de64247b
|
3 |
+
size 6779
|