hermeschen-ezcon commited on
Commit
72f1d3e
·
verified ·
1 Parent(s): 432f90b

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. config.json +2 -98
  3. tokenizer_config.json +1 -0
README.md CHANGED
@@ -9,7 +9,7 @@ tags:
9
  ---
10
 
11
  # EZCon/RolmOCR-4bit-mlx
12
- This model was converted to MLX format from [`reducto/RolmOCR`]() using mlx-vlm version **0.1.21**.
13
  Refer to the [original model card](https://huggingface.co/reducto/RolmOCR) for more details on the model.
14
  ## Use with mlx
15
 
 
9
  ---
10
 
11
  # EZCon/RolmOCR-4bit-mlx
12
+ This model was converted to MLX format from [`reducto/RolmOCR`]() using mlx-vlm version **0.1.26**.
13
  Refer to the [original model card](https://huggingface.co/reducto/RolmOCR) for more details on the model.
14
  ## Use with mlx
15
 
config.json CHANGED
@@ -79,109 +79,14 @@
79
  "suppress_tokens": null,
80
  "task_specific_params": null,
81
  "temperature": 1.0,
82
- "text_config": {
83
- "vocab_size": 152064,
84
- "max_position_embeddings": 128000,
85
- "hidden_size": 3584,
86
- "intermediate_size": 18944,
87
- "num_hidden_layers": 28,
88
- "num_attention_heads": 28,
89
- "use_sliding_window": false,
90
- "sliding_window": 32768,
91
- "max_window_layers": 28,
92
- "num_key_value_heads": 4,
93
- "hidden_act": "silu",
94
- "initializer_range": 0.02,
95
- "rms_norm_eps": 1e-06,
96
- "use_cache": false,
97
- "rope_theta": 1000000.0,
98
- "attention_dropout": 0.0,
99
- "rope_scaling": {
100
- "mrope_section": [
101
- 16,
102
- 24,
103
- 24
104
- ],
105
- "rope_type": "default",
106
- "type": "default"
107
- },
108
- "return_dict": true,
109
- "output_hidden_states": false,
110
- "output_attentions": false,
111
- "torchscript": false,
112
- "torch_dtype": "bfloat16",
113
- "use_bfloat16": false,
114
- "tf_legacy_loss": false,
115
- "pruned_heads": {},
116
- "tie_word_embeddings": false,
117
- "chunk_size_feed_forward": 0,
118
- "is_encoder_decoder": false,
119
- "is_decoder": false,
120
- "cross_attention_hidden_size": null,
121
- "add_cross_attention": false,
122
- "tie_encoder_decoder": false,
123
- "max_length": 20,
124
- "min_length": 0,
125
- "do_sample": false,
126
- "early_stopping": false,
127
- "num_beams": 1,
128
- "num_beam_groups": 1,
129
- "diversity_penalty": 0.0,
130
- "temperature": 1.0,
131
- "top_k": 50,
132
- "top_p": 1.0,
133
- "typical_p": 1.0,
134
- "repetition_penalty": 1.0,
135
- "length_penalty": 1.0,
136
- "no_repeat_ngram_size": 0,
137
- "encoder_no_repeat_ngram_size": 0,
138
- "bad_words_ids": null,
139
- "num_return_sequences": 1,
140
- "output_scores": false,
141
- "return_dict_in_generate": false,
142
- "forced_bos_token_id": null,
143
- "forced_eos_token_id": null,
144
- "remove_invalid_values": false,
145
- "exponential_decay_length_penalty": null,
146
- "suppress_tokens": null,
147
- "begin_suppress_tokens": null,
148
- "architectures": [
149
- "Qwen2_5_VLForConditionalGeneration"
150
- ],
151
- "finetuning_task": null,
152
- "id2label": {
153
- "0": "LABEL_0",
154
- "1": "LABEL_1"
155
- },
156
- "label2id": {
157
- "LABEL_0": 0,
158
- "LABEL_1": 1
159
- },
160
- "tokenizer_class": null,
161
- "prefix": null,
162
- "bos_token_id": 151643,
163
- "pad_token_id": null,
164
- "eos_token_id": 151645,
165
- "sep_token_id": null,
166
- "decoder_start_token_id": null,
167
- "task_specific_params": null,
168
- "problem_type": null,
169
- "_name_or_path": "",
170
- "_attn_implementation_autoset": false,
171
- "model_type": "qwen2_5_vl_text",
172
- "vision_end_token_id": 151653,
173
- "vision_start_token_id": 151652,
174
- "vision_token_id": 151654
175
- },
176
  "tf_legacy_loss": false,
177
  "tie_encoder_decoder": false,
178
  "tie_word_embeddings": false,
179
  "tokenizer_class": null,
180
  "top_k": 50,
181
  "top_p": 1.0,
182
- "torch_dtype": "bfloat16",
183
  "torchscript": false,
184
- "transformers_version": "4.52.0.dev0",
185
  "typical_p": 1.0,
186
  "use_bfloat16": false,
187
  "use_cache": false,
@@ -269,8 +174,7 @@
269
  23,
270
  31
271
  ],
272
- "out_hidden_size": 3584,
273
- "initializer_range": 0.02
274
  },
275
  "vision_end_token_id": 151653,
276
  "vision_start_token_id": 151652,
 
79
  "suppress_tokens": null,
80
  "task_specific_params": null,
81
  "temperature": 1.0,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
82
  "tf_legacy_loss": false,
83
  "tie_encoder_decoder": false,
84
  "tie_word_embeddings": false,
85
  "tokenizer_class": null,
86
  "top_k": 50,
87
  "top_p": 1.0,
 
88
  "torchscript": false,
89
+ "transformers_version": "4.51.3",
90
  "typical_p": 1.0,
91
  "use_bfloat16": false,
92
  "use_cache": false,
 
174
  23,
175
  31
176
  ],
177
+ "out_hidden_size": 3584
 
178
  },
179
  "vision_end_token_id": 151653,
180
  "vision_start_token_id": 151652,
tokenizer_config.json CHANGED
@@ -195,6 +195,7 @@
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
 
198
  "clean_up_tokenization_spaces": false,
199
  "eos_token": "<|im_end|>",
200
  "errors": "replace",
 
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
  "clean_up_tokenization_spaces": false,
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",