Add files using upload-large-folder tool
Browse files- README.md +3 -3
- model.safetensors +2 -2
- tokenizer_config.json +1 -0
README.md
CHANGED
@@ -1,16 +1,16 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
library_name: mlx
|
4 |
-
|
|
|
5 |
tags:
|
6 |
- mlx
|
7 |
-
pipeline_tag: text-generation
|
8 |
---
|
9 |
|
10 |
# mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit
|
11 |
|
12 |
This model [mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit](https://huggingface.co/mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit) was
|
13 |
-
converted to MLX format from [deepseek-ai/
|
14 |
using mlx-lm version **0.24.1**.
|
15 |
|
16 |
## Use with mlx
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
library_name: mlx
|
4 |
+
pipeline_tag: text-generation
|
5 |
+
base_model: deepseek-ai/DeepSeek-R1-0528-Qwen3-8B
|
6 |
tags:
|
7 |
- mlx
|
|
|
8 |
---
|
9 |
|
10 |
# mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit
|
11 |
|
12 |
This model [mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit](https://huggingface.co/mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit) was
|
13 |
+
converted to MLX format from [deepseek-ai/DeepSeek-R1-0528-Qwen3-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-0528-Qwen3-8B)
|
14 |
using mlx-lm version **0.24.1**.
|
15 |
|
16 |
## Use with mlx
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1810144c1beafe5c73ef740a787e26d94efdd846da4a5b60cd9e69a82083474
|
3 |
+
size 4607835164
|
tokenizer_config.json
CHANGED
@@ -229,6 +229,7 @@
|
|
229 |
}
|
230 |
},
|
231 |
"bos_token": "<|begin▁of▁sentence|>",
|
|
|
232 |
"clean_up_tokenization_spaces": false,
|
233 |
"eos_token": "<|end▁of▁sentence|>",
|
234 |
"extra_special_tokens": {},
|
|
|
229 |
}
|
230 |
},
|
231 |
"bos_token": "<|begin▁of▁sentence|>",
|
232 |
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='', is_first_sp=true, is_last_user=false) %}{%- for message in messages %}{%- if message['role'] == 'system' %}{%- if ns.is_first_sp %}{% set ns.system_prompt = ns.system_prompt + message['content'] %}{% set ns.is_first_sp = false %}{%- else %}{% set ns.system_prompt = ns.system_prompt + '\n\n' + message['content'] %}{%- endif %}{%- endif %}{%- endfor %}{{ bos_token }}{{ ns.system_prompt }}{%- for message in messages %}{% set content = message['content'] %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{%- set ns.is_first = false -%}{%- set ns.is_last_user = true -%}{{'<|User|>' + content + '<|Assistant|>'}}{%- endif %}{%- if message['role'] == 'assistant' %}{% if '</think>' in content %}{% set content = content.split('</think>')[-1] %}{% endif %}{% endif %}{%- if message['role'] == 'assistant' and message['tool_calls'] is defined and message['tool_calls'] is not none %}{%- set ns.is_last_user = false -%}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{%- endif %}{%- set ns.is_first = false %}{%- set ns.is_tool = false -%}{%- set ns.is_output_first = true %}{%- for tool in message['tool_calls'] %}{%- if not ns.is_first %}{%- if content is none %}{{'<|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\n' + '```json' + '\n' + tool['function']['arguments'] + '\n' + '```' + '<|tool▁call▁end|>'}}{%- else %}{{content + '<|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\n' + '```json' + '\n' + tool['function']['arguments'] + '\n' + '```' + '<|tool▁call▁end|>'}}{%- endif %}{%- set ns.is_first = true -%}{%- else %}{{'\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\n' + '```json' + '\n' + tool['function']['arguments'] + '\n' + '```' + '<|tool▁call▁end|>'}}{%- endif %}{%- endfor %}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- if message['role'] == 'assistant' and (message['tool_calls'] is not defined or message['tool_calls'] is none)%}{%- set ns.is_last_user = false -%}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + content + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{{content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_last_user = false -%}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + content + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'\n<|tool▁output▁begin|>' + content + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_last_user and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}",
|
233 |
"clean_up_tokenization_spaces": false,
|
234 |
"eos_token": "<|end▁of▁sentence|>",
|
235 |
"extra_special_tokens": {},
|