awni commited on
Commit
b9b5af4
·
verified ·
1 Parent(s): 3002913

Add files using upload-large-folder tool

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. model.safetensors +2 -2
  3. tokenizer_config.json +1 -0
README.md CHANGED
@@ -1,16 +1,16 @@
1
  ---
2
  license: mit
3
  library_name: mlx
4
- base_model: deepseek-ai/deepseek-r1-0528-Qwen3-8B
 
5
  tags:
6
  - mlx
7
- pipeline_tag: text-generation
8
  ---
9
 
10
  # mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit
11
 
12
  This model [mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit](https://huggingface.co/mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit) was
13
- converted to MLX format from [deepseek-ai/deepseek-r1-0528-Qwen3-8B](https://huggingface.co/deepseek-ai/deepseek-r1-0528-Qwen3-8B)
14
  using mlx-lm version **0.24.1**.
15
 
16
  ## Use with mlx
 
1
  ---
2
  license: mit
3
  library_name: mlx
4
+ pipeline_tag: text-generation
5
+ base_model: deepseek-ai/DeepSeek-R1-0528-Qwen3-8B
6
  tags:
7
  - mlx
 
8
  ---
9
 
10
  # mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit
11
 
12
  This model [mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit](https://huggingface.co/mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit) was
13
+ converted to MLX format from [deepseek-ai/DeepSeek-R1-0528-Qwen3-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-0528-Qwen3-8B)
14
  using mlx-lm version **0.24.1**.
15
 
16
  ## Use with mlx
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab1fd8637a3197dab44341c5b6b2617ea0695f79e91beb01d59e55c40791a070
3
- size 4607835174
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1810144c1beafe5c73ef740a787e26d94efdd846da4a5b60cd9e69a82083474
3
+ size 4607835164
tokenizer_config.json CHANGED
@@ -229,6 +229,7 @@
229
  }
230
  },
231
  "bos_token": "<|begin▁of▁sentence|>",
 
232
  "clean_up_tokenization_spaces": false,
233
  "eos_token": "<|end▁of▁sentence|>",
234
  "extra_special_tokens": {},
 
229
  }
230
  },
231
  "bos_token": "<|begin▁of▁sentence|>",
232
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='', is_first_sp=true, is_last_user=false) %}{%- for message in messages %}{%- if message['role'] == 'system' %}{%- if ns.is_first_sp %}{% set ns.system_prompt = ns.system_prompt + message['content'] %}{% set ns.is_first_sp = false %}{%- else %}{% set ns.system_prompt = ns.system_prompt + '\n\n' + message['content'] %}{%- endif %}{%- endif %}{%- endfor %}{{ bos_token }}{{ ns.system_prompt }}{%- for message in messages %}{% set content = message['content'] %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{%- set ns.is_first = false -%}{%- set ns.is_last_user = true -%}{{'<|User|>' + content + '<|Assistant|>'}}{%- endif %}{%- if message['role'] == 'assistant' %}{% if '</think>' in content %}{% set content = content.split('</think>')[-1] %}{% endif %}{% endif %}{%- if message['role'] == 'assistant' and message['tool_calls'] is defined and message['tool_calls'] is not none %}{%- set ns.is_last_user = false -%}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{%- endif %}{%- set ns.is_first = false %}{%- set ns.is_tool = false -%}{%- set ns.is_output_first = true %}{%- for tool in message['tool_calls'] %}{%- if not ns.is_first %}{%- if content is none %}{{'<|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\n' + '```json' + '\n' + tool['function']['arguments'] + '\n' + '```' + '<|tool▁call▁end|>'}}{%- else %}{{content + '<|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\n' + '```json' + '\n' + tool['function']['arguments'] + '\n' + '```' + '<|tool▁call▁end|>'}}{%- endif %}{%- set ns.is_first = true -%}{%- else %}{{'\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\n' + '```json' + '\n' + tool['function']['arguments'] + '\n' + '```' + '<|tool▁call▁end|>'}}{%- endif %}{%- endfor %}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- if message['role'] == 'assistant' and (message['tool_calls'] is not defined or message['tool_calls'] is none)%}{%- set ns.is_last_user = false -%}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + content + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{{content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_last_user = false -%}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + content + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'\n<|tool▁output▁begin|>' + content + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_last_user and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}",
233
  "clean_up_tokenization_spaces": false,
234
  "eos_token": "<|end▁of▁sentence|>",
235
  "extra_special_tokens": {},