Kikinoking commited on
Commit
5c451a4
·
verified ·
1 Parent(s): 5c1ea40

Upload Qwen3ForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +44 -24
  2. model.safetensors +2 -2
config.json CHANGED
@@ -5,31 +5,10 @@
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
8
- "config_groups": {
9
- "group_0": {
10
- "input_activations": {
11
- "num_bits": 8,
12
- "strategy": "tensor",
13
- "type": "int"
14
- },
15
- "targets": [
16
- "Linear"
17
- ],
18
- "weights": {
19
- "num_bits": 8,
20
- "strategy": "tensor",
21
- "type": "int"
22
- }
23
- }
24
- },
25
  "eos_token_id": 151643,
26
- "format": "naive-quantized",
27
  "head_dim": 128,
28
  "hidden_act": "silu",
29
  "hidden_size": 1024,
30
- "ignore": [
31
- "lm_head"
32
- ],
33
  "initializer_range": 0.02,
34
  "intermediate_size": 3072,
35
  "max_position_embeddings": 32768,
@@ -38,14 +17,55 @@
38
  "num_attention_heads": 16,
39
  "num_hidden_layers": 28,
40
  "num_key_value_heads": 8,
41
- "quant_method": "compressed-tensors",
42
- "quantization_status": "frozen",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  "rms_norm_eps": 1e-06,
44
  "rope_scaling": null,
45
  "rope_theta": 1000000,
46
  "sliding_window": null,
47
  "tie_word_embeddings": true,
48
- "torch_dtype": "float32",
49
  "transformers_version": "4.51.3",
50
  "use_cache": true,
51
  "use_sliding_window": false,
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "eos_token_id": 151643,
 
9
  "head_dim": 128,
10
  "hidden_act": "silu",
11
  "hidden_size": 1024,
 
 
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 3072,
14
  "max_position_embeddings": 32768,
 
17
  "num_attention_heads": 16,
18
  "num_hidden_layers": 28,
19
  "num_key_value_heads": 8,
20
+ "quantization_config": {
21
+ "config_groups": {
22
+ "group_0": {
23
+ "input_activations": {
24
+ "actorder": null,
25
+ "block_structure": null,
26
+ "dynamic": false,
27
+ "group_size": null,
28
+ "num_bits": 8,
29
+ "observer": "minmax",
30
+ "observer_kwargs": {},
31
+ "strategy": "tensor",
32
+ "symmetric": true,
33
+ "type": "int"
34
+ },
35
+ "output_activations": null,
36
+ "targets": [
37
+ "Linear"
38
+ ],
39
+ "weights": {
40
+ "actorder": null,
41
+ "block_structure": null,
42
+ "dynamic": false,
43
+ "group_size": null,
44
+ "num_bits": 8,
45
+ "observer": "minmax",
46
+ "observer_kwargs": {},
47
+ "strategy": "tensor",
48
+ "symmetric": true,
49
+ "type": "int"
50
+ }
51
+ }
52
+ },
53
+ "format": "naive-quantized",
54
+ "global_compression_ratio": null,
55
+ "ignore": [
56
+ "lm_head"
57
+ ],
58
+ "kv_cache_scheme": null,
59
+ "quant_method": "compressed-tensors",
60
+ "quantization_status": "compressed",
61
+ "sparsity_config": {}
62
+ },
63
  "rms_norm_eps": 1e-06,
64
  "rope_scaling": null,
65
  "rope_theta": 1000000,
66
  "sliding_window": null,
67
  "tie_word_embeddings": true,
68
+ "torch_dtype": "float16",
69
  "transformers_version": "4.51.3",
70
  "use_cache": true,
71
  "use_sliding_window": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dc66f94b88125434979932c9794cd8d5f946b32a03cc7b2c6faabcbfa8459ef
3
- size 2384234968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f989386eb1d54b4e8c90164597f5ed23e01318172950950e7c28bf5ccd072a4b
3
+ size 751776664