adefossez commited on
Commit
65dee9b
·
verified ·
1 Parent(s): d40c49d

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +40 -36
config.json CHANGED
@@ -1,39 +1,7 @@
1
  {
2
- "model_type": "hibiki",
3
- "lm_gen_config": {
4
- "temp": 0.8,
5
- "temp_text": 0.8,
6
- "top_k": 250,
7
- "top_k_text": 50
8
- },
9
- "dim": 2048,
10
- "text_card": 48000,
11
- "existing_text_padding_id": 3,
12
  "n_q": 16,
13
  "dep_q": 8,
14
- "card": 2048,
15
- "num_heads": 16,
16
- "num_layers": 16,
17
- "hidden_scale": 4.125,
18
- "causal": true,
19
- "layer_scale": null,
20
- "context": 500,
21
- "max_period": 100000,
22
- "gating": "silu",
23
- "norm": "rms_norm_f32",
24
- "positional_embedding": "rope",
25
- "depformer_dim": 1024,
26
- "depformer_dim_feedforward": 4224,
27
- "depformer_num_heads": 16,
28
- "depformer_num_layers": 6,
29
- "depformer_causal": true,
30
- "depformer_layer_scale": null,
31
- "depformer_multi_linear": true,
32
- "depformer_context": 16,
33
- "depformer_max_period": 10000,
34
- "depformer_gating": "silu",
35
- "depformer_pos_emb": "none",
36
- "depformer_weights_per_step": true,
37
  "delays": [
38
  0,
39
  0,
@@ -53,6 +21,26 @@
53
  2,
54
  2
55
  ],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
  "conditioners": {
57
  "description": {
58
  "type": "lut",
@@ -71,11 +59,27 @@
71
  }
72
  },
73
  "fuser": {
 
 
74
  "sum": [
75
  "description"
76
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  },
78
  "mimi_name": "mimi-pytorch-e351c8d8@125.safetensors",
79
- "moshi_name": "hibikim-pytorch-dc2cf5a5@80.safetensors",
80
- "tokenizer_name": "tokenizer_spm_48k_multi6_2.model"
81
  }
 
1
  {
2
+ "card": 2048,
 
 
 
 
 
 
 
 
 
3
  "n_q": 16,
4
  "dep_q": 8,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "delays": [
6
  0,
7
  0,
 
21
  2,
22
  2
23
  ],
24
+ "dim": 2048,
25
+ "text_card": 48000,
26
+ "existing_text_padding_id": 3,
27
+ "num_heads": 16,
28
+ "num_layers": 16,
29
+ "hidden_scale": 4.125,
30
+ "causal": true,
31
+ "layer_scale": null,
32
+ "context": 1500,
33
+ "max_period": 100000.0,
34
+ "gating": "silu",
35
+ "norm": "rms_norm_f32",
36
+ "positional_embedding": "rope",
37
+ "depformer_dim": 1024,
38
+ "depformer_num_heads": 16,
39
+ "depformer_num_layers": 6,
40
+ "depformer_dim_feedforward": null,
41
+ "depformer_multi_linear": true,
42
+ "depformer_pos_emb": "none",
43
+ "depformer_weights_per_step": true,
44
  "conditioners": {
45
  "description": {
46
  "type": "lut",
 
59
  }
60
  },
61
  "fuser": {
62
+ "cross_attention_pos_emb": false,
63
+ "cross_attention_pos_emb_scale": 1,
64
  "sum": [
65
  "description"
66
+ ],
67
+ "prepend": [],
68
+ "cross": []
69
+ },
70
+ "cross_attention": false,
71
+ "model_id": {
72
+ "sig": "37c6cfd6",
73
+ "epoch": 200
74
+ },
75
+ "model_type": "hibiki",
76
+ "lm_gen_config": {
77
+ "temp": 0.8,
78
+ "temp_text": 0.8,
79
+ "top_k": 250,
80
+ "top_k_text": 50
81
  },
82
  "mimi_name": "mimi-pytorch-e351c8d8@125.safetensors",
83
+ "tokenizer_name": "tokenizer_spm_48k_multi6_2.model",
84
+ "moshi_name": "hibikim-pytorch-37c6cfd6@200.safetensors"
85
  }