Upload 13 files
Browse files- Modelfile +2 -1
- adapter_config.json +5 -5
- adapter_model.safetensors +1 -1
- codellama-13b-qml.gguf +1 -1
- optimizer.pt +1 -1
- rng_state.pth +1 -1
- scheduler.pt +1 -1
- trainer_state.json +0 -0
- training_args.bin +1 -1
Modelfile
CHANGED
@@ -6,7 +6,8 @@ ADAPTER ./codellama-13b-qml.gguf
|
|
6 |
# Parameters optimized for code generation
|
7 |
PARAMETER temperature 0
|
8 |
PARAMETER num_predict 500
|
9 |
-
PARAMETER top_p
|
|
|
10 |
PARAMETER stop "<SUF>"
|
11 |
PARAMETER stop "<PRE>"
|
12 |
PARAMETER stop "</PRE>"
|
|
|
6 |
# Parameters optimized for code generation
|
7 |
PARAMETER temperature 0
|
8 |
PARAMETER num_predict 500
|
9 |
+
PARAMETER top_p 1
|
10 |
+
PARAMETER repeat_penalty 1.05
|
11 |
PARAMETER stop "<SUF>"
|
12 |
PARAMETER stop "<PRE>"
|
13 |
PARAMETER stop "</PRE>"
|
adapter_config.json
CHANGED
@@ -23,14 +23,14 @@
|
|
23 |
"rank_pattern": {},
|
24 |
"revision": null,
|
25 |
"target_modules": [
|
26 |
-
"up_proj",
|
27 |
"q_proj",
|
28 |
-
"
|
29 |
"o_proj",
|
30 |
-
"lm_head",
|
31 |
"gate_proj",
|
32 |
-
"
|
33 |
-
"
|
|
|
|
|
34 |
],
|
35 |
"task_type": "CAUSAL_LM",
|
36 |
"use_dora": false,
|
|
|
23 |
"rank_pattern": {},
|
24 |
"revision": null,
|
25 |
"target_modules": [
|
|
|
26 |
"q_proj",
|
27 |
+
"k_proj",
|
28 |
"o_proj",
|
|
|
29 |
"gate_proj",
|
30 |
+
"up_proj",
|
31 |
+
"lm_head",
|
32 |
+
"down_proj",
|
33 |
+
"v_proj"
|
34 |
],
|
35 |
"task_type": "CAUSAL_LM",
|
36 |
"use_dora": false,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1010972856
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ad1248e461e65d17445ce9cd6591bf3819149b646d5f2ed728563fff94798bec
|
3 |
size 1010972856
|
codellama-13b-qml.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1010935552
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96575ea056a540c9cc2d59b4c899b897f8f3178e6aaa5a02369c9b7ed0db0392
|
3 |
size 1010935552
|
optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 507119330
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b0424b5bb0d3f607e5c86bba2463c2b1b0124c1c9999c68bac1325e9efa160b
|
3 |
size 507119330
|
rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4da548e2683b691713e1bce09b17a3f8b68b352599b0f67e2a1ad51faa3a5ca
|
3 |
size 14244
|
scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1e14f542310c644a4bc29da20a8f89a9ba2f396d841056f00df29c8bfcd1650
|
3 |
size 1064
|
trainer_state.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5816
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5306b140cf8a31bb768c7ee63c9a1a7528de47035d9438a03161dd8779eda42a
|
3 |
size 5816
|