Spaces:
Sleeping
Sleeping
use selection
Browse files- app_local.py +6 -5
app_local.py
CHANGED
@@ -17,7 +17,7 @@ llm_model = None
|
|
17 |
def respond(
|
18 |
message,
|
19 |
history: list[tuple[str, str]],
|
20 |
-
|
21 |
system_message,
|
22 |
max_tokens,
|
23 |
temperature,
|
@@ -29,16 +29,17 @@ def respond(
|
|
29 |
|
30 |
global llm
|
31 |
global llm_model
|
32 |
-
|
33 |
-
|
|
|
34 |
llm = Llama(
|
35 |
-
model_path=f"models/{
|
36 |
flash_attn=True,
|
37 |
n_gpu_layers=81,
|
38 |
n_batch=1024,
|
39 |
n_ctx=8192,
|
40 |
)
|
41 |
-
llm_model =
|
42 |
|
43 |
provider = LlamaCppPythonProvider(llm)
|
44 |
|
|
|
17 |
def respond(
|
18 |
message,
|
19 |
history: list[tuple[str, str]],
|
20 |
+
selected_model,
|
21 |
system_message,
|
22 |
max_tokens,
|
23 |
temperature,
|
|
|
29 |
|
30 |
global llm
|
31 |
global llm_model
|
32 |
+
|
33 |
+
# Check if the selected model has changed
|
34 |
+
if llm is None or llm_model != selected_model:
|
35 |
llm = Llama(
|
36 |
+
model_path=f"models/{selected_model}",
|
37 |
flash_attn=True,
|
38 |
n_gpu_layers=81,
|
39 |
n_batch=1024,
|
40 |
n_ctx=8192,
|
41 |
)
|
42 |
+
llm_model = selected_model
|
43 |
|
44 |
provider = LlamaCppPythonProvider(llm)
|
45 |
|