Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -16,7 +16,6 @@ from langchain_core.runnables import RunnableLambda
|
|
16 |
from datetime import date
|
17 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
18 |
from setup import OLMO_MODEL
|
19 |
-
from langchain_core.callbacks import CallbackManager
|
20 |
model_path = OLMO_MODEL
|
21 |
|
22 |
# Environment variables
|
@@ -84,13 +83,9 @@ def get_chain(temperature):
|
|
84 |
retriever = load_retriever(docstore_path,chroma_path,embeddings,child_splitter,parent_splitter)
|
85 |
|
86 |
# Replace the local OLMOLLM with the Hugging Face model
|
87 |
-
callback_handler.on_llm_end = lambda response, *args, **kwargs: None
|
88 |
-
# Callbacks support token-wise streaming
|
89 |
-
callback_manager = CallbackManager([callback_handler])
|
90 |
llm = LlamaCpp(
|
91 |
model_path=str(model_path),
|
92 |
-
|
93 |
-
temperature=0.8,
|
94 |
max_tokens=3000,
|
95 |
verbose=False,
|
96 |
echo=False
|
|
|
16 |
from datetime import date
|
17 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
18 |
from setup import OLMO_MODEL
|
|
|
19 |
model_path = OLMO_MODEL
|
20 |
|
21 |
# Environment variables
|
|
|
83 |
retriever = load_retriever(docstore_path,chroma_path,embeddings,child_splitter,parent_splitter)
|
84 |
|
85 |
# Replace the local OLMOLLM with the Hugging Face model
|
|
|
|
|
|
|
86 |
llm = LlamaCpp(
|
87 |
model_path=str(model_path),
|
88 |
+
temperature=temperature,
|
|
|
89 |
max_tokens=3000,
|
90 |
verbose=False,
|
91 |
echo=False
|