boryasbora commited on
Commit
da56d07
·
verified ·
1 Parent(s): d6f9613

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -6
app.py CHANGED
@@ -16,7 +16,6 @@ from langchain_core.runnables import RunnableLambda
16
  from datetime import date
17
  from transformers import AutoModelForCausalLM, AutoTokenizer
18
  from setup import OLMO_MODEL
19
- from langchain_core.callbacks import CallbackManager
20
  model_path = OLMO_MODEL
21
 
22
  # Environment variables
@@ -84,13 +83,9 @@ def get_chain(temperature):
84
  retriever = load_retriever(docstore_path,chroma_path,embeddings,child_splitter,parent_splitter)
85
 
86
  # Replace the local OLMOLLM with the Hugging Face model
87
- callback_handler.on_llm_end = lambda response, *args, **kwargs: None
88
- # Callbacks support token-wise streaming
89
- callback_manager = CallbackManager([callback_handler])
90
  llm = LlamaCpp(
91
  model_path=str(model_path),
92
- callback_manager=callback_manager,
93
- temperature=0.8,
94
  max_tokens=3000,
95
  verbose=False,
96
  echo=False
 
16
  from datetime import date
17
  from transformers import AutoModelForCausalLM, AutoTokenizer
18
  from setup import OLMO_MODEL
 
19
  model_path = OLMO_MODEL
20
 
21
  # Environment variables
 
83
  retriever = load_retriever(docstore_path,chroma_path,embeddings,child_splitter,parent_splitter)
84
 
85
  # Replace the local OLMOLLM with the Hugging Face model
 
 
 
86
  llm = LlamaCpp(
87
  model_path=str(model_path),
88
+ temperature=temperature,
 
89
  max_tokens=3000,
90
  verbose=False,
91
  echo=False