Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -25,7 +25,7 @@ os.environ['LANGCHAIN_API_KEY'] = 'lsv2_pt_ce80aac3833643dd893527f566a06bf9_667d
|
|
25 |
|
26 |
@st.cache_resource
|
27 |
def load_model():
|
28 |
-
model_name = "
|
29 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
30 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
31 |
return model, tokenizer
|
@@ -48,7 +48,7 @@ def load_retriever(docstore_path,chroma_path,embeddings,child_splitter,parent_sp
|
|
48 |
docstore=store,
|
49 |
child_splitter=child_splitter,
|
50 |
parent_splitter=parent_splitter,
|
51 |
-
search_kwargs={"k":
|
52 |
)
|
53 |
return retriever
|
54 |
def inspect(state):
|
@@ -83,8 +83,8 @@ def get_chain(temperature):
|
|
83 |
"text-generation",
|
84 |
model=model,
|
85 |
tokenizer=tokenizer,
|
86 |
-
max_length=
|
87 |
-
max_new_tokens =
|
88 |
temperature=temperature,
|
89 |
top_p=0.95,
|
90 |
repetition_penalty=1.15
|
|
|
25 |
|
26 |
@st.cache_resource
|
27 |
def load_model():
|
28 |
+
model_name = "EleutherAI/gpt-neo-125M"
|
29 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
30 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
31 |
return model, tokenizer
|
|
|
48 |
docstore=store,
|
49 |
child_splitter=child_splitter,
|
50 |
parent_splitter=parent_splitter,
|
51 |
+
search_kwargs={"k": 2}
|
52 |
)
|
53 |
return retriever
|
54 |
def inspect(state):
|
|
|
83 |
"text-generation",
|
84 |
model=model,
|
85 |
tokenizer=tokenizer,
|
86 |
+
max_length=1800,
|
87 |
+
max_new_tokens = 200,
|
88 |
temperature=temperature,
|
89 |
top_p=0.95,
|
90 |
repetition_penalty=1.15
|