rodrigomasini commited on
Commit
3df52c2
·
verified ·
1 Parent(s): 4a9558b

Update alternative.py

Browse files
Files changed (1) hide show
  1. alternative.py +12 -5
alternative.py CHANGED
@@ -28,13 +28,14 @@ import subprocess
28
  from typing import Tuple
29
 
30
  import gradio as gr
 
31
  from dotenv import load_dotenv
32
- from langchain_community.embeddings import HuggingFaceInferenceAPIEmbeddings
33
- from langchain_community.llms import HuggingFaceEndpoint
34
  from scrapegraphai.graphs import SmartScraperGraph
35
  from scrapegraphai.utils import prettify_exec_info
36
 
37
- from bs4 import BeautifulSoup as Soup
 
38
  from langchain_community.document_loaders import (AsyncHtmlLoader,
39
  NewsURLLoader, PubMedLoader,
40
  PlaywrightURLLoader,
@@ -49,12 +50,18 @@ from langchain_community.document_loaders import (AsyncHtmlLoader,
49
  # ------------------------------------------------------------------------------
50
 
51
  # Load environment variables
 
52
  HUGGINGFACEHUB_API_TOKEN = os.getenv('HUGGINGFACEHUB_API_TOKEN')
53
 
 
 
 
54
  # Initialize the model instances
55
- repo_id = "mistralai/Mistral-7B-Instruct-v0.2"
56
  llm_model_instance = HuggingFaceEndpoint(
57
- repo_id=repo_id, model_kwargs={"temperature": 0.5, "max_length": 8192, "token": HUGGINGFACEHUB_API_TOKEN}
 
 
 
58
  )
59
 
60
  embedder_model_instance = HuggingFaceInferenceAPIEmbeddings(
 
28
  from typing import Tuple
29
 
30
  import gradio as gr
31
+ from bs4 import BeautifulSoup as Soup
32
  from dotenv import load_dotenv
33
+
 
34
  from scrapegraphai.graphs import SmartScraperGraph
35
  from scrapegraphai.utils import prettify_exec_info
36
 
37
+ from langchain_community.embeddings import HuggingFaceInferenceAPIEmbeddings
38
+ from langchain_community.llms.huggingface_endpoint import HuggingFaceEndpoint
39
  from langchain_community.document_loaders import (AsyncHtmlLoader,
40
  NewsURLLoader, PubMedLoader,
41
  PlaywrightURLLoader,
 
50
  # ------------------------------------------------------------------------------
51
 
52
  # Load environment variables
53
+ load_dotenv()
54
  HUGGINGFACEHUB_API_TOKEN = os.getenv('HUGGINGFACEHUB_API_TOKEN')
55
 
56
+ # Foudational Model HF repo ID
57
+ REPO_ID = "mistralai/Mistral-7B-Instruct-v0.3"
58
+
59
  # Initialize the model instances
 
60
  llm_model_instance = HuggingFaceEndpoint(
61
+ repo_id=repo_id,
62
+ huggingfacehub_api_token = HUGGINGFACEHUB_API_TOKEN,
63
+ task = "conversational",
64
+ **model_kwargs={"temperature": 0.5, "max_length": 8192, "timeout": 6000},
65
  )
66
 
67
  embedder_model_instance = HuggingFaceInferenceAPIEmbeddings(