Coool2 commited on
Commit
ad260e4
·
verified ·
1 Parent(s): 2560709

Update agent.py

Browse files
Files changed (1) hide show
  1. agent.py +18 -4
agent.py CHANGED
@@ -4,7 +4,6 @@ from llama_index.core import VectorStoreIndex, Document
4
  from llama_index.core.node_parser import SentenceWindowNodeParser, HierarchicalNodeParser
5
  from llama_index.core.postprocessor import SentenceTransformerRerank
6
  from llama_index.embeddings.huggingface import HuggingFaceEmbedding
7
- from llama_index.llms.huggingface_api import HuggingFaceInferenceAPI
8
  from llama_index.core.retrievers import VectorIndexRetriever
9
  from llama_index.core.query_engine import RetrieverQueryEngine
10
  from llama_index.readers.file import PDFReader, DocxReader, CSVReader, ImageReader
@@ -23,9 +22,24 @@ from llama_index.core.callbacks.base import CallbackManager
23
  from llama_index.core.callbacks.llama_debug import LlamaDebugHandler
24
  from llama_index.core import Settings
25
 
26
- proj_llm = OpenRouter(
27
- model="mistralai/mistral-small-3.1-24b-instruct:free",
28
- api_key=os.getenv("OPENROUTER_API_KEY"),
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  )
30
 
31
  embed_model = HuggingFaceEmbedding("BAAI/bge-small-en-v1.5")
 
4
  from llama_index.core.node_parser import SentenceWindowNodeParser, HierarchicalNodeParser
5
  from llama_index.core.postprocessor import SentenceTransformerRerank
6
  from llama_index.embeddings.huggingface import HuggingFaceEmbedding
 
7
  from llama_index.core.retrievers import VectorIndexRetriever
8
  from llama_index.core.query_engine import RetrieverQueryEngine
9
  from llama_index.readers.file import PDFReader, DocxReader, CSVReader, ImageReader
 
22
  from llama_index.core.callbacks.llama_debug import LlamaDebugHandler
23
  from llama_index.core import Settings
24
 
25
+ from transformers import AutoModelForCausalLM, AutoTokenizer
26
+ from llama_index.llms.huggingface import HuggingFaceLLM
27
+
28
+ model_id = "mistralai/Pixtral-12B-Base-2409" # or "mistralai/Mistral-7B-Instruct-v0.2"
29
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
30
+ model = AutoModelForCausalLM.from_pretrained(
31
+ model_id,
32
+ torch_dtype="auto", # or torch.float16 for FP16
33
+ device_map="auto" # will use all available GPUs
34
+ )
35
+
36
+ proj_llm = HuggingFaceLLM(
37
+ model=model,
38
+ tokenizer=tokenizer,
39
+ context_window=3900, # adjust as needed
40
+ max_new_tokens=512, # adjust as needed
41
+ device_map="auto", # ensures multi-GPU support
42
+ generate_kwargs={"temperature": 0.7, "top_p": 0.95}
43
  )
44
 
45
  embed_model = HuggingFaceEmbedding("BAAI/bge-small-en-v1.5")