Coool2 commited on
Commit
3860a34
·
verified ·
1 Parent(s): 4dea40a

Update agent.py

Browse files
Files changed (1) hide show
  1. agent.py +9 -2
agent.py CHANGED
@@ -64,7 +64,7 @@ logging.basicConfig(level=logging.INFO)
64
  logging.getLogger("llama_index.core.agent").setLevel(logging.DEBUG)
65
  logging.getLogger("llama_index.llms").setLevel(logging.DEBUG)
66
 
67
- model_id = "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8"
68
  proj_llm = HuggingFaceLLM(
69
  model_name=model_id,
70
  tokenizer_name=model_id,
@@ -399,7 +399,14 @@ from llama_index.llms.huggingface import HuggingFaceLLM
399
  # --- 1. Initialize a dedicated LLM for Code Generation ---
400
  # It's good practice to use a model specifically fine-tuned for coding.
401
  # This model is loaded only once for efficiency.
402
- code_llm = proj_llm
 
 
 
 
 
 
 
403
 
404
  def generate_python_code(query: str) -> str:
405
  """
 
64
  logging.getLogger("llama_index.core.agent").setLevel(logging.DEBUG)
65
  logging.getLogger("llama_index.llms").setLevel(logging.DEBUG)
66
 
67
+ model_id = "Qwen/Qwen2.5-7B-Instruct"
68
  proj_llm = HuggingFaceLLM(
69
  model_name=model_id,
70
  tokenizer_name=model_id,
 
399
  # --- 1. Initialize a dedicated LLM for Code Generation ---
400
  # It's good practice to use a model specifically fine-tuned for coding.
401
  # This model is loaded only once for efficiency.
402
+ code_llm = HuggingFaceLLM(
403
+ model_name="Qwen/Qwen2.5-Coder-3B",
404
+ tokenizer_name="Qwen/Qwen2.5-Coder-3B",
405
+ device_map="auto",
406
+ model_kwargs={"torch_dtype": "auto"},
407
+ # Set generation parameters for precise, non-creative code output
408
+ generate_kwargs={"temperature": 0.0, "do_sample": False}
409
+ )
410
 
411
  def generate_python_code(query: str) -> str:
412
  """