narainkumbari commited on
Commit
17cfcdd
·
1 Parent(s): c6f1011

Fix: Load CPU-safe model for HF Space

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -2,20 +2,20 @@ import streamlit as st
2
  import torch
3
  from transformers import AutoTokenizer, AutoModelForCausalLM
4
  import re
 
5
  from pydub import AudioSegment
6
  import speech_recognition as sr
7
  import io
8
 
9
  # Load model and tokenizer from local fine-tuned directory
10
- MODEL_PATH = "Tufan1/BioMedLM-Cardio-Fold4-CPU"
11
- tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
 
 
12
 
13
  # Force CPU-safe model loading
14
- model = AutoModelForCausalLM.from_pretrained(
15
- MODEL_PATH,
16
- torch_dtype=torch.float32,
17
- low_cpu_mem_usage=True
18
- ).to("cpu")
19
 
20
 
21
  # Dictionaries to decode user inputs
 
2
  import torch
3
  from transformers import AutoTokenizer, AutoModelForCausalLM
4
  import re
5
+ from peft import PeftModel
6
  from pydub import AudioSegment
7
  import speech_recognition as sr
8
  import io
9
 
10
  # Load model and tokenizer from local fine-tuned directory
11
+ # Define base and adapter model paths
12
+ BASE_MODEL = "stanfordai/BioMedLM" # or the path you used originally
13
+ ADAPTER_PATH = "Tufan1/BioMedLM-Cardio-Fold4-CPU"
14
+ tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL)
15
 
16
  # Force CPU-safe model loading
17
+ base_model = AutoModelForCausalLM.from_pretrained(BASE_MODEL, torch_dtype=torch.float32)
18
+ model = PeftModel.from_pretrained(base_model, ADAPTER_PATH, device_map=None).to("cpu")
 
 
 
19
 
20
 
21
  # Dictionaries to decode user inputs