Spaces:
Sleeping
Sleeping
Commit
·
17cfcdd
1
Parent(s):
c6f1011
Fix: Load CPU-safe model for HF Space
Browse files
app.py
CHANGED
@@ -2,20 +2,20 @@ import streamlit as st
|
|
2 |
import torch
|
3 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
4 |
import re
|
|
|
5 |
from pydub import AudioSegment
|
6 |
import speech_recognition as sr
|
7 |
import io
|
8 |
|
9 |
# Load model and tokenizer from local fine-tuned directory
|
10 |
-
|
11 |
-
|
|
|
|
|
12 |
|
13 |
# Force CPU-safe model loading
|
14 |
-
|
15 |
-
|
16 |
-
torch_dtype=torch.float32,
|
17 |
-
low_cpu_mem_usage=True
|
18 |
-
).to("cpu")
|
19 |
|
20 |
|
21 |
# Dictionaries to decode user inputs
|
|
|
2 |
import torch
|
3 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
4 |
import re
|
5 |
+
from peft import PeftModel
|
6 |
from pydub import AudioSegment
|
7 |
import speech_recognition as sr
|
8 |
import io
|
9 |
|
10 |
# Load model and tokenizer from local fine-tuned directory
|
11 |
+
# Define base and adapter model paths
|
12 |
+
BASE_MODEL = "stanfordai/BioMedLM" # or the path you used originally
|
13 |
+
ADAPTER_PATH = "Tufan1/BioMedLM-Cardio-Fold4-CPU"
|
14 |
+
tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL)
|
15 |
|
16 |
# Force CPU-safe model loading
|
17 |
+
base_model = AutoModelForCausalLM.from_pretrained(BASE_MODEL, torch_dtype=torch.float32)
|
18 |
+
model = PeftModel.from_pretrained(base_model, ADAPTER_PATH, device_map=None).to("cpu")
|
|
|
|
|
|
|
19 |
|
20 |
|
21 |
# Dictionaries to decode user inputs
|