warshanks commited on
Commit
e2b9b1a
·
1 Parent(s): 6f47c3f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -12,12 +12,12 @@ import spaces
12
  import torch
13
  from loguru import logger
14
  from PIL import Image
15
- from transformers import AutoProcessor, Gemma3ForConditionalGeneration, TextIteratorStreamer
16
 
17
  model_id = os.getenv("MODEL_ID", "google/medgemma-4b-it")
18
- processor = AutoProcessor.from_pretrained(model_id, padding_side="left")
19
- model = Gemma3ForConditionalGeneration.from_pretrained(
20
- model_id, device_map="auto", torch_dtype=torch.bfloat16, attn_implementation="eager"
21
  )
22
 
23
  MAX_NUM_IMAGES = int(os.getenv("MAX_NUM_IMAGES", "5"))
@@ -214,7 +214,7 @@ demo = gr.ChatInterface(
214
  textbox=gr.MultimodalTextbox(file_types=["image", ".mp4"], file_count="multiple", autofocus=True),
215
  multimodal=True,
216
  additional_inputs=[
217
- gr.Textbox(label="System Prompt", value=""),
218
  gr.Slider(label="Max New Tokens", minimum=100, maximum=8192, step=10, value=2048),
219
  ],
220
  stop_btn=False,
 
12
  import torch
13
  from loguru import logger
14
  from PIL import Image
15
+ from transformers import AutoProcessor, AutoModelForImageTextToText, TextIteratorStreamer
16
 
17
  model_id = os.getenv("MODEL_ID", "google/medgemma-4b-it")
18
+ processor = AutoProcessor.from_pretrained(model_id)
19
+ model = AutoModelForImageTextToText.from_pretrained(
20
+ model_id, device_map="auto", torch_dtype=torch.bfloat16
21
  )
22
 
23
  MAX_NUM_IMAGES = int(os.getenv("MAX_NUM_IMAGES", "5"))
 
214
  textbox=gr.MultimodalTextbox(file_types=["image", ".mp4"], file_count="multiple", autofocus=True),
215
  multimodal=True,
216
  additional_inputs=[
217
+ gr.Textbox(label="System Prompt", value="You are a helpful medical expert."),
218
  gr.Slider(label="Max New Tokens", minimum=100, maximum=8192, step=10, value=2048),
219
  ],
220
  stop_btn=False,