ginipick commited on
Commit
b94c22c
ยท
verified ยท
1 Parent(s): 191dec3

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +31 -8
app.py CHANGED
@@ -11,7 +11,11 @@ import spaces
11
  MODEL_ID = "naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B"
12
  MAX_NEW_TOKENS = 512
13
  CPU_THREAD_COUNT = 4 # ํ•„์š”์‹œ ์กฐ์ ˆ
14
- HF_TOKEN = os.getenv("HF_TOKEN") # Hugging Face ํ† ํฐ ํ™˜๊ฒฝ ๋ณ€์ˆ˜์—์„œ ๊ฐ€์ ธ์˜ค๊ธฐ
 
 
 
 
15
 
16
  # --- ์„ ํƒ ์‚ฌํ•ญ: CPU ์Šค๋ ˆ๋“œ ์„ค์ • ---
17
  # torch.set_num_threads(CPU_THREAD_COUNT)
@@ -39,18 +43,34 @@ try:
39
  device_map = "auto" if torch.cuda.is_available() else "cpu"
40
  dtype = torch.float16 if torch.cuda.is_available() else torch.float32
41
 
 
 
 
 
 
 
 
 
 
42
  tokenizer = AutoTokenizer.from_pretrained(
43
  MODEL_ID,
44
- trust_remote_code=True,
45
- token=HF_TOKEN # ํ† ํฐ ์ถ”๊ฐ€
46
  )
47
 
 
 
 
 
 
 
 
 
 
 
 
48
  model = AutoModelForCausalLM.from_pretrained(
49
  MODEL_ID,
50
- torch_dtype=dtype,
51
- device_map=device_map,
52
- trust_remote_code=True,
53
- token=HF_TOKEN # ํ† ํฐ ์ถ”๊ฐ€
54
  )
55
 
56
  model.eval()
@@ -233,10 +253,13 @@ examples = [
233
  ["์ œ์ฃผ๋„ ์—ฌํ–‰ ๊ณ„ํš์„ ์„ธ์šฐ๊ณ  ์žˆ๋Š”๋ฐ, 3๋ฐ• 4์ผ ์ถ”์ฒœ ์ฝ”์Šค ์ข€ ์•Œ๋ ค์ฃผ์„ธ์š”."],
234
  ]
235
 
 
 
 
236
  # ChatInterface๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ ์ž์ฒด Chatbot ์ปดํฌ๋„ŒํŠธ ๊ด€๋ฆฌ
237
  demo = gr.ChatInterface(
238
  fn=predict,
239
- title="๐Ÿค– HyperCLOVAX-SEED-Text-Instruct-0.5B",
240
  description=(
241
  f"**๋ชจ๋ธ:** {MODEL_ID}\n"
242
  ),
 
11
  MODEL_ID = "naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B"
12
  MAX_NEW_TOKENS = 512
13
  CPU_THREAD_COUNT = 4 # ํ•„์š”์‹œ ์กฐ์ ˆ
14
+
15
+ # Hugging Face ํ† ํฐ ์„ค์ • - ํ™˜๊ฒฝ ๋ณ€์ˆ˜์—์„œ ๊ฐ€์ ธ์˜ค๊ธฐ
16
+ HF_TOKEN = os.getenv("HF_TOKEN")
17
+ if not HF_TOKEN:
18
+ print("๊ฒฝ๊ณ : HF_TOKEN ํ™˜๊ฒฝ ๋ณ€์ˆ˜๊ฐ€ ์„ค์ •๋˜์ง€ ์•Š์•˜์Šต๋‹ˆ๋‹ค. ๋น„๊ณต๊ฐœ ๋ชจ๋ธ์— ์ ‘๊ทผํ•  ์ˆ˜ ์—†์„ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.")
19
 
20
  # --- ์„ ํƒ ์‚ฌํ•ญ: CPU ์Šค๋ ˆ๋“œ ์„ค์ • ---
21
  # torch.set_num_threads(CPU_THREAD_COUNT)
 
43
  device_map = "auto" if torch.cuda.is_available() else "cpu"
44
  dtype = torch.float16 if torch.cuda.is_available() else torch.float32
45
 
46
+ # ํ† ํฌ๋‚˜์ด์ € ๋กœ๋”ฉ
47
+ tokenizer_kwargs = {
48
+ "trust_remote_code": True
49
+ }
50
+
51
+ # HF_TOKEN์ด ์„ค์ •๋˜์–ด ์žˆ์œผ๋ฉด ์ถ”๊ฐ€
52
+ if HF_TOKEN:
53
+ tokenizer_kwargs["token"] = HF_TOKEN
54
+
55
  tokenizer = AutoTokenizer.from_pretrained(
56
  MODEL_ID,
57
+ **tokenizer_kwargs
 
58
  )
59
 
60
+ # ๋ชจ๋ธ ๋กœ๋”ฉ
61
+ model_kwargs = {
62
+ "torch_dtype": dtype,
63
+ "device_map": device_map,
64
+ "trust_remote_code": True
65
+ }
66
+
67
+ # HF_TOKEN์ด ์„ค์ •๋˜์–ด ์žˆ์œผ๋ฉด ์ถ”๊ฐ€
68
+ if HF_TOKEN:
69
+ model_kwargs["token"] = HF_TOKEN
70
+
71
  model = AutoModelForCausalLM.from_pretrained(
72
  MODEL_ID,
73
+ **model_kwargs
 
 
 
74
  )
75
 
76
  model.eval()
 
253
  ["์ œ์ฃผ๋„ ์—ฌํ–‰ ๊ณ„ํš์„ ์„ธ์šฐ๊ณ  ์žˆ๋Š”๋ฐ, 3๋ฐ• 4์ผ ์ถ”์ฒœ ์ฝ”์Šค ์ข€ ์•Œ๋ ค์ฃผ์„ธ์š”."],
254
  ]
255
 
256
+ # ๋ชจ๋ธ ์ด๋ฆ„์— ๋งž๊ฒŒ ํƒ€์ดํ‹€ ์กฐ์ •
257
+ title = "๐Ÿค– HyperCLOVAX-SEED-Vision-Instruct-3B"
258
+
259
  # ChatInterface๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ ์ž์ฒด Chatbot ์ปดํฌ๋„ŒํŠธ ๊ด€๋ฆฌ
260
  demo = gr.ChatInterface(
261
  fn=predict,
262
+ title=title,
263
  description=(
264
  f"**๋ชจ๋ธ:** {MODEL_ID}\n"
265
  ),