from transformers import pipeline | |
import spaces | |
# 1. ๋ชจ๋ธ ๋ก๋ (์ต์ด 1๋ฒ๋ง ๋ก๋๋จ) | |
generator = pipeline( | |
"text-generation", | |
model="dasomaru/gemma-3-4bit-it-demo", # ๋ค๊ฐ ์ ๋ก๋ํ ๋ชจ๋ธ ์ด๋ฆ | |
tokenizer="dasomaru/gemma-3-4bit-it-demo", | |
device=0, # CUDA:0 ์ฌ์ฉ (GPU). CPU๋ง ์์ผ๋ฉด device=-1 | |
max_new_tokens=512, | |
temperature=0.7, | |
top_p=0.9, | |
repetition_penalty=1.1 | |
) | |
# 2. ๋ต๋ณ ์์ฑ ํจ์ | |
def generate_answer(prompt: str) -> str: | |
""" | |
์ ๋ ฅ๋ฐ์ ํ๋กฌํํธ๋ก๋ถํฐ ๋ชจ๋ธ์ด ๋ต๋ณ์ ์์ฑํ๋ค. | |
""" | |
print(f"๐ต Prompt Length: {len(prompt)} characters") # ์ถ๊ฐ! | |
outputs = generator( | |
prompt, | |
do_sample=True, | |
top_k=50, | |
num_return_sequences=1 | |
) | |
return outputs[0]["generated_text"].strip() | |