Spaces:
Running
Running
File size: 1,071 Bytes
5a22686 5bafdec 02cfa98 0d51f33 1049d8c 0d51f33 5910e3a 0d51f33 7415a22 0d51f33 5a22686 0d51f33 d4781e3 0d51f33 d4781e3 0d51f33 f99bc36 08f3dee f99bc36 741ef63 5a22686 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 |
import gradio as gr
from transformers import AutoTokenizer,AutoModelForCausalLM
import torch
def haiku_generate():
model_id = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
device_map="auto",
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
low_cpu_mem_usage=True
)
model.eval()
prompt ="俳句だけを生成して"
inputs = tokenizer(prompt,return_tensors="pt")
if not torch.cuda.is_available():
inputs ={k:v.to("cpu") for k,v in inputs.items()}
with torch.no_glad():
outputs = model.generate(
**inputs,
max_new_tokens=25,
do_sample=True,
temperature=0.7,
top_p=0.9
)
return tokenizer.decode(outputs[0],skip_special_tokens=True)
with gr.Blocks() as demo:
output = gr.Textbox(label="俳句")
btn = gr.Button("生成")
btn.click(fn=haiku_generate,outputs=output)
demo.launch()
|