File size: 1,071 Bytes
5a22686
5bafdec
02cfa98
0d51f33
 
 
1049d8c
0d51f33
 
5910e3a
0d51f33
7415a22
0d51f33
5a22686
0d51f33
 
 
 
 
d4781e3
0d51f33
 
 
 
 
d4781e3
0d51f33
f99bc36
08f3dee
f99bc36
741ef63
5a22686
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
import gradio as gr
from transformers import AutoTokenizer,AutoModelForCausalLM
import torch
def haiku_generate():
    model_id = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
    tokenizer = AutoTokenizer.from_pretrained(model_id)
    model = AutoModelForCausalLM.from_pretrained(
        model_id,
        device_map="auto",
        torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
        low_cpu_mem_usage=True
    )
    model.eval()

    prompt ="俳句だけを生成して"
    inputs = tokenizer(prompt,return_tensors="pt")
    if not torch.cuda.is_available():
        inputs ={k:v.to("cpu") for k,v in inputs.items()}
    with torch.no_glad():
        outputs = model.generate(
            **inputs,
            max_new_tokens=25,
            do_sample=True,
            temperature=0.7,
            top_p=0.9
        )
    return tokenizer.decode(outputs[0],skip_special_tokens=True)
with gr.Blocks() as demo:
    output = gr.Textbox(label="俳句")
    btn = gr.Button("生成")
    btn.click(fn=haiku_generate,outputs=output)

demo.launch()