Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -44,13 +44,13 @@ def generate_responses(prompt, history):
|
|
44 |
|
45 |
async def custom_sampler_task():
|
46 |
generated_list = []
|
47 |
-
generator = creative_sampler.generate(wrapped_prompt, max_length=
|
48 |
for token in generator:
|
49 |
generated_list.append(token)
|
50 |
return tokenizer.decode(generated_list, skip_special_tokens=True)
|
51 |
|
52 |
custom_output = asyncio.run(custom_sampler_task())
|
53 |
-
standard_output = model1.generate(inputs, max_length=
|
54 |
standard_response = tokenizer.decode(standard_output[0][len(inputs[0]):], skip_special_tokens=True)
|
55 |
|
56 |
return standard_response.strip(), custom_output.strip()
|
|
|
44 |
|
45 |
async def custom_sampler_task():
|
46 |
generated_list = []
|
47 |
+
generator = creative_sampler.generate(wrapped_prompt, max_length=1024, temperature=1)
|
48 |
for token in generator:
|
49 |
generated_list.append(token)
|
50 |
return tokenizer.decode(generated_list, skip_special_tokens=True)
|
51 |
|
52 |
custom_output = asyncio.run(custom_sampler_task())
|
53 |
+
standard_output = model1.generate(inputs, max_length=1024, temperature=1)
|
54 |
standard_response = tokenizer.decode(standard_output[0][len(inputs[0]):], skip_special_tokens=True)
|
55 |
|
56 |
return standard_response.strip(), custom_output.strip()
|