import torchaudio import torch from vui.inference import render from vui.model import Vui model = Vui.from_pretrained().cuda() model.decoder = torch.compile(model.decoder, fullgraph=True, mode="max-autotune") for i in range(10): waveform = render( model, """Hey, here is some random stuff, usually something quite long as the shorter the text the less likely the model can cope! So cool yeah makes sense, would you be able to help me with something? Sure what is it?""", ) print(waveform.shape) torchaudio.save("out.opus", waveform[0].cpu(), 22050)