gguf quantized and fp8/16/32 scaled dia-1.6b

  • base model from nari-labs
  • text-to-speech synthesis

run it with gguf-connector

ggc s2

screenshot

Prompt Audio Sample
[S1] This is an open weights text to dialogue model.
[S2] You get full control over scripts and voices.
[S1] Wow. Amazing. (laughs)
[S2] Try it now on Git hub or Hugging Face.
🎧 dia-sample-1
[S1] Hey Connector, why your appearance looks so stupid?
[S2] Oh, really? maybe I ate too much smart beans.
[S1] Wow. Amazing. (laughs)
[S2] Let's go to get some more smart beans and you will become stupid as well.
🎧 dia-sample-2

review/reference

  • simply execute the command (ggc s2) above in console/terminal
  • note: model file(s) will be pulled to local cache automatically during the first launch; then opt to run it entirely offline; i.e., from local URL: http://127.0.0.1:7860 with lazy webui
  • gguf-connector (pypi)
Downloads last month
667
GGUF
Model size
1.61B params
Architecture
pig
Hardware compatibility
Log In to view the estimation

2-bit

4-bit

6-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for calcuis/dia-gguf

Base model

nari-labs/Dia-1.6B
Quantized
(5)
this model