Flashattention 2 support?
#14 opened about 1 month ago
by
t-albertge
attnmask
1
#13 opened about 1 month ago
by
Kamichanw
Question about the chat template which ignores add_generation_prompt
#12 opened 2 months ago
by
xukp20

How much VRAM/RAM is required to load this model?
1
#11 opened 5 months ago
by
dpkirchner
Training time
#10 opened 5 months ago
by
iHaag
4-bit LLaDA model
#9 opened 5 months ago
by
chentianqi
Impressive work
#8 opened 5 months ago
by
Daemontatox

what part of the code is diffusion?
π
1
1
#6 opened 5 months ago
by
fblgit

Model performance
2
#5 opened 5 months ago
by
icoicqico

That is awesome!
2
#4 opened 6 months ago
by
owao
Anybody has been able to run their chat.py model on a Mac?
8
#3 opened 6 months ago
by
neodymion
Gguf?
π
8
8
#2 opened 6 months ago
by
AlgorithmicKing

Add library_name and pipeline_tag to model card
π
1
#1 opened 6 months ago
by
nielsr
