What is the minimum vLLM version required to serve this model ?
👍
1
1
#6 opened 2 months ago
by
thameem-abbas
Is there a benefit of this version vs the original MXFP4?
#5 opened 3 months ago
by
SuperbEmphasis
Model does not generate tokens when served with 4 RTX 6000 ADA GPUs on vLLM
#4 opened 3 months ago
by
Esj-DL
gptq int4 MIX int 8 please please please champs!
#3 opened 3 months ago
by
groxaxo
seems stuck on last steps
#1 opened 4 months ago
by
Fernanda24