How was this quantized?
#1
by jlinux - opened
Can you share how this was quantized? I am unable to quantize using convert.py from llama.cpp and successfully load it with BPE or SPM vocab. Your insights are appreciated :).
Closing.. llama.cpp has a pending merge request to support which successfully generates GGUF.
jlinux changed discussion status to closed