Can you provide a 5.5bpw quantization of this model?

#2
by xldistance - opened

I need to use a longer context, but 8bpw video memory is not enough

Sign up or log in to comment