Can you provide a 5.5bpw quantization of this model?
#2
by
xldistance
- opened
I need to use a longer context, but 8bpw video memory is not enough
I need to use a longer context, but 8bpw video memory is not enough