Title says 8-bit while model card sentence says 4-bit

#1
by StanHatko - opened

The model title has Int8 in the name and it has the tag "8-bit precision". However, the first sentence of the model card says:

This is a GPTQ-quantized 4-bit version of huihui-ai/Llama-3.3-70B-Instruct-abliterated-finetuned.

If this is an 8-bit model, can you please fix the first sentence of the model card? Or the other way around if it's a 4-bit model?

Sorry for the confusion. This is copied from the 4-bit format, and it has been corrected. You can check the quantization
settings in the config.json file for more details.

Thank you for quickly responding and fixing this.

StanHatko changed discussion status to closed

Sign up or log in to comment