INT4 LoRA fantastic-tuning vs QLoRA: A user inquired about the dissimilarities between INT4 LoRA fine-tuning and QLoRA in terms of accuracy and speed. A different member explained that QLoRA with HQQ includes frozen quantized weights, does not use tinnygemm, and makes use of dequantizing alongside torch.matmul LLM inference inside of a font: