view reply In your bitsandbytes config, why are you decompressing the weights to torch.float32, when the native format of phi3 is torch.bfloat16? This seems like a waste of memory