From: Sigbjørn Skjæret Date: Sun, 15 Mar 2026 11:15:12 +0000 (+0100) Subject: convert : support contiguous method on lora tensors (#20489) X-Git-Tag: upstream/0.0.8611~254 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=89d0aec0420f45718bc7f960134e7237dd4c9712;p=pkg%2Fggml%2Fsources%2Fllama.cpp convert : support contiguous method on lora tensors (#20489) --- diff --git a/convert_lora_to_gguf.py b/convert_lora_to_gguf.py index b0adde8a8..871ce8242 100755 --- a/convert_lora_to_gguf.py +++ b/convert_lora_to_gguf.py @@ -128,6 +128,12 @@ class LoraTorchTensor: assert dim is None return self.shape + def contiguous(self) -> LoraTorchTensor: + return LoraTorchTensor( + self._lora_A.contiguous(), + self._lora_B.contiguous(), + ) + def reshape(self, *shape: int | tuple[int, ...]) -> LoraTorchTensor: if isinstance(shape[0], tuple): new_shape: tuple[int, ...] = shape[0]