From: Georgi Gerganov Date: Mon, 24 Apr 2023 15:52:25 +0000 (+0300) Subject: ggml : sync llama.cpp (fix GCC 8 build, close #99) X-Git-Tag: upstream/0.0.1642~1518 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=32f22c0e77dcba5cf0cf837df01811e36cd92f1e;p=pkg%2Fggml%2Fsources%2Fggml ggml : sync llama.cpp (fix GCC 8 build, close #99) --- diff --git a/src/ggml.c b/src/ggml.c index f8f73af3..6e46c0e5 100644 --- a/src/ggml.c +++ b/src/ggml.c @@ -436,7 +436,7 @@ static const size_t CACHE_LINE_SIZE_F32 = CACHE_LINE_SIZE/sizeof(float); static inline __m128i bytes_from_nibbles_16(const uint8_t * rsi) { // Load 8 bytes from memory - __m128i tmp = _mm_loadu_si64( ( const __m128i* )rsi ); + __m128i tmp = _mm_loadl_epi64( ( const __m128i* )rsi ); // Expand bytes into uint16_t values __m128i bytes = _mm_cvtepu8_epi16( tmp );