From 2a7cc09149bb26e6c73049025b707f2bd8932bab Mon Sep 17 00:00:00 2001 From: Kawrakow Date: Thu, 22 Jan 2026 13:20:23 +0200 Subject: [PATCH] Remove llamafile remnants (#1179) --- ggml/CMakeLists.txt | 1 - ggml/include/ggml.h | 1 - ggml/src/CMakeLists.txt | 10 ---------- ggml/src/ggml.c | 15 --------------- src/llama.cpp | 1 - 5 files changed, 28 deletions(-) diff --git a/ggml/CMakeLists.txt b/ggml/CMakeLists.txt index 095cdac9..c8987b2f 100644 --- a/ggml/CMakeLists.txt +++ b/ggml/CMakeLists.txt @@ -112,7 +112,6 @@ option(GGML_ACCELERATE "ggml: enable Accelerate framework" option(GGML_BLAS "ggml: use BLAS" ${GGML_BLAS_DEFAULT}) set(GGML_BLAS_VENDOR ${GGML_BLAS_VENDOR_DEFAULT} CACHE STRING "ggml: BLAS library vendor") -option(GGML_LLAMAFILE "ggml: use LLAMAFILE" OFF) option(GGML_IQK_MUL_MAT "ggml: use optimized iqk matrix multiplications" ON) option(GGML_CUDA "ggml: use CUDA" OFF) diff --git a/ggml/include/ggml.h b/ggml/include/ggml.h index 127a8ac6..fe7cb166 100644 --- a/ggml/include/ggml.h +++ b/ggml/include/ggml.h @@ -2999,7 +2999,6 @@ extern "C" { GGML_API int ggml_cpu_has_vsx (void); GGML_API int ggml_cpu_has_matmul_int8(void); GGML_API int ggml_cpu_has_cann (void); - GGML_API int ggml_cpu_has_llamafile (void); // // Internal types and functions exposed for tests and benchmarks diff --git a/ggml/src/CMakeLists.txt b/ggml/src/CMakeLists.txt index 818bafa2..f12f87fc 100644 --- a/ggml/src/CMakeLists.txt +++ b/ggml/src/CMakeLists.txt @@ -299,15 +299,6 @@ if (GGML_IQK_MUL_MAT) endif() endif() -if (GGML_LLAMAFILE) - message(STATUS "Using llamafile") - - add_compile_definitions(GGML_USE_LLAMAFILE) - - set(GGML_HEADERS_LLAMAFILE llamafile/sgemm.h) - set(GGML_SOURCES_LLAMAFILE llamafile/sgemm.cpp) -endif() - if (GGML_CUDA) cmake_minimum_required(VERSION 3.18) # for CMAKE_CUDA_ARCHITECTURES @@ -1534,7 +1525,6 @@ add_library(ggml ${GGML_SOURCES_VULKAN} ${GGML_HEADERS_VULKAN} ${GGML_SOURCES_ROCM} ${GGML_HEADERS_ROCM} ${GGML_SOURCES_BLAS} ${GGML_HEADERS_BLAS} - ${GGML_SOURCES_LLAMAFILE} ${GGML_HEADERS_LLAMAFILE} ${GGML_SOURCES_IQK_MM} ${GGML_HEADERS_IQK_MM} ${GGML_SOURCES_IQK} ${GGML_HEADERS_IQK} ${GGML_SOURCES_CANN} ${GGML_HEADERS_CANN} diff --git a/ggml/src/ggml.c b/ggml/src/ggml.c index c849e026..5633ac8a 100644 --- a/ggml/src/ggml.c +++ b/ggml/src/ggml.c @@ -54,13 +54,6 @@ #if defined(__ARM_FEATURE_SVE) int ggml_sve_cnt_b = 0; #endif -#if defined(__ARM_FEATURE_SVE) || defined(__ARM_FEATURE_MATMUL_INT8) -#undef GGML_USE_LLAMAFILE -#endif - -#ifdef GGML_USE_LLAMAFILE -#include -#endif #if defined(_MSC_VER) // disable "possible loss of data" to avoid hundreds of casts @@ -28670,14 +28663,6 @@ int ggml_cpu_has_cann(void) { #endif } -int ggml_cpu_has_llamafile(void) { -#if defined(GGML_USE_LLAMAFILE) - return 1; -#else - return 0; -#endif -} - int ggml_cpu_has_gpublas(void) { return ggml_cpu_has_cuda() || ggml_cpu_has_vulkan() || ggml_cpu_has_kompute() || ggml_cpu_has_sycl(); } diff --git a/src/llama.cpp b/src/llama.cpp index becc0626..54952e25 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -7883,7 +7883,6 @@ const char * llama_print_system_info(void) { s += "SSSE3 = " + std::to_string(ggml_cpu_has_ssse3()) + " | "; s += "VSX = " + std::to_string(ggml_cpu_has_vsx()) + " | "; s += "MATMUL_INT8 = " + std::to_string(ggml_cpu_has_matmul_int8()) + " | "; - s += "LLAMAFILE = " + std::to_string(ggml_cpu_has_llamafile()) + " | "; return s.c_str(); }