mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-01-26 17:20:01 +00:00
Remove llamafile remnants (#1179)
This commit is contained in:
@@ -112,7 +112,6 @@ option(GGML_ACCELERATE "ggml: enable Accelerate framework"
|
|||||||
option(GGML_BLAS "ggml: use BLAS" ${GGML_BLAS_DEFAULT})
|
option(GGML_BLAS "ggml: use BLAS" ${GGML_BLAS_DEFAULT})
|
||||||
set(GGML_BLAS_VENDOR ${GGML_BLAS_VENDOR_DEFAULT} CACHE STRING
|
set(GGML_BLAS_VENDOR ${GGML_BLAS_VENDOR_DEFAULT} CACHE STRING
|
||||||
"ggml: BLAS library vendor")
|
"ggml: BLAS library vendor")
|
||||||
option(GGML_LLAMAFILE "ggml: use LLAMAFILE" OFF)
|
|
||||||
option(GGML_IQK_MUL_MAT "ggml: use optimized iqk matrix multiplications" ON)
|
option(GGML_IQK_MUL_MAT "ggml: use optimized iqk matrix multiplications" ON)
|
||||||
|
|
||||||
option(GGML_CUDA "ggml: use CUDA" OFF)
|
option(GGML_CUDA "ggml: use CUDA" OFF)
|
||||||
|
|||||||
@@ -2999,7 +2999,6 @@ extern "C" {
|
|||||||
GGML_API int ggml_cpu_has_vsx (void);
|
GGML_API int ggml_cpu_has_vsx (void);
|
||||||
GGML_API int ggml_cpu_has_matmul_int8(void);
|
GGML_API int ggml_cpu_has_matmul_int8(void);
|
||||||
GGML_API int ggml_cpu_has_cann (void);
|
GGML_API int ggml_cpu_has_cann (void);
|
||||||
GGML_API int ggml_cpu_has_llamafile (void);
|
|
||||||
|
|
||||||
//
|
//
|
||||||
// Internal types and functions exposed for tests and benchmarks
|
// Internal types and functions exposed for tests and benchmarks
|
||||||
|
|||||||
@@ -299,15 +299,6 @@ if (GGML_IQK_MUL_MAT)
|
|||||||
endif()
|
endif()
|
||||||
endif()
|
endif()
|
||||||
|
|
||||||
if (GGML_LLAMAFILE)
|
|
||||||
message(STATUS "Using llamafile")
|
|
||||||
|
|
||||||
add_compile_definitions(GGML_USE_LLAMAFILE)
|
|
||||||
|
|
||||||
set(GGML_HEADERS_LLAMAFILE llamafile/sgemm.h)
|
|
||||||
set(GGML_SOURCES_LLAMAFILE llamafile/sgemm.cpp)
|
|
||||||
endif()
|
|
||||||
|
|
||||||
if (GGML_CUDA)
|
if (GGML_CUDA)
|
||||||
cmake_minimum_required(VERSION 3.18) # for CMAKE_CUDA_ARCHITECTURES
|
cmake_minimum_required(VERSION 3.18) # for CMAKE_CUDA_ARCHITECTURES
|
||||||
|
|
||||||
@@ -1534,7 +1525,6 @@ add_library(ggml
|
|||||||
${GGML_SOURCES_VULKAN} ${GGML_HEADERS_VULKAN}
|
${GGML_SOURCES_VULKAN} ${GGML_HEADERS_VULKAN}
|
||||||
${GGML_SOURCES_ROCM} ${GGML_HEADERS_ROCM}
|
${GGML_SOURCES_ROCM} ${GGML_HEADERS_ROCM}
|
||||||
${GGML_SOURCES_BLAS} ${GGML_HEADERS_BLAS}
|
${GGML_SOURCES_BLAS} ${GGML_HEADERS_BLAS}
|
||||||
${GGML_SOURCES_LLAMAFILE} ${GGML_HEADERS_LLAMAFILE}
|
|
||||||
${GGML_SOURCES_IQK_MM} ${GGML_HEADERS_IQK_MM}
|
${GGML_SOURCES_IQK_MM} ${GGML_HEADERS_IQK_MM}
|
||||||
${GGML_SOURCES_IQK} ${GGML_HEADERS_IQK}
|
${GGML_SOURCES_IQK} ${GGML_HEADERS_IQK}
|
||||||
${GGML_SOURCES_CANN} ${GGML_HEADERS_CANN}
|
${GGML_SOURCES_CANN} ${GGML_HEADERS_CANN}
|
||||||
|
|||||||
@@ -54,13 +54,6 @@
|
|||||||
#if defined(__ARM_FEATURE_SVE)
|
#if defined(__ARM_FEATURE_SVE)
|
||||||
int ggml_sve_cnt_b = 0;
|
int ggml_sve_cnt_b = 0;
|
||||||
#endif
|
#endif
|
||||||
#if defined(__ARM_FEATURE_SVE) || defined(__ARM_FEATURE_MATMUL_INT8)
|
|
||||||
#undef GGML_USE_LLAMAFILE
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifdef GGML_USE_LLAMAFILE
|
|
||||||
#include <llamafile/sgemm.h>
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#if defined(_MSC_VER)
|
#if defined(_MSC_VER)
|
||||||
// disable "possible loss of data" to avoid hundreds of casts
|
// disable "possible loss of data" to avoid hundreds of casts
|
||||||
@@ -28670,14 +28663,6 @@ int ggml_cpu_has_cann(void) {
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
int ggml_cpu_has_llamafile(void) {
|
|
||||||
#if defined(GGML_USE_LLAMAFILE)
|
|
||||||
return 1;
|
|
||||||
#else
|
|
||||||
return 0;
|
|
||||||
#endif
|
|
||||||
}
|
|
||||||
|
|
||||||
int ggml_cpu_has_gpublas(void) {
|
int ggml_cpu_has_gpublas(void) {
|
||||||
return ggml_cpu_has_cuda() || ggml_cpu_has_vulkan() || ggml_cpu_has_kompute() || ggml_cpu_has_sycl();
|
return ggml_cpu_has_cuda() || ggml_cpu_has_vulkan() || ggml_cpu_has_kompute() || ggml_cpu_has_sycl();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -7883,7 +7883,6 @@ const char * llama_print_system_info(void) {
|
|||||||
s += "SSSE3 = " + std::to_string(ggml_cpu_has_ssse3()) + " | ";
|
s += "SSSE3 = " + std::to_string(ggml_cpu_has_ssse3()) + " | ";
|
||||||
s += "VSX = " + std::to_string(ggml_cpu_has_vsx()) + " | ";
|
s += "VSX = " + std::to_string(ggml_cpu_has_vsx()) + " | ";
|
||||||
s += "MATMUL_INT8 = " + std::to_string(ggml_cpu_has_matmul_int8()) + " | ";
|
s += "MATMUL_INT8 = " + std::to_string(ggml_cpu_has_matmul_int8()) + " | ";
|
||||||
s += "LLAMAFILE = " + std::to_string(ggml_cpu_has_llamafile()) + " | ";
|
|
||||||
|
|
||||||
return s.c_str();
|
return s.c_str();
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user