Skip to content

Commit

Permalink
ggml : add and use ggml_cpu_has_llamafile() (ggerganov#8664)
Browse files Browse the repository at this point in the history
  • Loading branch information
ggerganov authored Jul 25, 2024
1 parent be6d7c0 commit eddcb52
Show file tree
Hide file tree
Showing 3 changed files with 10 additions and 5 deletions.
1 change: 1 addition & 0 deletions ggml/include/ggml.h
Original file line number Diff line number Diff line change
Expand Up @@ -2400,6 +2400,7 @@ extern "C" {
GGML_API int ggml_cpu_has_vsx (void);
GGML_API int ggml_cpu_has_matmul_int8(void);
GGML_API int ggml_cpu_has_cann (void);
GGML_API int ggml_cpu_has_llamafile (void);

//
// Internal types and functions exposed for tests and benchmarks
Expand Down
8 changes: 8 additions & 0 deletions ggml/src/ggml.c
Original file line number Diff line number Diff line change
Expand Up @@ -22005,6 +22005,14 @@ int ggml_cpu_has_cann(void) {
#endif
}

int ggml_cpu_has_llamafile(void) {
#if defined(GGML_USE_LLAMAFILE)
return 1;
#else
return 0;
#endif
}

int ggml_cpu_has_gpublas(void) {
return ggml_cpu_has_cuda() || ggml_cpu_has_vulkan() || ggml_cpu_has_kompute() || ggml_cpu_has_sycl();
}
Expand Down
6 changes: 1 addition & 5 deletions src/llama.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -19146,11 +19146,7 @@ const char * llama_print_system_info(void) {
s += "SSSE3 = " + std::to_string(ggml_cpu_has_ssse3()) + " | ";
s += "VSX = " + std::to_string(ggml_cpu_has_vsx()) + " | ";
s += "MATMUL_INT8 = " + std::to_string(ggml_cpu_has_matmul_int8()) + " | ";
#ifdef GGML_USE_LLAMAFILE
s += "LLAMAFILE = 1 | ";
#else
s += "LLAMAFILE = 0 | ";
#endif
s += "LLAMAFILE = " + std::to_string(ggml_cpu_has_llamafile()) + " | ";

return s.c_str();
}
Expand Down

0 comments on commit eddcb52

Please sign in to comment.