Spaces:
Running
Running
llama-bench : add support for the RPC backend (llama/7435)
Browse files
ggml.c
CHANGED
|
@@ -22872,6 +22872,14 @@ int ggml_cpu_has_sycl(void) {
|
|
| 22872 |
#endif
|
| 22873 |
}
|
| 22874 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22875 |
int ggml_cpu_has_gpublas(void) {
|
| 22876 |
return ggml_cpu_has_cuda() || ggml_cpu_has_clblast() || ggml_cpu_has_vulkan() || ggml_cpu_has_kompute() ||
|
| 22877 |
ggml_cpu_has_sycl();
|
|
|
|
| 22872 |
#endif
|
| 22873 |
}
|
| 22874 |
|
| 22875 |
+
int ggml_cpu_has_rpc(void) {
|
| 22876 |
+
#if defined(GGML_USE_RPC)
|
| 22877 |
+
return 1;
|
| 22878 |
+
#else
|
| 22879 |
+
return 0;
|
| 22880 |
+
#endif
|
| 22881 |
+
}
|
| 22882 |
+
|
| 22883 |
int ggml_cpu_has_gpublas(void) {
|
| 22884 |
return ggml_cpu_has_cuda() || ggml_cpu_has_clblast() || ggml_cpu_has_vulkan() || ggml_cpu_has_kompute() ||
|
| 22885 |
ggml_cpu_has_sycl();
|
ggml.h
CHANGED
|
@@ -2428,6 +2428,7 @@ extern "C" {
|
|
| 2428 |
GGML_API int ggml_cpu_has_sse3 (void);
|
| 2429 |
GGML_API int ggml_cpu_has_ssse3 (void);
|
| 2430 |
GGML_API int ggml_cpu_has_sycl (void);
|
|
|
|
| 2431 |
GGML_API int ggml_cpu_has_vsx (void);
|
| 2432 |
GGML_API int ggml_cpu_has_matmul_int8(void);
|
| 2433 |
|
|
|
|
| 2428 |
GGML_API int ggml_cpu_has_sse3 (void);
|
| 2429 |
GGML_API int ggml_cpu_has_ssse3 (void);
|
| 2430 |
GGML_API int ggml_cpu_has_sycl (void);
|
| 2431 |
+
GGML_API int ggml_cpu_has_rpc (void);
|
| 2432 |
GGML_API int ggml_cpu_has_vsx (void);
|
| 2433 |
GGML_API int ggml_cpu_has_matmul_int8(void);
|
| 2434 |
|