You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
for (size_t i = 0; i < llama_max_devices(); ++i) {
1452
1450
if (i < split_arg.size()) {
1453
-
params.tensor_split[i] = std::stof(split_arg[i]);
1451
+
params.tensor_split[i] = std::stof(split_arg[i]);
1454
1452
} else {
1455
-
params.tensor_split[i] = 0.0f;
1453
+
params.tensor_split[i] = 0.0f;
1456
1454
}
1457
1455
}
1458
-
#ifndef GGML_USE_CUDA_SYCL_VULKAN
1459
-
fprintf(stderr, "warning: llama.cpp was compiled without CUDA/SYCL/Vulkan. Setting a tensor split has no effect.\n");
1460
-
#endif// GGML_USE_CUDA_SYCL_VULKAN
1456
+
if (!llama_supports_gpu_offload()) {
1457
+
fprintf(stderr, "warning: llama.cpp was compiled without support for GPU offload. Setting a tensor split has no effect.\n");
1458
+
}
1461
1459
}
1462
1460
));
1463
1461
add_opt(llama_arg(
1464
1462
{"-mg", "--main-gpu"}, "INDEX",
1465
1463
format("the GPU to use for the model (with split-mode = none), or for intermediate results and KV (with split-mode = row) (default: %d)", params.main_gpu),
1466
1464
[](gpt_params & params, int value) {
1467
1465
params.main_gpu = value;
1468
-
#ifndef GGML_USE_CUDA_SYCL_VULKAN
1469
-
fprintf(stderr, "warning: llama.cpp was compiled without CUDA/SYCL/Vulkan. Setting the main GPU has no effect.\n");
1470
-
#endif// GGML_USE_CUDA_SYCL_VULKAN
1466
+
if (!llama_supports_gpu_offload()) {
1467
+
fprintf(stderr, "warning: llama.cpp was compiled without support for GPU offload. Setting the main GPU has no effect.\n");
0 commit comments