JohannesGaessler commited on
Commit
f16510d
·
1 Parent(s): e7b20b1

CUDA: fix unused warning in mmq.cu (llama/7442)

Browse files
Files changed (1) hide show
  1. ggml-cuda/mmq.cu +10 -0
ggml-cuda/mmq.cu CHANGED
@@ -1220,6 +1220,7 @@ template <bool need_check> static __global__ void
1220
  load_tiles_q4_0<arch_config.y, arch_config.nwarps, need_check>, VDR_Q4_0_Q8_1_MMQ, vec_dot_q4_0_q8_1_mul_mat>
1221
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1222
  #else
 
1223
  GGML_UNUSED(vec_dot_q4_0_q8_1_mul_mat);
1224
  NO_DEVICE_CODE;
1225
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1244,6 +1245,7 @@ template <bool need_check> static __global__ void
1244
  load_tiles_q4_1<arch_config.y, arch_config.nwarps, need_check>, VDR_Q4_1_Q8_1_MMQ, vec_dot_q4_1_q8_1_mul_mat>
1245
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1246
  #else
 
1247
  GGML_UNUSED(vec_dot_q4_1_q8_1_mul_mat);
1248
  NO_DEVICE_CODE;
1249
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1266,6 +1268,7 @@ template <bool need_check> static __global__ void
1266
  load_tiles_q5_0<arch_config.y, arch_config.nwarps, need_check>, VDR_Q5_0_Q8_1_MMQ, vec_dot_q5_0_q8_1_mul_mat>
1267
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1268
  #else
 
1269
  GGML_UNUSED(vec_dot_q5_0_q8_1_mul_mat);
1270
  NO_DEVICE_CODE;
1271
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1288,6 +1291,7 @@ mul_mat_q5_1(
1288
  load_tiles_q5_1<arch_config.y, arch_config.nwarps, need_check>, VDR_Q5_1_Q8_1_MMQ, vec_dot_q5_1_q8_1_mul_mat>
1289
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1290
  #else
 
1291
  GGML_UNUSED(vec_dot_q5_1_q8_1_mul_mat);
1292
  NO_DEVICE_CODE;
1293
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1310,6 +1314,7 @@ template <bool need_check> static __global__ void
1310
  load_tiles_q8_0<arch_config.y, arch_config.nwarps, need_check>, VDR_Q8_0_Q8_1_MMQ, vec_dot_q8_0_q8_1_mul_mat>
1311
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1312
  #else
 
1313
  GGML_UNUSED(vec_dot_q8_0_q8_1_mul_mat);
1314
  NO_DEVICE_CODE;
1315
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1332,6 +1337,7 @@ mul_mat_q2_K(
1332
  load_tiles_q2_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q2_K_Q8_1_MMQ, vec_dot_q2_K_q8_1_mul_mat>
1333
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1334
  #else
 
1335
  GGML_UNUSED(vec_dot_q2_K_q8_1_mul_mat);
1336
  NO_DEVICE_CODE;
1337
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1356,6 +1362,7 @@ template <bool need_check> static __global__ void
1356
  load_tiles_q3_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q3_K_Q8_1_MMQ, vec_dot_q3_K_q8_1_mul_mat>
1357
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1358
  #else
 
1359
  GGML_UNUSED(vec_dot_q3_K_q8_1_mul_mat);
1360
  NO_DEVICE_CODE;
1361
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1380,6 +1387,7 @@ template <bool need_check> static __global__ void
1380
  load_tiles_q4_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q4_K_Q8_1_MMQ, vec_dot_q4_K_q8_1_mul_mat>
1381
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1382
  #else
 
1383
  GGML_UNUSED(vec_dot_q4_K_q8_1_mul_mat);
1384
  NO_DEVICE_CODE;
1385
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1402,6 +1410,7 @@ mul_mat_q5_K(
1402
  load_tiles_q5_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q5_K_Q8_1_MMQ, vec_dot_q5_K_q8_1_mul_mat>
1403
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1404
  #else
 
1405
  GGML_UNUSED(vec_dot_q5_K_q8_1_mul_mat);
1406
  NO_DEVICE_CODE;
1407
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
@@ -1426,6 +1435,7 @@ template <bool need_check> static __global__ void
1426
  load_tiles_q6_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q6_K_Q8_1_MMQ, vec_dot_q6_K_q8_1_mul_mat>
1427
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1428
  #else
 
1429
  GGML_UNUSED(vec_dot_q6_K_q8_1_mul_mat);
1430
  NO_DEVICE_CODE;
1431
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1220
  load_tiles_q4_0<arch_config.y, arch_config.nwarps, need_check>, VDR_Q4_0_Q8_1_MMQ, vec_dot_q4_0_q8_1_mul_mat>
1221
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1222
  #else
1223
+ GGML_UNUSED(get_arch_config_device);
1224
  GGML_UNUSED(vec_dot_q4_0_q8_1_mul_mat);
1225
  NO_DEVICE_CODE;
1226
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1245
  load_tiles_q4_1<arch_config.y, arch_config.nwarps, need_check>, VDR_Q4_1_Q8_1_MMQ, vec_dot_q4_1_q8_1_mul_mat>
1246
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1247
  #else
1248
+ GGML_UNUSED(get_arch_config_device);
1249
  GGML_UNUSED(vec_dot_q4_1_q8_1_mul_mat);
1250
  NO_DEVICE_CODE;
1251
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1268
  load_tiles_q5_0<arch_config.y, arch_config.nwarps, need_check>, VDR_Q5_0_Q8_1_MMQ, vec_dot_q5_0_q8_1_mul_mat>
1269
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1270
  #else
1271
+ GGML_UNUSED(get_arch_config_device);
1272
  GGML_UNUSED(vec_dot_q5_0_q8_1_mul_mat);
1273
  NO_DEVICE_CODE;
1274
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1291
  load_tiles_q5_1<arch_config.y, arch_config.nwarps, need_check>, VDR_Q5_1_Q8_1_MMQ, vec_dot_q5_1_q8_1_mul_mat>
1292
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1293
  #else
1294
+ GGML_UNUSED(get_arch_config_device);
1295
  GGML_UNUSED(vec_dot_q5_1_q8_1_mul_mat);
1296
  NO_DEVICE_CODE;
1297
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1314
  load_tiles_q8_0<arch_config.y, arch_config.nwarps, need_check>, VDR_Q8_0_Q8_1_MMQ, vec_dot_q8_0_q8_1_mul_mat>
1315
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1316
  #else
1317
+ GGML_UNUSED(get_arch_config_device);
1318
  GGML_UNUSED(vec_dot_q8_0_q8_1_mul_mat);
1319
  NO_DEVICE_CODE;
1320
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1337
  load_tiles_q2_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q2_K_Q8_1_MMQ, vec_dot_q2_K_q8_1_mul_mat>
1338
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1339
  #else
1340
+ GGML_UNUSED(get_arch_config_device);
1341
  GGML_UNUSED(vec_dot_q2_K_q8_1_mul_mat);
1342
  NO_DEVICE_CODE;
1343
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1362
  load_tiles_q3_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q3_K_Q8_1_MMQ, vec_dot_q3_K_q8_1_mul_mat>
1363
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1364
  #else
1365
+ GGML_UNUSED(get_arch_config_device);
1366
  GGML_UNUSED(vec_dot_q3_K_q8_1_mul_mat);
1367
  NO_DEVICE_CODE;
1368
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1387
  load_tiles_q4_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q4_K_Q8_1_MMQ, vec_dot_q4_K_q8_1_mul_mat>
1388
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1389
  #else
1390
+ GGML_UNUSED(get_arch_config_device);
1391
  GGML_UNUSED(vec_dot_q4_K_q8_1_mul_mat);
1392
  NO_DEVICE_CODE;
1393
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1410
  load_tiles_q5_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q5_K_Q8_1_MMQ, vec_dot_q5_K_q8_1_mul_mat>
1411
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1412
  #else
1413
+ GGML_UNUSED(get_arch_config_device);
1414
  GGML_UNUSED(vec_dot_q5_K_q8_1_mul_mat);
1415
  NO_DEVICE_CODE;
1416
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A
 
1435
  load_tiles_q6_K<arch_config.y, arch_config.nwarps, need_check>, VDR_Q6_K_Q8_1_MMQ, vec_dot_q6_K_q8_1_mul_mat>
1436
  (vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
1437
  #else
1438
+ GGML_UNUSED(get_arch_config_device);
1439
  GGML_UNUSED(vec_dot_q6_K_q8_1_mul_mat);
1440
  NO_DEVICE_CODE;
1441
  #endif // __CUDA_ARCH__ >= MIN_CC_DP4A