File tree Expand file tree Collapse file tree 1 file changed +0
-16
lines changed Expand file tree Collapse file tree 1 file changed +0
-16
lines changed Original file line number Diff line number Diff line change @@ -262,22 +262,6 @@ static bool fp16_mma_hardware_available(const int cc) {
262
262
(GGML_CUDA_CC_IS_MTHREADS (cc) && cc >= GGML_CUDA_CC_QY2);
263
263
}
264
264
265
- static bool bf16_mma_hardware_available (const int cc) {
266
- return (GGML_CUDA_CC_IS_NVIDIA (cc) && cc >= GGML_CUDA_CC_AMPERE) || GGML_CUDA_CC_IS_CDNA (cc) || cc >= GGML_CUDA_CC_RDNA3;
267
- }
268
-
269
- static bool fp32_mma_hardware_available (const int cc) {
270
- return GGML_CUDA_CC_IS_CDNA (cc);
271
- }
272
-
273
- static bool bf16_mma_hardware_available (const int cc) {
274
- return (GGML_CUDA_CC_IS_NVIDIA (cc) && cc >= GGML_CUDA_CC_AMPERE) || GGML_CUDA_CC_IS_CDNA (cc) || cc >= GGML_CUDA_CC_RDNA3;
275
- }
276
-
277
- static bool fp32_mma_hardware_available (const int cc) {
278
- return GGML_CUDA_CC_IS_CDNA (cc);
279
- }
280
-
281
265
// Volta technically had FP16 tensor cores but they work very differently compared to Turing and later.
282
266
static bool new_mma_available (const int cc) {
283
267
return GGML_CUDA_CC_IS_NVIDIA (cc) && ggml_cuda_highest_compiled_arch (cc) >= GGML_CUDA_CC_TURING;
You can’t perform that action at this time.
0 commit comments