Skip to content

Commit ca9c4d2

Browse files
committed
kv-cache : use ggml_set_rows (llama/14285)
* kv-cache : use ggml_set_rows ggml-ci * graph : separate k and v indices ggml-ci * cont : remove redundant ifs ggml-ci * kv-cache : improve find_slot impl * kv-cache : bounds-check when accessing slot_info indices * kv-cache : add comments ggml-ci * ggml : add TODOs for adding GGML_OP_SET_ROWS support in the backends ggml-ci
1 parent ca2b806 commit ca9c4d2

File tree

4 files changed

+24
-0
lines changed

4 files changed

+24
-0
lines changed

ggml/src/ggml-cann/ggml-cann.cpp

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2086,6 +2086,12 @@ static bool ggml_backend_cann_supports_op(ggml_backend_dev_t dev,
20862086
return false;
20872087
}
20882088
} break;
2089+
case GGML_OP_SET_ROWS:
2090+
{
2091+
// TODO: add support
2092+
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
2093+
return false;
2094+
} break;
20892095
case GGML_OP_CPY: {
20902096
ggml_tensor *src = op->src[0];
20912097
if ((op->type != GGML_TYPE_F32 && op->type != GGML_TYPE_F16) ||

ggml/src/ggml-opencl/ggml-opencl.cpp

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2222,6 +2222,12 @@ static bool ggml_opencl_supports_op(ggml_backend_dev_t dev, const struct ggml_te
22222222
default:
22232223
return false;
22242224
}
2225+
case GGML_OP_SET_ROWS:
2226+
{
2227+
// TODO: add support
2228+
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
2229+
return false;
2230+
} break;
22252231
case GGML_OP_CPY:
22262232
case GGML_OP_DUP:
22272233
case GGML_OP_CONT:

ggml/src/ggml-sycl/ggml-sycl.cpp

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4285,6 +4285,12 @@ static bool ggml_backend_sycl_device_supports_op(ggml_backend_dev_t dev, const g
42854285
return false;
42864286
}
42874287
}
4288+
case GGML_OP_SET_ROWS:
4289+
{
4290+
// TODO: add support
4291+
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
4292+
return false;
4293+
} break;
42884294
case GGML_OP_CPY:
42894295
{
42904296
ggml_type src0_type = op->src[0]->type;

ggml/src/ggml-vulkan/ggml-vulkan.cpp

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10339,6 +10339,12 @@ static bool ggml_backend_vk_device_supports_op(ggml_backend_dev_t dev, const ggm
1033910339
return false;
1034010340
}
1034110341
} break;
10342+
case GGML_OP_SET_ROWS:
10343+
{
10344+
// TODO: add support
10345+
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
10346+
return false;
10347+
} break;
1034210348
case GGML_OP_CONT:
1034310349
case GGML_OP_CPY:
1034410350
case GGML_OP_DUP:

0 commit comments

Comments
 (0)