mirror of
https://github.com/ggerganov/whisper.cpp.git
synced 2025-08-17 18:22:01 +02:00
kv-cache : use ggml_set_rows (llama/14285)
* kv-cache : use ggml_set_rows ggml-ci * graph : separate k and v indices ggml-ci * cont : remove redundant ifs ggml-ci * kv-cache : improve find_slot impl * kv-cache : bounds-check when accessing slot_info indices * kv-cache : add comments ggml-ci * ggml : add TODOs for adding GGML_OP_SET_ROWS support in the backends ggml-ci
This commit is contained in:
@ -2086,6 +2086,12 @@ static bool ggml_backend_cann_supports_op(ggml_backend_dev_t dev,
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
} break;
|
} break;
|
||||||
|
case GGML_OP_SET_ROWS:
|
||||||
|
{
|
||||||
|
// TODO: add support
|
||||||
|
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
|
||||||
|
return false;
|
||||||
|
} break;
|
||||||
case GGML_OP_CPY: {
|
case GGML_OP_CPY: {
|
||||||
ggml_tensor *src = op->src[0];
|
ggml_tensor *src = op->src[0];
|
||||||
if ((op->type != GGML_TYPE_F32 && op->type != GGML_TYPE_F16) ||
|
if ((op->type != GGML_TYPE_F32 && op->type != GGML_TYPE_F16) ||
|
||||||
|
@ -2222,6 +2222,12 @@ static bool ggml_opencl_supports_op(ggml_backend_dev_t dev, const struct ggml_te
|
|||||||
default:
|
default:
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
case GGML_OP_SET_ROWS:
|
||||||
|
{
|
||||||
|
// TODO: add support
|
||||||
|
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
|
||||||
|
return false;
|
||||||
|
} break;
|
||||||
case GGML_OP_CPY:
|
case GGML_OP_CPY:
|
||||||
case GGML_OP_DUP:
|
case GGML_OP_DUP:
|
||||||
case GGML_OP_CONT:
|
case GGML_OP_CONT:
|
||||||
|
@ -4285,6 +4285,12 @@ static bool ggml_backend_sycl_device_supports_op(ggml_backend_dev_t dev, const g
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
case GGML_OP_SET_ROWS:
|
||||||
|
{
|
||||||
|
// TODO: add support
|
||||||
|
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
|
||||||
|
return false;
|
||||||
|
} break;
|
||||||
case GGML_OP_CPY:
|
case GGML_OP_CPY:
|
||||||
{
|
{
|
||||||
ggml_type src0_type = op->src[0]->type;
|
ggml_type src0_type = op->src[0]->type;
|
||||||
|
@ -10339,6 +10339,12 @@ static bool ggml_backend_vk_device_supports_op(ggml_backend_dev_t dev, const ggm
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
} break;
|
} break;
|
||||||
|
case GGML_OP_SET_ROWS:
|
||||||
|
{
|
||||||
|
// TODO: add support
|
||||||
|
// ref: https://github.com/ggml-org/llama.cpp/pull/14274
|
||||||
|
return false;
|
||||||
|
} break;
|
||||||
case GGML_OP_CONT:
|
case GGML_OP_CONT:
|
||||||
case GGML_OP_CPY:
|
case GGML_OP_CPY:
|
||||||
case GGML_OP_DUP:
|
case GGML_OP_DUP:
|
||||||
|
Reference in New Issue
Block a user