mirror of
https://github.com/ggerganov/whisper.cpp.git
synced 2025-06-05 01:07:19 +02:00
45 lines
1.5 KiB
C
45 lines
1.5 KiB
C
#pragma once
|
|
|
|
#include "llama.h"
|
|
#include "llama-io.h"
|
|
#include "llama-memory.h"
|
|
|
|
struct llama_kv_cache : public llama_memory_i {
|
|
virtual ~llama_kv_cache() = default;
|
|
|
|
// split the input batch into a set of ubatches and verify that they can fit into the cache
|
|
// return a state object containing the ubatches and KV cache state required to process them
|
|
// check the llama_memory_state_i::get_status() for the result
|
|
virtual llama_memory_state_ptr init_batch(
|
|
const llama_batch & batch,
|
|
uint32_t n_ubatch,
|
|
bool embd_pooled,
|
|
bool logits_all) = 0;
|
|
|
|
// simulate full cache, used for allocating worst-case compute buffers
|
|
virtual llama_memory_state_ptr init_full() = 0;
|
|
|
|
// process any pending defrag/shift/etc. operations
|
|
// optionally call once before processing a new batch
|
|
// return true if any operations were performed
|
|
virtual bool update(llama_context & lctx) = 0;
|
|
|
|
// schedule a defrag if the fragmentation threshold is exceeded. otherwise, do nothing
|
|
// TODO: change to
|
|
// llama_memory_state_ptr init_defrag(float thold) = 0;
|
|
//
|
|
virtual void defrag_sched(float thold) = 0;
|
|
|
|
// getters
|
|
virtual bool get_can_shift() const = 0;
|
|
|
|
bool get_can_edit() const override { return get_can_shift(); }
|
|
|
|
//
|
|
// state write/read
|
|
//
|
|
|
|
virtual void state_write(llama_io_write_i & io, llama_seq_id seq_id = -1) const = 0;
|
|
virtual void state_read (llama_io_read_i & io, llama_seq_id seq_id = -1) = 0;
|
|
};
|