mirror of
https://github.com/ggerganov/whisper.cpp.git
synced 2024-11-07 16:44:13 +01:00
b6c5f49b78
* whisper : add whisper_batch * whisper : move kv_self to whisper_state * whisper : full batched decoding support * whisper : fix memory leak in whisper_batch * whisper : fix mem leak again + remove oboslete function * whisper : clear kv cache when using whisper_decode API * whisper : speed-up sampling * whisper : fix decoders initializer * bench : add batch size 5 bench * whisper : add comment about the KV cache size * whisper : add check for max number of decoders * whisper : avoid starting sampling threads with bs=1 * whisper : enable beam-search by default * cuda : sync llama.cpp fixes |
||
---|---|---|
.. | ||
bench-all.sh | ||
bench-wts.sh | ||
bench.py | ||
convert-all.sh | ||
deploy-wasm.sh | ||
quantize-all.sh | ||
sha-all.sh | ||
sync-ggml.sh |