whisper.cpp/ggml
Johannes Gäßler 24d8534bd8 CPU/CUDA: Gemma 2 FlashAttention support (llama/8542)
* CPU/CUDA: Gemma 2 FlashAttention support

* apply logit_softcap to scale in kernel

* disable logit softcapping tests on Metal

* remove metal check
2024-08-28 13:22:20 +03:00
..
cmake whisper : reorganize source code + improve CMake (#2256) 2024-06-26 19:34:09 +03:00
include CPU/CUDA: Gemma 2 FlashAttention support (llama/8542) 2024-08-28 13:22:20 +03:00
src CPU/CUDA: Gemma 2 FlashAttention support (llama/8542) 2024-08-28 13:22:20 +03:00
.gitignore whisper : reorganize source code + improve CMake (#2256) 2024-06-26 19:34:09 +03:00
CMakeLists.txt cmake : remove unused option GGML_CURL (llama/9011) 2024-08-28 13:22:20 +03:00
ggml_vk_generate_shaders.py whisper : reorganize source code + improve CMake (#2256) 2024-06-26 19:34:09 +03:00