Compare commits

..

2 Commits

Author SHA1 Message Date
13c5446759 Update ggml-cuda/mmvq.cu
Co-authored-by: Johannes Gäßler <johannesg@5d6.de>
2024-06-11 17:37:32 +03:00
9df6298a91 cuda : fix bounds check for src0 rows in MMVQ kernel 2024-06-11 11:30:12 +03:00
3 changed files with 3 additions and 3 deletions

View File

@ -459,7 +459,7 @@ jobs:
path: build/bin/${{ matrix.build }}
windows-cublas:
runs-on: windows-2019
runs-on: windows-latest
strategy:
matrix:

View File

@ -297,10 +297,10 @@ ggml-cuda/%.o: ggml-cuda/%.cu ggml-cuda/%.cuh ggml.h ggml-common.h ggml-cuda/com
ggml-cuda.o: ggml-cuda.cu ggml-cuda.h ggml.h ggml-backend.h ggml-backend-impl.h ggml-common.h $(wildcard ggml-cuda/*.cuh)
$(NVCC) $(NVCCFLAGS) $(CXXFLAGS) -Wno-pedantic -c $< -o $@
endif
whisper-mel-cuda.o: whisper-mel-cuda.cu whisper.h ggml.h ggml-backend.h whisper-mel.hpp whisper-mel-cuda.hpp
$(NVCC) $(NVCCFLAGS) $(CXXFLAGS) -Wno-pedantic -c $< -o $@
endif
ifdef WHISPER_HIPBLAS
ROCM_PATH ?= /opt/rocm

View File

@ -3167,7 +3167,7 @@ struct mel_calc_cpu : public whisper_mel_calc {
}
whisper_mel_calc * whisper_mel_calc_create(ggml_backend_t backend, const whisper_filters & filters) {
#if defined(GGML_USE_CUDA) && !defined(GGML_USE_HIPBLAS)
#if GGML_USE_CUDA
if (ggml_backend_is_cuda(backend)) {
auto ret = whisper_mel_calc_create_cuda(backend, filters);
// run a warmup to avoid the first kernel launch overhead (thus we get the best perf even on the first run)