Commit Graph

  • 1f14567ee6
    ggml : do not define GGML_USE_CUDA when building with GGML_BACKEND_DL (llama/11211) Radoslav Gerganov 2025-01-13 13:31:41 +0200
  • 618d94abb4
    Vulkan: Fix float16 use on devices without float16 support + fix subgroup_size_control validation error (llama/11161) 0cc4m 2025-01-10 06:39:33 +0100
  • 2d6f599774
    llama: add support for QRWKV6 model architecture (llama/11001) Molly Sophia 2025-01-10 09:58:08 +0800
  • fe7bb8849d
    SYCL: Refactor ggml_sycl_compute_forward (llama/11121) Akarshan Biswas 2025-01-10 05:43:03 +0530
  • 40aa3fa643
    fix: add missing msg in static_assert (llama/11143) hydai 2025-01-09 04:03:28 +0800
  • 3272320d98
    llamafile : ppc64le MMA INT8 implementation (llama/10912) amritahs-ibm 2025-01-08 16:24:19 +0530
  • e322e918a3
    Disable GL_KHR_cooperative_matrix Vulkan extension if not available. (llama/11117) Mathieu Baudier 2025-01-08 09:18:13 +0100
  • f2031c56c2
    fix: Vulkan shader gen binary path when Cross-compiling (llama/11096) ag2s20150909 2025-01-08 16:17:29 +0800
  • a48be79914
    GGUF: C++ refactor, backend support, misc fixes (llama/11030) Johannes Gäßler 2025-01-07 18:01:58 +0100
  • 6679500ba3
    ggml-backend : only offload from host buffers (fix) (llama/11124) Diego Devesa 2025-01-07 16:11:57 +0100
  • c324f37090
    ggml-backend : only offload from host buffers (llama/11120) Diego Devesa 2025-01-07 12:38:05 +0100
  • f0783516ac
    rpc : code cleanup (llama/11107) Radoslav Gerganov 2025-01-07 08:37:02 +0200
  • c52b2f6d50
    SYCL: Use get_multi_ptr instead of deprecated get_pointer in wkv6 (llama/11087) Akarshan Biswas 2025-01-07 11:56:07 +0530
  • 9325f4af05
    CUDA: add BF16 support (llama/11093) Johannes Gäßler 2025-01-06 02:33:52 +0100
  • fbe4db4881
    Vulkan: Add device-specific blacklist for coopmat for the AMD proprietary driver (llama/11074) 0cc4m 2025-01-04 21:09:59 +0100
  • aababf16c8
    Support for models with non-512-aligned tensors over RPC. (llama/11047) matt23654 2025-01-04 16:10:30 +0000
  • 95cd1d3276
    fix: Vulkan shader gen binary path (llama/11037) Gilad S. 2025-01-04 10:17:31 +0200
  • 19c147b26d
    ggml : allow loading backend with env variable (ggml/1059) Radoslav Gerganov 2025-01-05 09:50:37 +0200
  • 507e230f1e
    scripts : sync opencl, gguf Georgi Gerganov 2025-01-14 09:42:16 +0200
  • 3037f1d5ee Update whisper.objc xcode project with files that have been removed and add new files that were missing Corey Earwood 2025-01-13 22:19:12 -0700
  • 3915a1b1f4 Disable GL_KHR_cooperative_matrix Vulkan extension if not available Pepijn de Vos 2025-01-13 22:01:56 +0100
  • eb68324c86
    whisper : fix gpu device selection (#2728) Georgi Gerganov 2025-01-13 13:11:37 +0200
  • c719c5be54
    whisper : fix gpu device selection Georgi Gerganov 2025-01-13 09:56:32 +0200
  • e940fbf283
    server : fix build (#2718) Georgi Gerganov 2025-01-13 08:57:33 +0200
  • 35d0e02c72
    talk-llama : sync llama.cpp (#2709) Georgi Gerganov 2025-01-13 08:55:48 +0200
  • 45d3faf961
    server : generate unique tmp filenames (#2718) NETZkultur GmbH 2025-01-13 07:55:21 +0100
  • f1fcab6eca
    Merge 7b7c9eb005 into 2ab2eb5110 Don Mahurin 2025-01-10 02:38:36 +0100
  • 6feb5f2690
    Use Unique Filenames for FFmpeg Conversion to Prevent File Overwrites NETZkultur GmbH 2025-01-09 15:48:55 +0100
  • 2ab2eb5110
    whisper : add whisper_full_get_segment_no_speech_prob_from_state (#2716) Sandro Hanea 2025-01-09 15:21:07 +0100
  • e18231285e Exposed whisper_full_get_segment_no_speech_prob_from_state in addition to context based retrieval Sandro Hanea 2025-01-09 11:46:24 +0000
  • b82d305282
    readme : add docker instructions (#2711) Jayant 2025-01-07 12:20:51 +0100
  • 744c2c431e
    Adding back docker instructions in v1.7.4 Jayant 2025-01-06 17:49:16 +0100
  • 9d2aafa153
    talk-llama : sync llama.cpp Georgi Gerganov 2025-01-06 15:25:33 +0200
  • 885e31368d
    docs: Fix main -> whisper-cli in download scripts (#2707) Adam Jones 2025-01-06 13:17:57 +0000
  • 8a9ad7844d
    release : v1.7.4 v1.7.4 Georgi Gerganov 2025-01-06 15:13:48 +0200
  • eb874b3a3c
    ci : cont Georgi Gerganov 2025-01-06 10:46:10 +0200
  • eb78e3a3f1
    ci : fix ubuntu runner names Georgi Gerganov 2025-01-06 09:29:10 +0200
  • d1a467f30a docs: Fix main -> whisper-cli in download scripts Adam Jones 2025-01-06 01:45:22 +0000
  • f99263e420 Run vad_simple on entire pcmf32, not on the last step Tamotsu Takahashi 2025-01-05 08:47:22 +0900
  • ece3ff88f6
    cli : fix segfault on missing argument (#2700) Yusuf Redžić 2025-01-04 09:47:41 +0100
  • 9366544991 ci : fix arm builds Georgi Gerganov 2025-01-03 16:24:02 +0200
  • 95583942ed sync : ggml Georgi Gerganov 2025-01-03 14:11:23 +0200
  • 2e93cb6a2f ggml : do not install metal source when embed library (ggml/1054) Georgi Gerganov 2025-01-03 14:11:20 +0200
  • de5cd60d1c metal : avoid uint (llama/11019) Georgi Gerganov 2025-01-03 11:26:14 +0200
  • 3fcba3e58b ggml : fixes for AVXVNNI instruction set with MSVC and Clang (llama/11027) Srihari-mcw 2024-12-31 19:53:33 +0530
  • cea5f1c52f vulkan: optimize mul_mat for small values of N (llama/10991) Jeff Bolz 2024-12-30 11:27:11 -0600
  • 2112462db4 vulkan: im2col and matmul optimizations for stable diffusion (llama/10942) Jeff Bolz 2024-12-29 03:16:34 -0600
  • fc84ecd445 vulkan: Use push constant offset to handle misaligned descriptors (llama/10987) Jeff Bolz 2024-12-29 02:35:11 -0600
  • 8de1e99907 vulkan: multi-row k quants (llama/10846) Eve 2024-12-26 10:54:44 -0500
  • 499af9294a examples, ggml : fix GCC compiler warnings (llama/10983) Peter 2024-12-27 00:59:11 +1100
  • bcf937c216 ggml : more perfo with llamafile tinyblas on x86_64 (llama/10714) Djip007 2024-12-24 18:54:49 +0100
  • b8d90953d7 ggml : use wstring for backend search paths (llama/10960) Diego Devesa 2024-12-24 04:05:27 +0100
  • 60a422147b ggml : fix arm enabled features check (llama/10961) Diego Devesa 2024-12-24 04:05:17 +0100
  • 3387415bad ggml : fix const usage in SSE path (llama/10962) Diego Devesa 2024-12-23 20:25:52 +0100
  • 536ca3ec89 ggml : fix run-time on FreeBSD in get_executable_path() (llama/10948) yuri@FreeBSD 2024-12-22 16:20:11 -0800
  • a4bb983190 vulkan: build fixes for 32b (llama/10927) Jeff Bolz 2024-12-22 03:44:01 -0600
  • 39c205f555 vulkan: optimize coopmat2 dequant functions (llama/10855) Jeff Bolz 2024-12-21 01:04:45 -0600
  • 6d502f33dc ggml-cpu: replace NEON asm with intrinsics in ggml_gemv_q4_0_4x8_q8_0() (llama/10874) Adrien Gallouët 2024-12-21 00:33:37 +0100
  • 5ea27d089d SYCL: Migrate away from deprecated ggml_tensor->backend (llama/10840) Akarshan Biswas 2024-12-20 21:01:28 +0530
  • 1462d92588 ggml : add test for SVE and disable when it fails (llama/10906) Diego Devesa 2024-12-20 13:31:28 +0100
  • 7ba1a41f47 ggml: fix arm build with gcc (llama/10895) Adrien Gallouët 2024-12-19 14:20:41 +0100
  • 5ea088636f ggml : fix arm build (llama/10890) Diego Devesa 2024-12-18 23:21:42 +0100
  • f32ddb3b1c tts : add OuteTTS support (llama/10784) Georgi Gerganov 2024-12-18 19:27:21 +0200
  • 79b75ece03 tests: add tests for GGUF (llama/10830) Johannes Gäßler 2024-12-17 19:09:35 +0100
  • 6348d73e55 ggml : improve inputs log sched_print_assignments (ggml/1053) Daniel Bevenius 2024-12-19 03:50:12 +0100
  • 5e60c6de3b
    ci : fix arm builds Georgi Gerganov 2025-01-03 16:24:02 +0200
  • 09d49febbf cli : fix segfault on missing argument Yusuf Redzic 2025-01-03 17:48:43 +0100
  • 1deb9a6151
    sync : ggml Georgi Gerganov 2025-01-03 14:11:23 +0200
  • 0024db855b
    ggml : do not install metal source when embed library (ggml/1054) Georgi Gerganov 2025-01-03 14:11:20 +0200
  • 6a0441b002
    metal : avoid uint (llama/11019) Georgi Gerganov 2025-01-03 11:26:14 +0200
  • 11d52f5d35
    ggml : fixes for AVXVNNI instruction set with MSVC and Clang (llama/11027) Srihari-mcw 2024-12-31 19:53:33 +0530
  • b9ca755a59
    vulkan: optimize mul_mat for small values of N (llama/10991) Jeff Bolz 2024-12-30 11:27:11 -0600
  • c43ba3787f
    vulkan: im2col and matmul optimizations for stable diffusion (llama/10942) Jeff Bolz 2024-12-29 03:16:34 -0600
  • 980d41ec7a
    vulkan: Use push constant offset to handle misaligned descriptors (llama/10987) Jeff Bolz 2024-12-29 02:35:11 -0600
  • aa0887e8af
    vulkan: multi-row k quants (llama/10846) Eve 2024-12-26 10:54:44 -0500
  • 55a5cf35a2
    examples, ggml : fix GCC compiler warnings (llama/10983) Peter 2024-12-27 00:59:11 +1100
  • fbaadc216d
    ggml : more perfo with llamafile tinyblas on x86_64 (llama/10714) Djip007 2024-12-24 18:54:49 +0100
  • 3edc4b0db1
    ggml : use wstring for backend search paths (llama/10960) Diego Devesa 2024-12-24 04:05:27 +0100
  • 6d27ca5bb7
    ggml : fix arm enabled features check (llama/10961) Diego Devesa 2024-12-24 04:05:17 +0100
  • ac6af9e766
    ggml : fix const usage in SSE path (llama/10962) Diego Devesa 2024-12-23 20:25:52 +0100
  • b1151f92a0
    ggml : fix run-time on FreeBSD in get_executable_path() (llama/10948) yuri@FreeBSD 2024-12-22 16:20:11 -0800
  • 6c01a1eb4c
    vulkan: build fixes for 32b (llama/10927) Jeff Bolz 2024-12-22 03:44:01 -0600
  • 0914a27c53
    vulkan: optimize coopmat2 dequant functions (llama/10855) Jeff Bolz 2024-12-21 01:04:45 -0600
  • 18bfd3181f
    ggml-cpu: replace NEON asm with intrinsics in ggml_gemv_q4_0_4x8_q8_0() (llama/10874) Adrien Gallouët 2024-12-21 00:33:37 +0100
  • f86cf2b1e1
    SYCL: Migrate away from deprecated ggml_tensor->backend (llama/10840) Akarshan Biswas 2024-12-20 21:01:28 +0530
  • 011ca37a19
    ggml : add test for SVE and disable when it fails (llama/10906) Diego Devesa 2024-12-20 13:31:28 +0100
  • 58d1a1b4b9
    ggml: fix arm build with gcc (llama/10895) Adrien Gallouët 2024-12-19 14:20:41 +0100
  • 7927fce84a
    ggml : fix arm build (llama/10890) Diego Devesa 2024-12-18 23:21:42 +0100
  • 3aa63a8c2f
    tts : add OuteTTS support (llama/10784) Georgi Gerganov 2024-12-18 19:27:21 +0200
  • 0d4d69a9cc
    tests: add tests for GGUF (llama/10830) Johannes Gäßler 2024-12-17 19:09:35 +0100
  • a371a941fb
    ggml : improve inputs log sched_print_assignments (ggml/1053) Daniel Bevenius 2024-12-19 03:50:12 +0100
  • 0a84581f20 Make stream more test-friendly Tamotsu Takahashi 2025-01-03 15:28:47 +0900
  • fb36a1538a
    readme : fix real-time audio input example build instructions (#2692) Samuel Durante 2025-01-02 07:05:38 -0300
  • c81b8b910b
    objc : rename ggml-cpu-aarch64.c to .cpp (#2687) Alter 2025-01-02 10:05:09 +0000
  • 85b60f31d0
    docs : replace Core ML with OpenVINO (#2686) Konosuke Sakai 2025-01-02 19:03:02 +0900
  • 425d3add59 Fix windows build Tamotsu Takahashi 2025-01-02 13:53:26 +0900
  • 17c7600416 Fix inconsistency of ifdef Tamotsu Takahashi 2025-01-02 13:35:06 +0900
  • 61222da541 Fix windows build (include fcntl.h) Tamotsu Takahashi 2025-01-02 12:44:39 +0900
  • 03b25dd7f3 Remove unused n_new_line Tamotsu Takahashi 2025-01-02 12:40:11 +0900
  • 75099f9f87 Fix armv7-linux build Tamotsu Takahashi 2025-01-02 12:32:01 +0900