Commit Graph

  • 380838dce8 Add Whisper::Context#full_parallel Kitaiti Makoto 2024-11-16 04:19:26 +0900
  • f5753ce863 Add test for Whisper::Context#full_parallel Kitaiti Makoto 2024-11-16 04:19:20 +0900
  • a1066c467d Add description to Whisper::Context#full Kitaiti Makoto 2024-11-16 04:18:58 +0900
  • 79ec5498b7 Add additional signature for Whisper::Context#full Kitaiti Makoto 2024-11-14 18:56:23 +0900
  • ff44e911bb Add document of Whisper::Context#full [skip ci] Kitaiti Makoto 2024-11-14 18:52:47 +0900
  • f71e330d6c Add tests for Whisper::Error Kitaiti Makoto 2024-11-14 18:37:10 +0900
  • fbc4d0d07f Add Whisper::Context#full Kitaiti Makoto 2024-11-14 18:37:01 +0900
  • 4ddb3f3528 Add tests for Whisper::Context#full Kitaiti Makoto 2024-11-14 08:08:43 +0900
  • 8c6a9b8bb6
    ruby : Follow source tree change (#2580) KITAITI Makoto 2024-11-22 00:04:29 +0900
  • 628e404425 Use ternary operator Kitaiti Makoto 2024-11-21 22:50:51 +0900
  • 7b90773d47 Use GitHub workflow setting for dependency definition Kitaiti Makoto 2024-11-21 22:45:01 +0900
  • 67a67a5b7b Fix paths in GitHub workflow for Ruby bindings Kitaiti Makoto 2024-11-21 22:28:18 +0900
  • b0aeef2d52
    ci : fix windows builds to use 2019 gg/ci-fix-windows Georgi Gerganov 2024-11-21 14:28:14 +0200
  • 647696eef5 Follow whisper.cpp log level change Kitaiti Makoto 2024-11-21 21:26:04 +0900
  • f31aa20843 Update whispercpp.gemspec Kitaiti Makoto 2024-11-21 21:21:55 +0900
  • 5b9997c424 Follow whisper.cpp source tree change Kitaiti Makoto 2024-11-21 21:14:28 +0900
  • 60c293e943 openvino : Pass CPU threads parameter Karthick J 2024-11-21 12:43:14 +0530
  • b67bdc9430
    disable gg/objc Georgi Gerganov 2024-11-20 23:18:58 +0200
  • 5e966f7844
    try3 Georgi Gerganov 2024-11-20 22:02:49 +0200
  • 54005478af
    try2 Georgi Gerganov 2024-11-20 21:42:58 +0200
  • 49c389b40a
    examples : try to fix objc CI Georgi Gerganov 2024-11-20 21:28:43 +0200
  • 37c88027e1 whisper : use backend registry (#0) Georgi Gerganov 2024-11-20 15:32:34 +0200
  • 9db070a3c5 ggml/sched : do not skip views in pre-assignments slaren 2024-11-20 13:25:08 +0100
  • 7fd8d9c220 whisper : adapt to new ggml (wip) Georgi Gerganov 2024-11-19 19:09:07 +0200
  • 06e059b8f8 talk-llama : sync llama.cpp Georgi Gerganov 2024-11-19 19:08:57 +0200
  • c9f49d5f9d sync : ggml Georgi Gerganov 2024-11-19 19:04:21 +0200
  • f4c1d7df39 ggml : sync resolve (skip) (#0) Georgi Gerganov 2024-11-19 19:03:47 +0200
  • 339b8e559c Add required ggml-base and backend libs to cmake pkg (llama/10407) bandoti 2024-11-19 12:10:30 -0400
  • 5f6d6919b4 cuda : fix CUDA_FLAGS not being applied (llama/10403) Diego Devesa 2024-11-19 14:29:38 +0100
  • 8ee767732f sycl : Add option to set the SYCL architecture for all targets (llama/10266) Romain Biessy 2024-11-19 09:02:23 +0100
  • 45f1f9144f vulkan: Optimize soft_max (llama/10301) Jeff Bolz 2024-11-19 01:25:17 -0600
  • 53589c8f12 sycl: Revert MUL_MAT_OP support changes (llama/10385) Alberto Cabrera Pérez 2024-11-19 00:50:04 +0000
  • 7ac2f17fac cuda : only use native when supported by cmake (llama/10389) Diego Devesa 2024-11-18 18:43:40 +0100
  • 48862c7b27 vulkan: remove use of null initializer (llama/10372) Jeff Bolz 2024-11-18 08:28:42 -0600
  • 44f7d9f4e3 metal : fox offset integer overflows in im2col (ggml/1015) Plamen Minev 2024-11-18 15:02:27 +0200
  • fd12302587 Vulkan: Fix device info output format specifiers (llama/10366) 0cc4m 2024-11-18 11:02:43 +0100
  • f80bef4630 metal : add GGML_UNARY_OP_ELU kernel (ggml/1018) PAB 2024-11-18 10:02:49 +0100
  • 161b443514 CUDA: fix MMV kernel being used for FP16 src1 (llama/10357) Johannes Gäßler 2024-11-17 23:20:42 +0100
  • ef7fbe1c66 CMake: fix typo in comment [no ci] (llama/10360) Johannes Gäßler 2024-11-17 12:59:38 +0100
  • 0879d3599e llama : only use default buffer types for the KV cache (llama/10358) Diego Devesa 2024-11-17 12:25:45 +0100
  • 2a444dc5bd metal : refactor kernel args into structs (llama/10238) Georgi Gerganov 2024-11-17 11:23:01 +0200
  • 45cf1634dc ggml : fix undefined reference to 'getcpu' (llama/10354) FirstTimeEZ 2024-11-17 21:39:22 +1300
  • dcb2922d1d CUDA: remove DMMV, consolidate F16 mult mat vec (llama/10318) Johannes Gäßler 2024-11-17 09:09:55 +0100
  • 3c5c751174 CMake: default to -arch=native for CUDA build (llama/10320) Johannes Gäßler 2024-11-17 09:06:34 +0100
  • 24ad19d0e9 ggml : fix possible buffer use after free in sched reserve (llama/9930) Diego Devesa 2024-11-17 07:31:17 +0100
  • bd574b05af ggml : inttypes.h -> cinttypes (llama/0) Georgi Gerganov 2024-11-16 23:40:39 +0200
  • 7e0eafcb1e ggml : adapt AMX to tensor->grad removal (llama/0) Georgi Gerganov 2024-11-16 21:38:01 +0200
  • 75670ae673 ggml : fix compile warnings (llama/0) Georgi Gerganov 2024-11-16 21:32:41 +0200
  • d4fcdf602b llamafile : fix include path (llama/0) Georgi Gerganov 2024-11-16 17:58:56 +0200
  • 1bebb1a116 vulkan: Optimize some mat-vec mul quant shaders (llama/10296) Jeff Bolz 2024-11-16 00:26:57 -0600
  • ee437cde59 ggml : optimize Q4_0 into Q4_0_X_Y repack (llama/10324) Dan Johansson 2024-11-16 01:53:37 +0100
  • c1506d38cf Make updates to fix issues with clang-cl builds while using AVX512 flags (llama/10314) Srihari-mcw 2024-11-16 02:57:00 +0530
  • c9541741e6 ggml: new optimization interface (ggml/988) Johannes Gäßler 2024-11-16 13:49:35 +0100
  • 6a55015dc4 ggml : remove duplicated sources from the last sync (ggml/1017) Georgi Gerganov 2024-11-15 23:52:31 +0200
  • 7e86030d4d ggml : fix some build issues slaren 2024-11-15 20:20:54 +0100
  • 401fbea326 sync : leftovers (ggml/0) Georgi Gerganov 2024-11-15 21:43:41 +0200
  • 44d1cbdfe9 cmake : restore CMakeLists.txt (llama/10256) Georgi Gerganov 2024-11-15 21:35:51 +0200
  • 3216efef2e AVX BF16 and single scale quant optimizations (llama/10212) Eve 2024-11-15 11:47:58 +0000
  • 2c0484ebf7 sycl: Use syclcompat::dp4a (llama/10267) Romain Biessy 2024-11-15 04:09:12 +0100
  • 3298916e5e backend cpu: add online flow for aarch64 Q4_0 GEMV/GEMM kernels (llama/9921) Charles Xu 2024-11-15 01:28:50 +0100
  • 746bf2596f ggml : build backends as libraries (llama/10256) Diego Devesa 2024-11-14 18:04:35 +0100
  • 5f7e094ccb scripts : update sync Georgi Gerganov 2024-11-19 18:59:18 +0200
  • e6114173b8
    whisper : use backend registry (#0) Georgi Gerganov 2024-11-20 15:32:34 +0200
  • 85ff4f974e Fix crash in ggml_vk_print_gpu_info Juliusz Chroboczek 2024-11-20 16:35:35 +0100
  • c800966378 ggml/sched : do not skip views in pre-assignments slaren 2024-11-20 13:25:08 +0100
  • 8c24c64924
    whisper : adapt to new ggml (wip) Georgi Gerganov 2024-11-19 19:09:07 +0200
  • 4e1f516ecc
    talk-llama : sync llama.cpp Georgi Gerganov 2024-11-19 19:08:57 +0200
  • 0eddc9fcbc
    sync : ggml Georgi Gerganov 2024-11-19 19:04:21 +0200
  • 52799f9082
    ggml : sync resolve (skip) (#0) Georgi Gerganov 2024-11-19 19:03:47 +0200
  • bfaf1fc76f
    Add required ggml-base and backend libs to cmake pkg (llama/10407) bandoti 2024-11-19 12:10:30 -0400
  • 166237d07e
    cuda : fix CUDA_FLAGS not being applied (llama/10403) Diego Devesa 2024-11-19 14:29:38 +0100
  • d2aaf9ecfc
    sycl : Add option to set the SYCL architecture for all targets (llama/10266) Romain Biessy 2024-11-19 09:02:23 +0100
  • 29894ef822
    vulkan: Optimize soft_max (llama/10301) Jeff Bolz 2024-11-19 01:25:17 -0600
  • 8d6e30fb61
    sycl: Revert MUL_MAT_OP support changes (llama/10385) Alberto Cabrera Pérez 2024-11-19 00:50:04 +0000
  • 761d310e78
    cuda : only use native when supported by cmake (llama/10389) Diego Devesa 2024-11-18 18:43:40 +0100
  • c4f4639466
    vulkan: remove use of null initializer (llama/10372) Jeff Bolz 2024-11-18 08:28:42 -0600
  • c157f624e2
    metal : fox offset integer overflows in im2col (ggml/1015) Plamen Minev 2024-11-18 15:02:27 +0200
  • 748d633638
    Vulkan: Fix device info output format specifiers (llama/10366) 0cc4m 2024-11-18 11:02:43 +0100
  • 937684c822
    metal : add GGML_UNARY_OP_ELU kernel (ggml/1018) PAB 2024-11-18 10:02:49 +0100
  • 58b5fc45b9
    CUDA: fix MMV kernel being used for FP16 src1 (llama/10357) Johannes Gäßler 2024-11-17 23:20:42 +0100
  • fcd8ea6aff
    CMake: fix typo in comment [no ci] (llama/10360) Johannes Gäßler 2024-11-17 12:59:38 +0100
  • 6b4de57e65
    llama : only use default buffer types for the KV cache (llama/10358) Diego Devesa 2024-11-17 12:25:45 +0100
  • dca00d8374
    metal : refactor kernel args into structs (llama/10238) Georgi Gerganov 2024-11-17 11:23:01 +0200
  • a901ba0716
    ggml : fix undefined reference to 'getcpu' (llama/10354) FirstTimeEZ 2024-11-17 21:39:22 +1300
  • 8bd8688888
    CUDA: remove DMMV, consolidate F16 mult mat vec (llama/10318) Johannes Gäßler 2024-11-17 09:09:55 +0100
  • 77ea626d26
    CMake: default to -arch=native for CUDA build (llama/10320) Johannes Gäßler 2024-11-17 09:06:34 +0100
  • c96434f2b3
    ggml : fix possible buffer use after free in sched reserve (llama/9930) Diego Devesa 2024-11-17 07:31:17 +0100
  • 3f1a78d6f8
    ggml : inttypes.h -> cinttypes (llama/0) Georgi Gerganov 2024-11-16 23:40:39 +0200
  • 600728ea21
    ggml : adapt AMX to tensor->grad removal (llama/0) Georgi Gerganov 2024-11-16 21:38:01 +0200
  • e726307095
    ggml : fix compile warnings (llama/0) Georgi Gerganov 2024-11-16 21:32:41 +0200
  • 7caa6b2e83
    llamafile : fix include path (llama/0) Georgi Gerganov 2024-11-16 17:58:56 +0200
  • 68b198b438
    vulkan: Optimize some mat-vec mul quant shaders (llama/10296) Jeff Bolz 2024-11-16 00:26:57 -0600
  • 49ca4814be
    ggml : optimize Q4_0 into Q4_0_X_Y repack (llama/10324) Dan Johansson 2024-11-16 01:53:37 +0100
  • 4b8ddfbda7
    Make updates to fix issues with clang-cl builds while using AVX512 flags (llama/10314) Srihari-mcw 2024-11-16 02:57:00 +0530
  • adf81dc329
    ggml: new optimization interface (ggml/988) Johannes Gäßler 2024-11-16 13:49:35 +0100
  • f33c7ea0c5
    ggml : remove duplicated sources from the last sync (ggml/1017) Georgi Gerganov 2024-11-15 23:52:31 +0200
  • 83c77397e4
    ggml : fix some build issues slaren 2024-11-15 20:20:54 +0100
  • 8dffd6444c
    sync : leftovers (ggml/0) Georgi Gerganov 2024-11-15 21:43:41 +0200
  • 1d49a2e7a2
    cmake : restore CMakeLists.txt (llama/10256) Georgi Gerganov 2024-11-15 21:35:51 +0200
  • 0df66d6586
    AVX BF16 and single scale quant optimizations (llama/10212) Eve 2024-11-15 11:47:58 +0000