mirror of
https://github.com/ggerganov/whisper.cpp.git
synced 2024-12-28 17:49:05 +01:00
models : change HF hosting from dataset to model
This commit is contained in:
parent
09e9068007
commit
1beff6f66d
@ -466,7 +466,7 @@ The original models are converted to a custom binary format. This allows to pack
|
|||||||
You can download the converted models using the [models/download-ggml-model.sh](models/download-ggml-model.sh) script
|
You can download the converted models using the [models/download-ggml-model.sh](models/download-ggml-model.sh) script
|
||||||
or manually from here:
|
or manually from here:
|
||||||
|
|
||||||
- https://huggingface.co/datasets/ggerganov/whisper.cpp
|
- https://huggingface.co/ggerganov/whisper.cpp
|
||||||
- https://ggml.ggerganov.com
|
- https://ggml.ggerganov.com
|
||||||
|
|
||||||
For more details, see the conversion script [models/convert-pt-to-ggml.py](models/convert-pt-to-ggml.py) or the README
|
For more details, see the conversion script [models/convert-pt-to-ggml.py](models/convert-pt-to-ggml.py) or the README
|
||||||
|
@ -17,9 +17,9 @@ import (
|
|||||||
// CONSTANTS
|
// CONSTANTS
|
||||||
|
|
||||||
const (
|
const (
|
||||||
srcUrl = "https://huggingface.co/datasets/ggerganov/whisper.cpp/resolve/main" // The location of the models
|
srcUrl = "https://huggingface.co/ggerganov/whisper.cpp/resolve/main" // The location of the models
|
||||||
srcExt = ".bin" // Filename extension
|
srcExt = ".bin" // Filename extension
|
||||||
bufSize = 1024 * 64 // Size of the buffer used for downloading the model
|
bufSize = 1024 * 64 // Size of the buffer used for downloading the model
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
|
@ -31,7 +31,7 @@ To run this, you will need a ggml GPT-2 model: [instructions](https://github.com
|
|||||||
Alternatively, you can simply download the smallest ggml GPT-2 117M model (240 MB) like this:
|
Alternatively, you can simply download the smallest ggml GPT-2 117M model (240 MB) like this:
|
||||||
|
|
||||||
```
|
```
|
||||||
wget --quiet --show-progress -O models/ggml-gpt-2-117M.bin https://huggingface.co/datasets/ggerganov/ggml/raw/main/ggml-model-gpt-2-117M.bin
|
wget --quiet --show-progress -O models/ggml-gpt-2-117M.bin https://huggingface.co/ggerganov/ggml/raw/main/ggml-model-gpt-2-117M.bin
|
||||||
```
|
```
|
||||||
|
|
||||||
## TTS
|
## TTS
|
||||||
|
@ -6,7 +6,7 @@ using the [convert-pt-to-ggml.py](convert-pt-to-ggml.py) script. You can either
|
|||||||
the `ggml` files yourself using the conversion script, or you can use the [download-ggml-model.sh](download-ggml-model.sh)
|
the `ggml` files yourself using the conversion script, or you can use the [download-ggml-model.sh](download-ggml-model.sh)
|
||||||
script to download the already converted models. Currently, they are hosted on the following locations:
|
script to download the already converted models. Currently, they are hosted on the following locations:
|
||||||
|
|
||||||
- https://huggingface.co/datasets/ggerganov/whisper.cpp
|
- https://huggingface.co/ggerganov/whisper.cpp
|
||||||
- https://ggml.ggerganov.com
|
- https://ggml.ggerganov.com
|
||||||
|
|
||||||
Sample usage:
|
Sample usage:
|
||||||
@ -23,7 +23,7 @@ You can now use it like this:
|
|||||||
|
|
||||||
A third option to obtain the model files is to download them from Hugging Face:
|
A third option to obtain the model files is to download them from Hugging Face:
|
||||||
|
|
||||||
https://huggingface.co/datasets/ggerganov/whisper.cpp/tree/main
|
https://huggingface.co/ggerganov/whisper.cpp/tree/main
|
||||||
|
|
||||||
## Available models
|
## Available models
|
||||||
|
|
||||||
|
@ -40,7 +40,7 @@ if exist "ggml-%model%.bin" (
|
|||||||
goto :eof
|
goto :eof
|
||||||
)
|
)
|
||||||
|
|
||||||
PowerShell -NoProfile -ExecutionPolicy Bypass -Command "Invoke-WebRequest -Uri https://huggingface.co/datasets/ggerganov/whisper.cpp/resolve/main/ggml-%model%.bin -OutFile ggml-%model%.bin"
|
PowerShell -NoProfile -ExecutionPolicy Bypass -Command "Invoke-WebRequest -Uri https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-%model%.bin -OutFile ggml-%model%.bin"
|
||||||
|
|
||||||
if %ERRORLEVEL% neq 0 (
|
if %ERRORLEVEL% neq 0 (
|
||||||
echo Failed to download ggml model %model%
|
echo Failed to download ggml model %model%
|
||||||
|
@ -6,7 +6,7 @@
|
|||||||
#src="https://ggml.ggerganov.com"
|
#src="https://ggml.ggerganov.com"
|
||||||
#pfx="ggml-model-whisper"
|
#pfx="ggml-model-whisper"
|
||||||
|
|
||||||
src="https://huggingface.co/datasets/ggerganov/whisper.cpp"
|
src="https://huggingface.co/ggerganov/whisper.cpp"
|
||||||
pfx="resolve/main/ggml"
|
pfx="resolve/main/ggml"
|
||||||
|
|
||||||
# get the path of this script
|
# get the path of this script
|
||||||
|
@ -631,7 +631,6 @@ struct whisper_context {
|
|||||||
int64_t t_load_us = 0;
|
int64_t t_load_us = 0;
|
||||||
int64_t t_start_us = 0;
|
int64_t t_start_us = 0;
|
||||||
|
|
||||||
|
|
||||||
ggml_type wtype = ggml_type::GGML_TYPE_F16; // weight type (FP32 or FP16)
|
ggml_type wtype = ggml_type::GGML_TYPE_F16; // weight type (FP32 or FP16)
|
||||||
|
|
||||||
whisper_model model;
|
whisper_model model;
|
||||||
|
Loading…
Reference in New Issue
Block a user