mirror of
https://github.com/ggerganov/whisper.cpp.git
synced 2024-12-26 00:29:21 +01:00
main : fix sampling time + add max_context parameter
This commit is contained in:
parent
72e9cdd6bf
commit
85d6e1e1e7
@ -42,6 +42,7 @@ struct whisper_params {
|
|||||||
int32_t n_threads = std::min(4, (int32_t) std::thread::hardware_concurrency());
|
int32_t n_threads = std::min(4, (int32_t) std::thread::hardware_concurrency());
|
||||||
int32_t offset_t_ms = 0;
|
int32_t offset_t_ms = 0;
|
||||||
int32_t offset_n = 0;
|
int32_t offset_n = 0;
|
||||||
|
int32_t max_context = -1;
|
||||||
|
|
||||||
bool verbose = false;
|
bool verbose = false;
|
||||||
bool translate = false;
|
bool translate = false;
|
||||||
@ -77,6 +78,8 @@ bool whisper_params_parse(int argc, char ** argv, whisper_params & params) {
|
|||||||
params.offset_t_ms = std::stoi(argv[++i]);
|
params.offset_t_ms = std::stoi(argv[++i]);
|
||||||
} else if (arg == "-on" || arg == "--offset-n") {
|
} else if (arg == "-on" || arg == "--offset-n") {
|
||||||
params.offset_n = std::stoi(argv[++i]);
|
params.offset_n = std::stoi(argv[++i]);
|
||||||
|
} else if (arg == "-mc" || arg == "--max-context") {
|
||||||
|
params.max_context = std::stoi(argv[++i]);
|
||||||
} else if (arg == "-v" || arg == "--verbose") {
|
} else if (arg == "-v" || arg == "--verbose") {
|
||||||
params.verbose = true;
|
params.verbose = true;
|
||||||
} else if (arg == "--translate") {
|
} else if (arg == "--translate") {
|
||||||
@ -127,6 +130,7 @@ void whisper_print_usage(int argc, char ** argv, const whisper_params & params)
|
|||||||
fprintf(stderr, " -t N, --threads N number of threads to use during computation (default: %d)\n", params.n_threads);
|
fprintf(stderr, " -t N, --threads N number of threads to use during computation (default: %d)\n", params.n_threads);
|
||||||
fprintf(stderr, " -ot N, --offset-t N time offset in milliseconds (default: %d)\n", params.offset_t_ms);
|
fprintf(stderr, " -ot N, --offset-t N time offset in milliseconds (default: %d)\n", params.offset_t_ms);
|
||||||
fprintf(stderr, " -on N, --offset-n N segment index offset (default: %d)\n", params.offset_n);
|
fprintf(stderr, " -on N, --offset-n N segment index offset (default: %d)\n", params.offset_n);
|
||||||
|
fprintf(stderr, " -mc N, --max-context N maximum number of text context tokens to store (default: max)\n");
|
||||||
fprintf(stderr, " -v, --verbose verbose output\n");
|
fprintf(stderr, " -v, --verbose verbose output\n");
|
||||||
fprintf(stderr, " --translate translate from source language to english\n");
|
fprintf(stderr, " --translate translate from source language to english\n");
|
||||||
fprintf(stderr, " -otxt, --output-txt output result in a text file\n");
|
fprintf(stderr, " -otxt, --output-txt output result in a text file\n");
|
||||||
@ -380,6 +384,8 @@ int main(int argc, char ** argv) {
|
|||||||
wparams.translate = params.translate;
|
wparams.translate = params.translate;
|
||||||
wparams.language = params.language.c_str();
|
wparams.language = params.language.c_str();
|
||||||
wparams.n_threads = params.n_threads;
|
wparams.n_threads = params.n_threads;
|
||||||
|
wparams.n_processors = 1;
|
||||||
|
wparams.n_max_text_ctx = params.max_context >= 0 ? params.max_context : wparams.n_max_text_ctx;
|
||||||
wparams.offset_ms = params.offset_t_ms;
|
wparams.offset_ms = params.offset_t_ms;
|
||||||
|
|
||||||
// this callback is called on each new segment
|
// this callback is called on each new segment
|
||||||
|
22
whisper.cpp
22
whisper.cpp
@ -211,14 +211,6 @@ struct whisper_vocab {
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
struct whisper_token_data {
|
|
||||||
whisper_token id; // token id
|
|
||||||
whisper_token tid; // forced timestamp token id
|
|
||||||
|
|
||||||
float p; // probability of the token
|
|
||||||
float pt; // probability of the timestamp token
|
|
||||||
};
|
|
||||||
|
|
||||||
struct whisper_segment {
|
struct whisper_segment {
|
||||||
int64_t t0;
|
int64_t t0;
|
||||||
int64_t t1;
|
int64_t t1;
|
||||||
@ -2219,7 +2211,7 @@ int whisper_decode(struct whisper_context * ctx, const whisper_token * tokens, i
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
whisper_token whisper_sample_best(struct whisper_context * ctx) {
|
whisper_token_data whisper_sample_best(struct whisper_context * ctx) {
|
||||||
const int64_t t_start_sample_us = ggml_time_us();
|
const int64_t t_start_sample_us = ggml_time_us();
|
||||||
|
|
||||||
// TODO: simplify
|
// TODO: simplify
|
||||||
@ -2227,7 +2219,7 @@ whisper_token whisper_sample_best(struct whisper_context * ctx) {
|
|||||||
|
|
||||||
ctx->t_sample_us += ggml_time_us() - t_start_sample_us;
|
ctx->t_sample_us += ggml_time_us() - t_start_sample_us;
|
||||||
|
|
||||||
return res.id;
|
return res;
|
||||||
}
|
}
|
||||||
|
|
||||||
whisper_token whisper_sample_timestamp(struct whisper_context * ctx) {
|
whisper_token whisper_sample_timestamp(struct whisper_context * ctx) {
|
||||||
@ -2330,8 +2322,9 @@ struct whisper_full_params whisper_full_default_params(enum whisper_sampling_str
|
|||||||
/*.strategy =*/ WHISPER_SAMPLING_GREEDY,
|
/*.strategy =*/ WHISPER_SAMPLING_GREEDY,
|
||||||
|
|
||||||
/*.n_threads =*/ std::min(4, (int32_t) std::thread::hardware_concurrency()),
|
/*.n_threads =*/ std::min(4, (int32_t) std::thread::hardware_concurrency()),
|
||||||
/*.offset_ms =*/ 0,
|
|
||||||
/*.n_processors =*/ 1,
|
/*.n_processors =*/ 1,
|
||||||
|
/*.n_max_text_ctx =*/ 16384,
|
||||||
|
/*.offset_ms =*/ 0,
|
||||||
|
|
||||||
/*.translate =*/ false,
|
/*.translate =*/ false,
|
||||||
/*.no_context =*/ false,
|
/*.no_context =*/ false,
|
||||||
@ -2362,8 +2355,9 @@ struct whisper_full_params whisper_full_default_params(enum whisper_sampling_str
|
|||||||
/*.strategy =*/ WHISPER_SAMPLING_BEAM_SEARCH,
|
/*.strategy =*/ WHISPER_SAMPLING_BEAM_SEARCH,
|
||||||
|
|
||||||
/*.n_threads =*/ std::min(4, (int32_t) std::thread::hardware_concurrency()),
|
/*.n_threads =*/ std::min(4, (int32_t) std::thread::hardware_concurrency()),
|
||||||
/*.offset_ms =*/ 0,
|
|
||||||
/*.n_processors =*/ 1,
|
/*.n_processors =*/ 1,
|
||||||
|
/*.n_max_text_ctx =*/ 16384,
|
||||||
|
/*.offset_ms =*/ 0,
|
||||||
|
|
||||||
/*.translate =*/ false,
|
/*.translate =*/ false,
|
||||||
/*.no_context =*/ false,
|
/*.no_context =*/ false,
|
||||||
@ -2470,7 +2464,7 @@ int whisper_full(
|
|||||||
|
|
||||||
// if we have already generated some text, use it as a prompt to condition the next generation
|
// if we have already generated some text, use it as a prompt to condition the next generation
|
||||||
if (prompt_past.size() > 0) {
|
if (prompt_past.size() > 0) {
|
||||||
int n_take = std::min(whisper_n_text_ctx(ctx)/2, int(prompt_past.size()));
|
int n_take = std::min(std::min(params.n_max_text_ctx, whisper_n_text_ctx(ctx)/2), int(prompt_past.size()));
|
||||||
|
|
||||||
prompt = { whisper_token_prev(ctx) };
|
prompt = { whisper_token_prev(ctx) };
|
||||||
prompt.insert(prompt.begin() + 1, prompt_past.end() - n_take, prompt_past.end());
|
prompt.insert(prompt.begin() + 1, prompt_past.end() - n_take, prompt_past.end());
|
||||||
@ -2512,7 +2506,7 @@ int whisper_full(
|
|||||||
// feel free to experiment!
|
// feel free to experiment!
|
||||||
//
|
//
|
||||||
{
|
{
|
||||||
auto token = whisper_sample_best(ctx->vocab, ctx->probs.data() + (ctx->probs.size() - ctx->vocab.n_vocab));
|
auto token = whisper_sample_best(ctx);
|
||||||
|
|
||||||
if (i == 0) {
|
if (i == 0) {
|
||||||
token.tid = whisper_token_beg(ctx);
|
token.tid = whisper_token_beg(ctx);
|
||||||
|
13
whisper.h
13
whisper.h
@ -68,6 +68,14 @@ extern "C" {
|
|||||||
|
|
||||||
typedef int whisper_token;
|
typedef int whisper_token;
|
||||||
|
|
||||||
|
struct whisper_token_data {
|
||||||
|
whisper_token id; // token id
|
||||||
|
whisper_token tid; // forced timestamp token id
|
||||||
|
|
||||||
|
float p; // probability of the token
|
||||||
|
float pt; // probability of the timestamp token
|
||||||
|
};
|
||||||
|
|
||||||
// Allocates all memory needed for the model and loads the model from the given file.
|
// Allocates all memory needed for the model and loads the model from the given file.
|
||||||
// Returns NULL on failure.
|
// Returns NULL on failure.
|
||||||
WHISPER_API struct whisper_context * whisper_init(const char * path_model);
|
WHISPER_API struct whisper_context * whisper_init(const char * path_model);
|
||||||
@ -122,7 +130,7 @@ extern "C" {
|
|||||||
// You can also implement your own sampling method using the whisper_get_probs() function.
|
// You can also implement your own sampling method using the whisper_get_probs() function.
|
||||||
// whisper_sample_best() returns the token with the highest probability
|
// whisper_sample_best() returns the token with the highest probability
|
||||||
// whisper_sample_timestamp() returns the most probable timestamp token
|
// whisper_sample_timestamp() returns the most probable timestamp token
|
||||||
WHISPER_API whisper_token whisper_sample_best(struct whisper_context * ctx);
|
WHISPER_API whisper_token_data whisper_sample_best(struct whisper_context * ctx);
|
||||||
WHISPER_API whisper_token whisper_sample_timestamp(struct whisper_context * ctx);
|
WHISPER_API whisper_token whisper_sample_timestamp(struct whisper_context * ctx);
|
||||||
|
|
||||||
// Return the id of the specified language, returns -1 if not found
|
// Return the id of the specified language, returns -1 if not found
|
||||||
@ -171,8 +179,9 @@ extern "C" {
|
|||||||
enum whisper_sampling_strategy strategy;
|
enum whisper_sampling_strategy strategy;
|
||||||
|
|
||||||
int n_threads;
|
int n_threads;
|
||||||
int offset_ms;
|
|
||||||
int n_processors;
|
int n_processors;
|
||||||
|
int n_max_text_ctx;
|
||||||
|
int offset_ms;
|
||||||
|
|
||||||
bool translate;
|
bool translate;
|
||||||
bool no_context;
|
bool no_context;
|
||||||
|
Loading…
Reference in New Issue
Block a user