| #include "arg.h" |
| #include "common.h" |
| #include "sampling.h" |
| #include "speculative.h" |
| #include "log.h" |
| #include "llama.h" |
|
|
| #include <clocale> |
| #include <cstdio> |
| #include <cstring> |
| #include <string> |
| #include <vector> |
|
|
| int main(int argc, char ** argv) { |
| std::setlocale(LC_NUMERIC, "C"); |
|
|
| common_params params; |
|
|
| if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_SPECULATIVE)) { |
| return 1; |
| } |
|
|
| if (params.n_predict < -1) { |
| LOG_ERR("%s: --n-predict must be >= -1\n", __func__); |
| return 1; |
| } |
|
|
| common_init(); |
|
|
| if (params.speculative.mparams_dft.path.empty()) { |
| LOG_ERR("%s: --model-draft is required\n", __func__); |
| return 1; |
| } |
|
|
| |
| llama_backend_init(); |
| llama_numa_init(params.numa); |
|
|
| llama_model * model_tgt = NULL; |
|
|
| llama_context * ctx_tgt = NULL; |
|
|
| |
| auto llama_init_tgt = common_init_from_params(params); |
|
|
| model_tgt = llama_init_tgt->model(); |
| ctx_tgt = llama_init_tgt->context(); |
|
|
| const llama_vocab * vocab = llama_model_get_vocab(model_tgt); |
|
|
| |
| llama_model_ptr model_dft; |
|
|
| |
| { |
| const auto & params_spec = params.speculative; |
|
|
| auto params_dft = params; |
|
|
| params_dft.n_parallel = 1; |
| params_dft.n_ctx = params_spec.n_ctx; |
| params_dft.n_batch = llama_n_ctx_seq(ctx_tgt); |
| params_dft.devices = params_spec.devices; |
| params_dft.model = params_spec.mparams_dft; |
| params_dft.n_gpu_layers = params_spec.n_gpu_layers; |
|
|
| if (params_spec.cpuparams.n_threads > 0) { |
| params_dft.cpuparams.n_threads = params.speculative.cpuparams.n_threads; |
| params_dft.cpuparams_batch.n_threads = params.speculative.cpuparams_batch.n_threads; |
| } |
|
|
| params_dft.tensor_buft_overrides = params.speculative.tensor_buft_overrides; |
|
|
| auto mparams_dft = common_model_params_to_llama(params_dft); |
|
|
| model_dft.reset(llama_model_load_from_file(params_dft.model.path.c_str(), mparams_dft)); |
| if (model_dft == nullptr) { |
| LOG_ERR("failed to load draft model, '%s'\n", params_dft.model.path.c_str()); |
| return 1; |
| } |
|
|
| params.speculative.model_dft = model_dft.get(); |
| params.speculative.cparams_dft = common_context_params_to_llama(params_dft); |
| } |
|
|
| |
| std::vector<llama_token> inp; |
| inp = common_tokenize(ctx_tgt, params.prompt, true, true); |
|
|
| if (llama_n_ctx(ctx_tgt) < (uint32_t) inp.size()) { |
| LOG_ERR("%s: the prompt exceeds the context size (%d tokens, ctx %d)\n", __func__, (int) inp.size(), llama_n_ctx(ctx_tgt)); |
|
|
| return 1; |
| } |
|
|
| if (llama_n_batch(ctx_tgt) < (uint32_t) inp.size()) { |
| LOG_ERR("%s: the prompt exceeds the batch size (%d tokens, batch %d)\n", __func__, (int) inp.size(), llama_n_batch(ctx_tgt)); |
|
|
| return 1; |
| } |
|
|
| LOG("\n\n"); |
|
|
| for (auto id : inp) { |
| LOG("%s", common_token_to_piece(ctx_tgt, id).c_str()); |
| } |
|
|
| int n_predict = 0; |
| int n_drafted = 0; |
| int n_accept = 0; |
|
|
| |
| bool has_eos = false; |
|
|
| |
| |
| |
|
|
| const auto t_enc_start = ggml_time_us(); |
|
|
| |
| struct common_sampler * smpl = common_sampler_init(model_tgt, params.sampling); |
|
|
| |
| llama_decode(ctx_tgt, llama_batch_get_one(inp.data(), inp.size() - 1)); |
|
|
| |
| llama_token id_last = inp.back(); |
|
|
| |
| llama_tokens prompt_tgt(inp.begin(), inp.end() - 1); |
| prompt_tgt.reserve(llama_n_ctx(ctx_tgt)); |
|
|
| int n_past = inp.size() - 1; |
|
|
| |
| const auto & params_spec = params.speculative; |
|
|
| struct common_speculative * spec = common_speculative_init(params.speculative, ctx_tgt); |
|
|
| common_speculative_begin(spec, prompt_tgt); |
|
|
| llama_batch batch_tgt = llama_batch_init(llama_n_batch(ctx_tgt), 0, 1); |
|
|
| const auto t_enc_end = ggml_time_us(); |
|
|
| const auto t_dec_start = ggml_time_us(); |
|
|
| while (true) { |
| |
| |
| |
| |
| |
| |
| |
| llama_tokens draft = common_speculative_draft(spec, params_spec, prompt_tgt, id_last); |
|
|
| |
|
|
| |
| common_batch_clear(batch_tgt); |
| common_batch_add (batch_tgt, id_last, n_past++, { 0 }, true); |
|
|
| |
| { |
| |
| if (draft.size() < (size_t) params_spec.n_min) { |
| draft.clear(); |
| } |
|
|
| for (size_t i = 0; i < draft.size(); ++i) { |
| common_batch_add(batch_tgt, draft[i], n_past + i, { 0 }, true); |
| } |
|
|
| |
|
|
| llama_decode(ctx_tgt, batch_tgt); |
| } |
|
|
| |
| |
| |
| |
| |
| |
| |
| const auto ids = common_sampler_sample_and_accept_n(smpl, ctx_tgt, draft); |
|
|
| |
|
|
| GGML_ASSERT(ids.size() > 0); |
|
|
| n_past += ids.size() - 1; |
| n_drafted += draft.size(); |
| n_accept += ids.size() - 1; |
| n_predict += ids.size(); |
|
|
| |
| |
| |
| |
| |
| for (size_t i = 0; i < ids.size(); ++i) { |
| prompt_tgt.push_back(id_last); |
|
|
| id_last = ids[i]; |
|
|
| if (llama_vocab_is_eog(vocab, id_last)) { |
| has_eos = true; |
| break; |
| } |
|
|
| const std::string token_str = common_token_to_piece(ctx_tgt, id_last); |
|
|
| if (params.use_color && i + 1 < ids.size()) { |
| LOG("\u001b[%dm%s\u001b[37m", (36 - 0 % 6), token_str.c_str()); |
| } else { |
| LOG("%s", token_str.c_str()); |
| } |
| } |
|
|
| LOG_DBG("accepted %d/%d draft tokens, the last target token is: (%d)\n", (int) ids.size() - 1, (int) draft.size(), id_last); |
|
|
| { |
| LOG_DBG("clear kv cache from any extra tokens, n_past = %d\n", n_past); |
|
|
| llama_memory_seq_rm(llama_get_memory(ctx_tgt), 0, n_past, -1); |
| } |
|
|
| if ((params.n_predict >= 0 && n_predict > params.n_predict) || has_eos) { |
| break; |
| } |
| } |
|
|
| auto t_dec_end = ggml_time_us(); |
|
|
| const int n_input = inp.size(); |
|
|
| LOG("\n\n"); |
|
|
| LOG_INF("encoded %4d tokens in %8.3f seconds, speed: %8.3f t/s\n", n_input, (t_enc_end - t_enc_start) / 1e6f, inp.size() / ((t_enc_end - t_enc_start) / 1e6f)); |
| LOG_INF("decoded %4d tokens in %8.3f seconds, speed: %8.3f t/s\n", n_predict, (t_dec_end - t_dec_start) / 1e6f, n_predict / ((t_dec_end - t_dec_start) / 1e6f)); |
|
|
| LOG_INF("\n"); |
| LOG_INF("n_draft = %d\n", params_spec.n_max); |
| LOG_INF("n_predict = %d\n", n_predict); |
| LOG_INF("n_drafted = %d\n", n_drafted); |
| LOG_INF("n_accept = %d\n", n_accept); |
| LOG_INF("accept = %.3f%%\n", 100.0f * n_accept / n_drafted); |
|
|
| LOG_INF("\n"); |
| LOG_INF("draft:\n\n"); |
|
|
| LOG_INF("\n"); |
| LOG_INF("target:\n\n"); |
| common_perf_print(ctx_tgt, smpl); |
|
|
| llama_batch_free(batch_tgt); |
|
|
| common_sampler_free(smpl); |
| common_speculative_free(spec); |
|
|
| llama_backend_free(); |
|
|
| LOG("\n\n"); |
|
|
| return 0; |
| } |
|
|