| #include "sampling.h" |
|
|
| #include "common.h" |
| #include "log.h" |
| #include "reasoning-budget.h" |
|
|
| #include <algorithm> |
| #include <cmath> |
| #include <cstring> |
| #include <unordered_map> |
|
|
| |
| |
| template<typename T> |
| struct ring_buffer { |
| ring_buffer(size_t cap) : capacity(cap), data(cap) {} |
|
|
| T & front() { |
| if (sz == 0) { |
| throw std::runtime_error("ring buffer is empty"); |
| } |
| return data[first]; |
| } |
|
|
| const T & front() const { |
| if (sz == 0) { |
| throw std::runtime_error("ring buffer is empty"); |
| } |
| return data[first]; |
| } |
|
|
| T & back() { |
| if (sz == 0) { |
| throw std::runtime_error("ring buffer is empty"); |
| } |
| return data[pos]; |
| } |
|
|
| const T & back() const { |
| if (sz == 0) { |
| throw std::runtime_error("ring buffer is empty"); |
| } |
| return data[pos]; |
| } |
|
|
| void push_back(const T & value) { |
| if (sz == capacity) { |
| |
| first = (first + 1) % capacity; |
| } else { |
| sz++; |
| } |
| data[pos] = value; |
| pos = (pos + 1) % capacity; |
| } |
|
|
| T pop_front() { |
| if (sz == 0) { |
| throw std::runtime_error("ring buffer is empty"); |
| } |
| T value = data[first]; |
| first = (first + 1) % capacity; |
| sz--; |
| return value; |
| } |
|
|
| const T & rat(size_t i) const { |
| if (i >= sz) { |
| throw std::runtime_error("ring buffer: index out of bounds"); |
| } |
| return data[(first + sz - i - 1) % capacity]; |
| } |
|
|
| std::vector<T> to_vector() const { |
| std::vector<T> result; |
| result.reserve(sz); |
| for (size_t i = 0; i < sz; i++) { |
| result.push_back(data[(first + i) % capacity]); |
| } |
| return result; |
| } |
|
|
| void clear() { |
| |
| sz = 0; |
| first = 0; |
| pos = 0; |
| } |
|
|
| bool empty() const { |
| return sz == 0; |
| } |
|
|
| size_t size() const { |
| return sz; |
| } |
|
|
| size_t capacity = 0; |
| size_t sz = 0; |
| size_t first = 0; |
| size_t pos = 0; |
| std::vector<T> data; |
| }; |
|
|
| struct common_sampler { |
| common_params_sampling params; |
|
|
| struct llama_sampler * grmr; |
| struct llama_sampler * chain; |
|
|
| ring_buffer<llama_token> prev; |
|
|
| std::vector<llama_token_data> cur; |
|
|
| llama_token_data_array cur_p; |
|
|
| void reset() { |
| prev.clear(); |
|
|
| llama_sampler_reset(chain); |
| } |
|
|
| void set_logits(struct llama_context * ctx, int idx) { |
| const float * sampled_probs = llama_get_sampled_probs_ith (ctx, idx); |
| const float * sampled_logits = llama_get_sampled_logits_ith (ctx, idx); |
| const llama_token * sampled_ids = llama_get_sampled_candidates_ith(ctx, idx); |
|
|
| const llama_model * model = llama_get_model(ctx); |
| const llama_vocab * vocab = llama_model_get_vocab(model); |
|
|
| const int n_vocab = llama_vocab_n_tokens(vocab); |
|
|
| if (sampled_probs) { |
| const uint32_t sampled_probs_count = llama_get_sampled_probs_count_ith(ctx, idx); |
| cur.resize(sampled_probs_count); |
| for (uint32_t i = 0; i < sampled_probs_count; ++i) { |
| cur[i] = llama_token_data{sampled_ids[i], sampled_logits[i], sampled_probs[i]}; |
| } |
| } else if (sampled_logits) { |
| const uint32_t sampled_logits_count = llama_get_sampled_logits_count_ith(ctx, idx); |
| cur.resize(sampled_logits_count); |
| for (uint32_t i = 0; i < sampled_logits_count; i++) { |
| cur[i] = llama_token_data{sampled_ids[i], sampled_logits[i], 0.0f}; |
| } |
| } else { |
| const auto * logits = llama_get_logits_ith(ctx, idx); |
| GGML_ASSERT(logits != nullptr); |
| cur.resize(n_vocab); |
| for (llama_token token_id = 0; token_id < n_vocab; token_id++) { |
| cur[token_id] = llama_token_data{token_id, logits[token_id], 0.0f}; |
| } |
| } |
|
|
| cur_p = { cur.data(), cur.size(), -1, false }; |
| } |
|
|
| common_time_meas tm() { |
| return common_time_meas(t_total_us, params.no_perf); |
| } |
|
|
| mutable int64_t t_total_us = 0; |
| }; |
|
|
| std::string common_params_sampling::print() const { |
| char result[1024]; |
|
|
| snprintf(result, sizeof(result), |
| "\trepeat_last_n = %d, repeat_penalty = %.3f, frequency_penalty = %.3f, presence_penalty = %.3f\n" |
| "\tdry_multiplier = %.3f, dry_base = %.3f, dry_allowed_length = %d, dry_penalty_last_n = %d\n" |
| "\ttop_k = %d, top_p = %.3f, min_p = %.3f, xtc_probability = %.3f, xtc_threshold = %.3f, typical_p = %.3f, top_n_sigma = %.3f, temp = %.3f\n" |
| "\tmirostat = %d, mirostat_lr = %.3f, mirostat_ent = %.3f, adaptive_target = %.3f, adaptive_decay = %.3f", |
| penalty_last_n, penalty_repeat, penalty_freq, penalty_present, |
| dry_multiplier, dry_base, dry_allowed_length, dry_penalty_last_n, |
| top_k, top_p, min_p, xtc_probability, xtc_threshold, typ_p, top_n_sigma, temp, |
| mirostat, mirostat_eta, mirostat_tau, adaptive_target, adaptive_decay); |
|
|
| return std::string(result); |
| } |
|
|
| struct common_sampler * common_sampler_init(const struct llama_model * model, struct common_params_sampling & params) { |
| const llama_vocab * vocab = llama_model_get_vocab(model); |
|
|
| llama_sampler_chain_params lparams = llama_sampler_chain_default_params(); |
|
|
| lparams.no_perf = params.no_perf; |
|
|
| llama_sampler * grmr = nullptr; |
| llama_sampler * chain = llama_sampler_chain_init(lparams); |
|
|
| std::vector<llama_sampler *> samplers; |
|
|
| if (params.grammar.compare(0, 11, "%llguidance") == 0) { |
| #ifdef LLAMA_USE_LLGUIDANCE |
| grmr = llama_sampler_init_llg(vocab, "lark", params.grammar.c_str()); |
| #else |
| GGML_ABORT("llguidance (cmake -DLLAMA_LLGUIDANCE=ON) is not enabled"); |
| #endif |
| } else { |
| std::vector<std::string> trigger_patterns; |
| std::vector<llama_token> trigger_tokens; |
| for (const auto & trigger : params.grammar_triggers) { |
| switch (trigger.type) { |
| case COMMON_GRAMMAR_TRIGGER_TYPE_WORD: |
| { |
| const auto & word = trigger.value; |
| trigger_patterns.push_back(regex_escape(word)); |
| break; |
| } |
| case COMMON_GRAMMAR_TRIGGER_TYPE_PATTERN: |
| { |
| trigger_patterns.push_back(trigger.value); |
| break; |
| } |
| case COMMON_GRAMMAR_TRIGGER_TYPE_PATTERN_FULL: |
| { |
| const auto & pattern = trigger.value; |
| std::string anchored = "^$"; |
| if (!pattern.empty()) { |
| anchored = (pattern.front() != '^' ? "^" : "") |
| + pattern |
| + (pattern.back() != '$' ? "$" : ""); |
| } |
| trigger_patterns.push_back(anchored); |
| break; |
| } |
| case COMMON_GRAMMAR_TRIGGER_TYPE_TOKEN: |
| { |
| const auto token = trigger.token; |
| trigger_tokens.push_back(token); |
| break; |
| } |
| default: |
| GGML_ASSERT(false && "unknown trigger type"); |
| } |
| } |
|
|
| std::vector<const char *> trigger_patterns_c; |
| trigger_patterns_c.reserve(trigger_patterns.size()); |
| for (const auto & regex : trigger_patterns) { |
| trigger_patterns_c.push_back(regex.c_str()); |
| } |
|
|
| if (!params.grammar.empty()) { |
| if (params.grammar_lazy) { |
| grmr = llama_sampler_init_grammar_lazy_patterns(vocab, params.grammar.c_str(), "root", |
| trigger_patterns_c.data(), trigger_patterns_c.size(), |
| trigger_tokens.data(), trigger_tokens.size()); |
| } else { |
| grmr = llama_sampler_init_grammar(vocab, params.grammar.c_str(), "root"); |
| } |
| } |
| } |
|
|
| |
| if (params.reasoning_budget_tokens >= 0 && !params.reasoning_budget_forced.empty()) { |
| samplers.push_back(common_reasoning_budget_init( |
| vocab, |
| params.reasoning_budget_start, |
| params.reasoning_budget_end, |
| params.reasoning_budget_forced, |
| params.reasoning_budget_tokens, |
| params.reasoning_budget_activate_immediately ? REASONING_BUDGET_COUNTING : REASONING_BUDGET_IDLE)); |
| } |
|
|
| if (params.has_logit_bias()) { |
| samplers.push_back(llama_sampler_init_logit_bias(llama_vocab_n_tokens(vocab), params.logit_bias.size(), params.logit_bias.data())); |
| } |
|
|
| if (params.mirostat == 0) { |
|
|
| bool use_adaptive_p = false; |
|
|
| for (const auto & cnstr : params.samplers) { |
| switch (cnstr) { |
| case COMMON_SAMPLER_TYPE_DRY: |
| { |
| std::vector<const char *> c_breakers; |
| c_breakers.reserve(params.dry_sequence_breakers.size()); |
| for (const auto & str : params.dry_sequence_breakers) { |
| c_breakers.push_back(str.c_str()); |
| } |
| samplers.push_back(llama_sampler_init_dry(vocab, llama_model_n_ctx_train(model), params.dry_multiplier, params.dry_base, params.dry_allowed_length, params.dry_penalty_last_n, c_breakers.data(), c_breakers.size())); |
| } |
| break; |
| case COMMON_SAMPLER_TYPE_TOP_K: |
| samplers.push_back(llama_sampler_init_top_k(params.top_k)); |
| break; |
| case COMMON_SAMPLER_TYPE_TOP_P: |
| samplers.push_back(llama_sampler_init_top_p(params.top_p, params.min_keep)); |
| break; |
| case COMMON_SAMPLER_TYPE_TOP_N_SIGMA: |
| samplers.push_back(llama_sampler_init_top_n_sigma(params.top_n_sigma)); |
| break; |
| case COMMON_SAMPLER_TYPE_MIN_P: |
| samplers.push_back(llama_sampler_init_min_p(params.min_p, params.min_keep)); |
| break; |
| case COMMON_SAMPLER_TYPE_XTC: |
| samplers.push_back(llama_sampler_init_xtc(params.xtc_probability, params.xtc_threshold, params.min_keep, params.seed)); |
| break; |
| case COMMON_SAMPLER_TYPE_TYPICAL_P: |
| samplers.push_back(llama_sampler_init_typical(params.typ_p, params.min_keep)); |
| break; |
| case COMMON_SAMPLER_TYPE_TEMPERATURE: |
| samplers.push_back(llama_sampler_init_temp_ext(params.temp, params.dynatemp_range, params.dynatemp_exponent)); |
| break; |
| case COMMON_SAMPLER_TYPE_INFILL: |
| samplers.push_back(llama_sampler_init_infill(vocab)); |
| break; |
| case COMMON_SAMPLER_TYPE_PENALTIES: |
| samplers.push_back(llama_sampler_init_penalties(params.penalty_last_n, params.penalty_repeat, params.penalty_freq, params.penalty_present)); |
| break; |
| case COMMON_SAMPLER_TYPE_ADAPTIVE_P: |
| |
| |
| |
| |
| use_adaptive_p = true; |
| break; |
| default: |
| GGML_ASSERT(false && "unknown sampler type"); |
| } |
| } |
| if (use_adaptive_p) { |
| |
| samplers.push_back(llama_sampler_init_adaptive_p(params.adaptive_target, params.adaptive_decay, params.seed)); |
| } else { |
| |
| samplers.push_back(llama_sampler_init_dist(params.seed)); |
| } |
| } else if (params.mirostat == 1) { |
| samplers.push_back(llama_sampler_init_temp(params.temp)); |
| samplers.push_back(llama_sampler_init_mirostat(llama_vocab_n_tokens(vocab), params.seed, params.mirostat_tau, params.mirostat_eta, 100)); |
| } else if (params.mirostat == 2) { |
| samplers.push_back(llama_sampler_init_temp(params.temp)); |
| samplers.push_back(llama_sampler_init_mirostat_v2(params.seed, params.mirostat_tau, params.mirostat_eta)); |
| } else { |
| GGML_ASSERT(false && "unknown mirostat version"); |
| } |
|
|
| for (auto * smpl : samplers) { |
| llama_sampler_chain_add(chain, smpl); |
| } |
|
|
| if (grmr && params.backend_sampling) { |
| LOG_WRN("%s: backend sampling is not compatible with grammar, disabling\n", __func__); |
|
|
| params.backend_sampling = false; |
| } |
|
|
| auto * result = new common_sampler { |
| params, |
| grmr, |
| chain, |
| ring_buffer<llama_token>(std::max(32, params.n_prev)), |
| {}, |
| {}, |
| }; |
|
|
| return result; |
| } |
|
|
| void common_sampler_free(struct common_sampler * gsmpl) { |
| if (!gsmpl) { |
| return; |
| } |
|
|
| llama_sampler_free(gsmpl->grmr); |
| llama_sampler_free(gsmpl->chain); |
|
|
| delete gsmpl; |
| } |
|
|
| void common_sampler_accept(struct common_sampler * gsmpl, llama_token token, bool accept_grammar) { |
| if (!gsmpl) { |
| return; |
| } |
|
|
| const auto tm = gsmpl->tm(); |
|
|
| if (gsmpl->grmr && accept_grammar) { |
| llama_sampler_accept(gsmpl->grmr, token); |
| } |
|
|
| llama_sampler_accept(gsmpl->chain, token); |
|
|
| gsmpl->prev.push_back(token); |
| } |
|
|
| void common_sampler_reset(struct common_sampler * gsmpl) { |
| if (!gsmpl) { |
| return; |
| } |
|
|
| gsmpl->reset(); |
| } |
|
|
| struct common_sampler * common_sampler_clone(common_sampler * gsmpl) { |
| return new common_sampler { |
| gsmpl->params, |
| llama_sampler_clone(gsmpl->grmr), |
| llama_sampler_clone(gsmpl->chain), |
| gsmpl->prev, |
| gsmpl->cur, |
| gsmpl->cur_p, |
| }; |
| } |
|
|
| void common_perf_print(const struct llama_context * ctx, const struct common_sampler * gsmpl) { |
| |
|
|
| const double t_sampling_ms = gsmpl ? 1e-3*gsmpl->t_total_us : 0; |
|
|
| llama_perf_sampler_data data_smpl; |
| llama_perf_context_data data_ctx; |
|
|
| memset(&data_smpl, 0, sizeof(data_smpl)); |
| memset(&data_ctx, 0, sizeof(data_ctx)); |
|
|
| if (gsmpl) { |
| auto & data = data_smpl; |
|
|
| data = llama_perf_sampler(gsmpl->chain); |
|
|
| |
| LOG_INF("%s: sampling time = %10.2f ms\n", __func__, t_sampling_ms); |
| LOG_INF("%s: samplers time = %10.2f ms / %5d tokens\n", __func__, data.t_sample_ms, data.n_sample); |
| } |
|
|
| if (ctx) { |
| auto & data = data_ctx; |
|
|
| data = llama_perf_context(ctx); |
|
|
| const double t_end_ms = 1e-3 * ggml_time_us(); |
|
|
| const double t_total_ms = t_end_ms - data.t_start_ms; |
| const double t_unacc_ms = t_total_ms - (t_sampling_ms + data.t_p_eval_ms + data.t_eval_ms); |
| const double t_unacc_pc = 100.0 * t_unacc_ms / t_total_ms; |
|
|
| LOG_INF("%s: load time = %10.2f ms\n", __func__, data.t_load_ms); |
| LOG_INF("%s: prompt eval time = %10.2f ms / %5d tokens (%8.2f ms per token, %8.2f tokens per second)\n", |
| __func__, data.t_p_eval_ms, data.n_p_eval, data.t_p_eval_ms / data.n_p_eval, 1e3 / data.t_p_eval_ms * data.n_p_eval); |
| LOG_INF("%s: eval time = %10.2f ms / %5d runs (%8.2f ms per token, %8.2f tokens per second)\n", |
| __func__, data.t_eval_ms, data.n_eval, data.t_eval_ms / data.n_eval, 1e3 / data.t_eval_ms * data.n_eval); |
| LOG_INF("%s: total time = %10.2f ms / %5d tokens\n", __func__, (t_end_ms - data.t_start_ms), (data.n_p_eval + data.n_eval)); |
| LOG_INF("%s: unaccounted time = %10.2f ms / %5.1f %% (total - sampling - prompt eval - eval) / (total)\n", __func__, t_unacc_ms, t_unacc_pc); |
| LOG_INF("%s: graphs reused = %10d\n", __func__, data.n_reused); |
|
|
| llama_memory_breakdown_print(ctx); |
| } |
| } |
|
|
| struct llama_sampler * common_sampler_get(const struct common_sampler * gsmpl) { |
| if (!gsmpl) { |
| return nullptr; |
| } |
|
|
| return gsmpl->chain; |
| } |
|
|
| llama_token common_sampler_sample(struct common_sampler * gsmpl, struct llama_context * ctx, int idx, bool grammar_first) { |
| llama_synchronize(ctx); |
|
|
| |
| const auto tm = gsmpl->tm(); |
|
|
| llama_token id = LLAMA_TOKEN_NULL; |
|
|
| auto & grmr = gsmpl->grmr; |
| auto & chain = gsmpl->chain; |
| auto & cur_p = gsmpl->cur_p; |
|
|
| |
| |
| { |
| id = llama_get_sampled_token_ith(ctx, idx); |
|
|
| if (id != LLAMA_TOKEN_NULL) { |
| LOG_DBG("%s: Backend sampler selected token: '%d'. Will not run any CPU samplers\n", __func__, id); |
|
|
| GGML_ASSERT(!gsmpl->grmr && "using grammar in combination with backend sampling is not supported"); |
|
|
| |
| gsmpl->cur.resize(1); |
| gsmpl->cur[0] = { id, 0.0f, 1.0f }; |
| cur_p = { gsmpl->cur.data(), gsmpl->cur.size(), 0, true }; |
|
|
| return id; |
| } |
| } |
|
|
| gsmpl->set_logits(ctx, idx); |
|
|
| if (grammar_first) { |
| llama_sampler_apply(grmr, &cur_p); |
| } |
|
|
| llama_sampler_apply(chain, &cur_p); |
|
|
| id = cur_p.data[cur_p.selected].id; |
|
|
| if (grammar_first) { |
| return id; |
| } |
|
|
| |
| { |
| llama_token_data single_token_data = { id, 1.0f, 0.0f }; |
| llama_token_data_array single_token_data_array = { &single_token_data, 1, -1, false }; |
|
|
| llama_sampler_apply(grmr, &single_token_data_array); |
|
|
| const bool is_valid = single_token_data_array.data[0].logit != -INFINITY; |
| if (is_valid) { |
| return id; |
| } |
| } |
|
|
| |
| |
| gsmpl->set_logits(ctx, idx); |
|
|
| llama_sampler_apply(grmr, &cur_p); |
| llama_sampler_apply(chain, &cur_p); |
|
|
| GGML_ASSERT(cur_p.selected != -1 && "no selected token during sampling - check your sampling configuration"); |
|
|
| id = cur_p.data[cur_p.selected].id; |
|
|
| return id; |
| } |
|
|
| std::vector<llama_token> common_sampler_sample_and_accept_n(struct common_sampler * gsmpl, struct llama_context * ctx, const std::vector<int> & idxs, const llama_tokens & draft, bool grammar_first) { |
| GGML_ASSERT(idxs.size() == draft.size() + 1 && "idxs.size() must be draft.size() + 1"); |
|
|
| std::vector<llama_token> result; |
| result.reserve(idxs.size()); |
|
|
| size_t i = 0; |
| for (; i < draft.size(); i++) { |
| const llama_token id = common_sampler_sample(gsmpl, ctx, idxs[i], grammar_first); |
|
|
| common_sampler_accept(gsmpl, id, true); |
|
|
| result.push_back(id); |
|
|
| if (draft[i] != id) { |
| break; |
| } |
| } |
|
|
| if (i == draft.size()) { |
| const llama_token id = common_sampler_sample(gsmpl, ctx, idxs[i], grammar_first); |
|
|
| common_sampler_accept(gsmpl, id, true); |
|
|
| result.push_back(id); |
| } |
|
|
| return result; |
| } |
|
|
| std::vector<llama_token> common_sampler_sample_and_accept_n(struct common_sampler * gsmpl, struct llama_context * ctx, const llama_tokens & draft, bool grammar_first) { |
| std::vector<int> idxs(draft.size() + 1); |
| for (size_t i = 0; i < idxs.size(); ++i) { |
| idxs[i] = i; |
| } |
|
|
| return common_sampler_sample_and_accept_n(gsmpl, ctx, idxs, draft, grammar_first); |
| } |
|
|
| uint32_t common_sampler_get_seed(const struct common_sampler * gsmpl) { |
| return llama_sampler_get_seed(gsmpl->chain); |
| } |
|
|
| |
|
|
| llama_token_data_array * common_sampler_get_candidates(struct common_sampler * gsmpl, bool do_sort) { |
| const auto tm = gsmpl->tm(); |
|
|
| auto * res = &gsmpl->cur_p; |
|
|
| if (do_sort && !res->sorted) { |
| |
| const llama_token id = res->data[res->selected].id; |
|
|
| std::sort(res->data, res->data + res->size, [](const llama_token_data & a, const llama_token_data & b) { |
| return a.p > b.p; |
| }); |
|
|
| |
| for (size_t i = 0; i < res->size; ++i) { |
| if (res->data[i].id == id) { |
| res->selected = i; |
| break; |
| } |
| } |
|
|
| res->sorted = true; |
| } |
|
|
| return res; |
| } |
|
|
| llama_token common_sampler_last(const struct common_sampler * gsmpl) { |
| return gsmpl->prev.rat(0); |
| } |
|
|
| std::string common_sampler_print(const struct common_sampler * gsmpl) { |
| std::string result = "logits "; |
|
|
| for (int i = 0; i < llama_sampler_chain_n(gsmpl->chain); i++) { |
| const auto * smpl = llama_sampler_chain_get(gsmpl->chain, i); |
| result += std::string("-> "); |
| result += std::string(llama_sampler_name(smpl)) + " "; |
| } |
|
|
| return result; |
| } |
|
|
| std::string common_sampler_prev_str(common_sampler * gsmpl, llama_context * ctx_main, int n) { |
| n = std::min(n, (int) gsmpl->prev.size()); |
|
|
| if (n <= 0) { |
| return ""; |
| } |
|
|
| std::string result; |
| result.reserve(8*n); |
|
|
| for (int i = n - 1; i >= 0; i--) { |
| const llama_token id = gsmpl->prev.rat(i); |
|
|
| GGML_ASSERT(id != LLAMA_TOKEN_NULL && "null token in the sampling history - should not happen"); |
|
|
| result += common_token_to_piece(ctx_main, id); |
| } |
|
|
| return result; |
| } |
|
|
| char common_sampler_type_to_chr(enum common_sampler_type cnstr) { |
| switch (cnstr) { |
| case COMMON_SAMPLER_TYPE_DRY: return 'd'; |
| case COMMON_SAMPLER_TYPE_TOP_K: return 'k'; |
| case COMMON_SAMPLER_TYPE_TYPICAL_P: return 'y'; |
| case COMMON_SAMPLER_TYPE_TOP_P: return 'p'; |
| case COMMON_SAMPLER_TYPE_TOP_N_SIGMA: return 's'; |
| case COMMON_SAMPLER_TYPE_MIN_P: return 'm'; |
| case COMMON_SAMPLER_TYPE_TEMPERATURE: return 't'; |
| case COMMON_SAMPLER_TYPE_XTC: return 'x'; |
| case COMMON_SAMPLER_TYPE_INFILL: return 'i'; |
| case COMMON_SAMPLER_TYPE_PENALTIES: return 'e'; |
| case COMMON_SAMPLER_TYPE_ADAPTIVE_P: return 'a'; |
| default : return '?'; |
| } |
| } |
|
|
| std::string common_sampler_type_to_str(enum common_sampler_type cnstr) { |
| switch (cnstr) { |
| case COMMON_SAMPLER_TYPE_DRY: return "dry"; |
| case COMMON_SAMPLER_TYPE_TOP_K: return "top_k"; |
| case COMMON_SAMPLER_TYPE_TYPICAL_P: return "typ_p"; |
| case COMMON_SAMPLER_TYPE_TOP_P: return "top_p"; |
| case COMMON_SAMPLER_TYPE_TOP_N_SIGMA: return "top_n_sigma"; |
| case COMMON_SAMPLER_TYPE_MIN_P: return "min_p"; |
| case COMMON_SAMPLER_TYPE_TEMPERATURE: return "temperature"; |
| case COMMON_SAMPLER_TYPE_XTC: return "xtc"; |
| case COMMON_SAMPLER_TYPE_INFILL: return "infill"; |
| case COMMON_SAMPLER_TYPE_PENALTIES: return "penalties"; |
| case COMMON_SAMPLER_TYPE_ADAPTIVE_P: return "adaptive_p"; |
| default : return ""; |
| } |
| } |
|
|
| std::vector<common_sampler_type> common_sampler_types_from_names(const std::vector<std::string> & names, bool allow_alt_names) { |
| std::unordered_map<std::string, common_sampler_type> sampler_canonical_name_map { |
| { "dry", COMMON_SAMPLER_TYPE_DRY }, |
| { "top_k", COMMON_SAMPLER_TYPE_TOP_K }, |
| { "top_p", COMMON_SAMPLER_TYPE_TOP_P }, |
| { "top_n_sigma", COMMON_SAMPLER_TYPE_TOP_N_SIGMA }, |
| { "typ_p", COMMON_SAMPLER_TYPE_TYPICAL_P }, |
| { "min_p", COMMON_SAMPLER_TYPE_MIN_P }, |
| { "temperature", COMMON_SAMPLER_TYPE_TEMPERATURE }, |
| { "xtc", COMMON_SAMPLER_TYPE_XTC }, |
| { "infill", COMMON_SAMPLER_TYPE_INFILL }, |
| { "penalties", COMMON_SAMPLER_TYPE_PENALTIES }, |
| { "adaptive_p", COMMON_SAMPLER_TYPE_ADAPTIVE_P }, |
| }; |
|
|
| |
| |
| std::unordered_map<std::string, common_sampler_type> sampler_alt_name_map { |
| { "top-k", COMMON_SAMPLER_TYPE_TOP_K }, |
| { "top-p", COMMON_SAMPLER_TYPE_TOP_P }, |
| { "top-n-sigma", COMMON_SAMPLER_TYPE_TOP_N_SIGMA }, |
| { "nucleus", COMMON_SAMPLER_TYPE_TOP_P }, |
| { "typical-p", COMMON_SAMPLER_TYPE_TYPICAL_P }, |
| { "typical", COMMON_SAMPLER_TYPE_TYPICAL_P }, |
| { "typ-p", COMMON_SAMPLER_TYPE_TYPICAL_P }, |
| { "typ", COMMON_SAMPLER_TYPE_TYPICAL_P }, |
| { "min-p", COMMON_SAMPLER_TYPE_MIN_P }, |
| { "temp", COMMON_SAMPLER_TYPE_TEMPERATURE }, |
| { "adaptive-p", COMMON_SAMPLER_TYPE_ADAPTIVE_P }, |
| }; |
|
|
| std::vector<common_sampler_type> samplers; |
| samplers.reserve(names.size()); |
|
|
| for (const auto & name : names) { |
| auto sampler = sampler_canonical_name_map.find(name); |
| if (sampler != sampler_canonical_name_map.end()) { |
| samplers.push_back(sampler->second); |
| continue; |
| } |
| if (allow_alt_names) { |
| sampler = sampler_alt_name_map.find(name); |
| if (sampler != sampler_alt_name_map.end()) { |
| samplers.push_back(sampler->second); |
| continue; |
| } |
| } |
| LOG_WRN("%s: unable to match sampler by name '%s'\n", __func__, name.c_str()); |
| } |
|
|
| return samplers; |
| } |
|
|
| std::vector<common_sampler_type> common_sampler_types_from_chars(const std::string & chars) { |
| std::unordered_map<char, common_sampler_type> sampler_name_map = { |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_DRY), COMMON_SAMPLER_TYPE_DRY }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_TOP_K), COMMON_SAMPLER_TYPE_TOP_K }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_TYPICAL_P), COMMON_SAMPLER_TYPE_TYPICAL_P }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_TOP_P), COMMON_SAMPLER_TYPE_TOP_P }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_TOP_N_SIGMA), COMMON_SAMPLER_TYPE_TOP_N_SIGMA }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_MIN_P), COMMON_SAMPLER_TYPE_MIN_P }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_TEMPERATURE), COMMON_SAMPLER_TYPE_TEMPERATURE }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_XTC), COMMON_SAMPLER_TYPE_XTC }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_INFILL), COMMON_SAMPLER_TYPE_INFILL }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_PENALTIES), COMMON_SAMPLER_TYPE_PENALTIES }, |
| { common_sampler_type_to_chr(COMMON_SAMPLER_TYPE_ADAPTIVE_P), COMMON_SAMPLER_TYPE_ADAPTIVE_P }, |
| }; |
|
|
| std::vector<common_sampler_type> samplers; |
| samplers.reserve(chars.size()); |
|
|
| for (const auto & c : chars) { |
| const auto sampler = sampler_name_map.find(c); |
| if (sampler != sampler_name_map.end()) { |
| samplers.push_back(sampler->second); |
| } else { |
| LOG_WRN("%s: unable to match sampler by char '%c'\n", __func__, c); |
| } |
| } |
|
|
| return samplers; |
| } |
|
|