Apply suggestions from code review
Co-authored-by: slaren <slarengh@gmail.com>
This commit is contained in:
parent
e75d00797e
commit
2dcf37c10b
13 changed files with 48 additions and 48 deletions
|
@ -1,7 +1,7 @@
|
||||||
#include "sampling.h"
|
#include "sampling.h"
|
||||||
|
|
||||||
struct llama_sampling_context * llama_sampling_init(const struct llama_sampling_params & params) {
|
struct llama_sampling_context * llama_sampling_init(const struct llama_sampling_params & params) {
|
||||||
auto result = new llama_sampling_context();
|
auto * result = new llama_sampling_context();
|
||||||
|
|
||||||
result->params = params;
|
result->params = params;
|
||||||
result->grammar = nullptr;
|
result->grammar = nullptr;
|
||||||
|
@ -197,8 +197,8 @@ static llama_token llama_sampling_sample_impl(
|
||||||
}
|
}
|
||||||
|
|
||||||
// apply params.logit_bias map
|
// apply params.logit_bias map
|
||||||
for (const auto & logit_bia : params.logit_bias) {
|
for (const auto & logit_bias : params.logit_bias) {
|
||||||
logits[logit_bia.first] += logit_bia.second;
|
logits[logit_bias.first] += logit_bias.second;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (ctx_cfg) {
|
if (ctx_cfg) {
|
||||||
|
|
|
@ -18,7 +18,7 @@ struct random_uniform_distribution {
|
||||||
};
|
};
|
||||||
|
|
||||||
struct train_state * init_train_state() {
|
struct train_state * init_train_state() {
|
||||||
auto state = new struct train_state;
|
auto * state = new struct train_state;
|
||||||
state->train_its = 0;
|
state->train_its = 0;
|
||||||
state->train_samples = 0;
|
state->train_samples = 0;
|
||||||
state->train_tokens = 0;
|
state->train_tokens = 0;
|
||||||
|
@ -46,12 +46,12 @@ void free_train_state(struct train_state * state) {
|
||||||
struct random_normal_distribution * init_random_normal_distribution(
|
struct random_normal_distribution * init_random_normal_distribution(
|
||||||
int seed, float mean, float std, float min, float max
|
int seed, float mean, float std, float min, float max
|
||||||
) {
|
) {
|
||||||
auto rnd = new random_normal_distribution{std::mt19937(seed), std::normal_distribution<float>{mean, std}, min, max};
|
auto * rnd = new random_normal_distribution{std::mt19937(seed), std::normal_distribution<float>{mean, std}, min, max};
|
||||||
return rnd;
|
return rnd;
|
||||||
}
|
}
|
||||||
|
|
||||||
struct random_uniform_distribution * init_random_uniform_distribution(int seed, float min, float max) {
|
struct random_uniform_distribution * init_random_uniform_distribution(int seed, float min, float max) {
|
||||||
auto rnd = new random_uniform_distribution{std::mt19937(seed), std::uniform_real_distribution<float>{min, max}};
|
auto * rnd = new random_uniform_distribution{std::mt19937(seed), std::uniform_real_distribution<float>{min, max}};
|
||||||
return rnd;
|
return rnd;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1379,7 +1379,7 @@ void finish_processing_train_args(struct train_params_common * params) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void train_opt_callback(void * vdata, int accum_step, float * sched, bool * cancel) {
|
void train_opt_callback(void * vdata, int accum_step, float * sched, bool * cancel) {
|
||||||
auto data = (struct train_opt_callback_data *) vdata;
|
auto * data = (struct train_opt_callback_data *) vdata;
|
||||||
struct train_params_common * params = data->params;
|
struct train_params_common * params = data->params;
|
||||||
struct train_state * train = data->train;
|
struct train_state * train = data->train;
|
||||||
struct ggml_opt_context * opt = train->opt;
|
struct ggml_opt_context * opt = train->opt;
|
||||||
|
|
|
@ -225,7 +225,7 @@ static void free_lora(struct lora_data * lora) {
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct lora_data * load_lora(struct lora_info * info) {
|
static struct lora_data * load_lora(struct lora_info * info) {
|
||||||
auto result = new struct lora_data;
|
auto * result = new struct lora_data;
|
||||||
result->info = *info;
|
result->info = *info;
|
||||||
result->ctx = NULL;
|
result->ctx = NULL;
|
||||||
result->lora_r = 1;
|
result->lora_r = 1;
|
||||||
|
@ -371,7 +371,7 @@ static void export_lora(struct export_lora_params * params) {
|
||||||
// load all loras
|
// load all loras
|
||||||
std::vector<struct lora_data *> loras;
|
std::vector<struct lora_data *> loras;
|
||||||
for (auto & i : params->lora) {
|
for (auto & i : params->lora) {
|
||||||
auto lora = load_lora(&i);
|
auto * lora = load_lora(&i);
|
||||||
if (lora) {
|
if (lora) {
|
||||||
loras.push_back(lora);
|
loras.push_back(lora);
|
||||||
}
|
}
|
||||||
|
@ -455,7 +455,7 @@ static void export_lora(struct export_lora_params * params) {
|
||||||
gguf_free(gguf_in);
|
gguf_free(gguf_in);
|
||||||
|
|
||||||
// free loras
|
// free loras
|
||||||
for (auto& lora : loras) {
|
for (auto * lora : loras) {
|
||||||
free_lora(lora);
|
free_lora(lora);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -801,7 +801,7 @@ static struct ggml_tensor * llama_build_lora_finetune_graphs(
|
||||||
|
|
||||||
// allocating checkpoints in one block to reduce memory fragmentation
|
// allocating checkpoints in one block to reduce memory fragmentation
|
||||||
// note: they will be freed in reverse order
|
// note: they will be freed in reverse order
|
||||||
for (auto& checkpoint : checkpoints) {
|
for (auto * checkpoint : checkpoints) {
|
||||||
if (checkpoint->data == NULL && checkpoint->view_src == NULL) {
|
if (checkpoint->data == NULL && checkpoint->view_src == NULL) {
|
||||||
ggml_allocr_alloc(alloc, checkpoint);
|
ggml_allocr_alloc(alloc, checkpoint);
|
||||||
}
|
}
|
||||||
|
@ -1471,7 +1471,7 @@ struct save_train_files_data {
|
||||||
};
|
};
|
||||||
|
|
||||||
static void save_train_files(void * vdata, struct train_state * train) {
|
static void save_train_files(void * vdata, struct train_state * train) {
|
||||||
auto data = (struct save_train_files_data *) vdata;
|
auto * data = (struct save_train_files_data *) vdata;
|
||||||
|
|
||||||
int64_t iter = train->opt->iter;
|
int64_t iter = train->opt->iter;
|
||||||
|
|
||||||
|
@ -1815,7 +1815,7 @@ int main(int argc, char ** argv) {
|
||||||
++token_noccurs[train_token];
|
++token_noccurs[train_token];
|
||||||
}
|
}
|
||||||
int n_unique_tokens = 0;
|
int n_unique_tokens = 0;
|
||||||
for (unsigned long long token_noccur : token_noccurs) {
|
for (size_t token_noccur : token_noccurs) {
|
||||||
if (token_noccur == 0) continue;
|
if (token_noccur == 0) continue;
|
||||||
++n_unique_tokens;
|
++n_unique_tokens;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1011,21 +1011,21 @@ struct clip_ctx * clip_model_load(const char * fname, const int verbosity = 1) {
|
||||||
vision_model.mm_2_w = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 2, "weight"));
|
vision_model.mm_2_w = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 2, "weight"));
|
||||||
vision_model.mm_2_b = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 2, "bias"));
|
vision_model.mm_2_b = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 2, "bias"));
|
||||||
} catch (std::runtime_error & e) {
|
} catch (std::runtime_error & e) {
|
||||||
static_cast<void>(e);
|
GGML_UNUSED(e);
|
||||||
}
|
}
|
||||||
try {
|
try {
|
||||||
// Yi-type llava
|
// Yi-type llava
|
||||||
vision_model.mm_3_w = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 3, "weight"));
|
vision_model.mm_3_w = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 3, "weight"));
|
||||||
vision_model.mm_3_b = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 3, "bias"));
|
vision_model.mm_3_b = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 3, "bias"));
|
||||||
} catch (std::runtime_error & e) {
|
} catch (std::runtime_error & e) {
|
||||||
static_cast<void>(e);
|
GGML_UNUSED(e);
|
||||||
}
|
}
|
||||||
try {
|
try {
|
||||||
// Yi-type llava
|
// Yi-type llava
|
||||||
vision_model.mm_4_w = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 4, "weight"));
|
vision_model.mm_4_w = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 4, "weight"));
|
||||||
vision_model.mm_4_b = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 4, "bias"));
|
vision_model.mm_4_b = get_tensor(new_clip->ctx_data, format(TN_LLAVA_PROJ, 4, "bias"));
|
||||||
} catch (std::runtime_error & e) {
|
} catch (std::runtime_error & e) {
|
||||||
static_cast<void>(e);
|
GGML_UNUSED(e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else if (new_clip->proj_type == PROJECTOR_TYPE_LDP) {
|
else if (new_clip->proj_type == PROJECTOR_TYPE_LDP) {
|
||||||
|
|
|
@ -451,7 +451,7 @@ static struct ggml_tensor * llama_build_train_graphs(
|
||||||
|
|
||||||
// allocating checkpoints in one block to reduce memory fragmentation
|
// allocating checkpoints in one block to reduce memory fragmentation
|
||||||
// note: they will be freed in reverse order
|
// note: they will be freed in reverse order
|
||||||
for (auto& checkpoint : checkpoints) {
|
for (auto * checkpoint : checkpoints) {
|
||||||
if (checkpoint->data == NULL && checkpoint->view_src == NULL) {
|
if (checkpoint->data == NULL && checkpoint->view_src == NULL) {
|
||||||
ggml_allocr_alloc(alloc, checkpoint);
|
ggml_allocr_alloc(alloc, checkpoint);
|
||||||
}
|
}
|
||||||
|
@ -923,7 +923,7 @@ struct save_train_files_data {
|
||||||
};
|
};
|
||||||
|
|
||||||
static void save_train_files(void * vdata, struct train_state * train) {
|
static void save_train_files(void * vdata, struct train_state * train) {
|
||||||
auto data = (struct save_train_files_data *) vdata;
|
auto * data = (struct save_train_files_data *) vdata;
|
||||||
int64_t iter = train->opt->iter;
|
int64_t iter = train->opt->iter;
|
||||||
|
|
||||||
if (strlen(data->fn_checkpoint_out) > 0) {
|
if (strlen(data->fn_checkpoint_out) > 0) {
|
||||||
|
|
10
llama.cpp
10
llama.cpp
|
@ -7374,7 +7374,7 @@ private:
|
||||||
bpe_encoded_words.reserve(text.size());
|
bpe_encoded_words.reserve(text.size());
|
||||||
|
|
||||||
auto cps = codepoints_from_utf8(text);
|
auto cps = codepoints_from_utf8(text);
|
||||||
for (unsigned int cp : cps)
|
for (uint32_t cp : cps)
|
||||||
text_utf.emplace_back(codepoint_to_utf8(cp));
|
text_utf.emplace_back(codepoint_to_utf8(cp));
|
||||||
|
|
||||||
for (int i = 0; i < (int)text_utf.size(); i++) {
|
for (int i = 0; i < (int)text_utf.size(); i++) {
|
||||||
|
@ -8089,7 +8089,7 @@ void llama_grammar_free(struct llama_grammar * grammar) {
|
||||||
}
|
}
|
||||||
|
|
||||||
struct llama_grammar * llama_grammar_copy(const struct llama_grammar * grammar) {
|
struct llama_grammar * llama_grammar_copy(const struct llama_grammar * grammar) {
|
||||||
auto result = new llama_grammar{ grammar->rules, grammar->stacks, grammar->partial_utf8 };
|
auto * result = new llama_grammar{ grammar->rules, grammar->stacks, grammar->partial_utf8 };
|
||||||
|
|
||||||
// redirect elements in stacks to point to new rules
|
// redirect elements in stacks to point to new rules
|
||||||
for (size_t is = 0; is < result->stacks.size(); is++) {
|
for (size_t is = 0; is < result->stacks.size(); is++) {
|
||||||
|
@ -9654,7 +9654,7 @@ static void llama_model_quantize_internal(const std::string & fname_inp, const s
|
||||||
|
|
||||||
if (tot_count > 0) {
|
if (tot_count > 0) {
|
||||||
LLAMA_LOG_INFO(" | hist: ");
|
LLAMA_LOG_INFO(" | hist: ");
|
||||||
for (long long i : hist_cur) {
|
for (int64_t i : hist_cur) {
|
||||||
LLAMA_LOG_INFO("%5.3f ", i / float(nelements));
|
LLAMA_LOG_INFO("%5.3f ", i / float(nelements));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -10101,7 +10101,7 @@ struct llama_model * llama_load_model_from_file(
|
||||||
struct llama_model_params params) {
|
struct llama_model_params params) {
|
||||||
ggml_time_init();
|
ggml_time_init();
|
||||||
|
|
||||||
auto model = new llama_model;
|
auto * model = new llama_model;
|
||||||
|
|
||||||
unsigned cur_percentage = 0;
|
unsigned cur_percentage = 0;
|
||||||
if (params.progress_callback == NULL) {
|
if (params.progress_callback == NULL) {
|
||||||
|
@ -10147,7 +10147,7 @@ struct llama_context * llama_new_context_with_model(
|
||||||
return nullptr;
|
return nullptr;
|
||||||
}
|
}
|
||||||
|
|
||||||
auto ctx = new llama_context(*model);
|
auto * ctx = new llama_context(*model);
|
||||||
|
|
||||||
const auto & hparams = model->hparams;
|
const auto & hparams = model->hparams;
|
||||||
auto & cparams = ctx->cparams;
|
auto & cparams = ctx->cparams;
|
||||||
|
|
|
@ -64,7 +64,7 @@ static void init_tensor_uniform(ggml_tensor * tensor, float min = -1.0f, float m
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
ggml_quantize_chunk(tensor->type, data.data(), dataq.data(), 0, int(size/tensor->ne[0]),
|
ggml_quantize_chunk(tensor->type, data.data(), dataq.data(), 0, int(size/tensor->ne[0]),
|
||||||
static_cast<int>(tensor->ne[0]), hist, im);
|
int(tensor->ne[0]), hist, im);
|
||||||
ggml_backend_tensor_set(tensor, dataq.data(), 0, dataq.size());
|
ggml_backend_tensor_set(tensor, dataq.data(), 0, dataq.size());
|
||||||
} else if (tensor->type == GGML_TYPE_I8 || tensor->type == GGML_TYPE_I16 || tensor->type == GGML_TYPE_I32) {
|
} else if (tensor->type == GGML_TYPE_I8 || tensor->type == GGML_TYPE_I16 || tensor->type == GGML_TYPE_I32) {
|
||||||
// This is going to create some weird integers though.
|
// This is going to create some weird integers though.
|
||||||
|
@ -288,9 +288,9 @@ struct test_case {
|
||||||
virtual size_t op_size(ggml_tensor * t) {
|
virtual size_t op_size(ggml_tensor * t) {
|
||||||
size_t size = ggml_nbytes(t);
|
size_t size = ggml_nbytes(t);
|
||||||
// add source tensors
|
// add source tensors
|
||||||
for (auto& el : t->src) {
|
for (auto * src : t->src) {
|
||||||
if (el) {
|
if (src) {
|
||||||
size += ggml_nbytes(el);
|
size += ggml_nbytes(src);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return size;
|
return size;
|
||||||
|
@ -423,7 +423,7 @@ struct test_case {
|
||||||
};
|
};
|
||||||
|
|
||||||
auto callback = [](int index, ggml_tensor * t1, ggml_tensor * t2, void * user_data) -> bool {
|
auto callback = [](int index, ggml_tensor * t1, ggml_tensor * t2, void * user_data) -> bool {
|
||||||
auto ud = (callback_userdata *) user_data;
|
auto * ud = (callback_userdata *) user_data;
|
||||||
const char * bn1 = ggml_backend_name(ud->backend1);
|
const char * bn1 = ggml_backend_name(ud->backend1);
|
||||||
const char * bn2 = ggml_backend_name(ud->backend2);
|
const char * bn2 = ggml_backend_name(ud->backend2);
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue