Merge branch 'master' into sycl
This commit is contained in:
commit
238ec31aeb
11 changed files with 190 additions and 100 deletions
26
.devops/main-intel.Dockerfile
Normal file
26
.devops/main-intel.Dockerfile
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
ARG ONEAPI_VERSION=2024.0.1-devel-ubuntu22.04
|
||||||
|
ARG UBUNTU_VERSION=22.04
|
||||||
|
|
||||||
|
FROM intel/hpckit:$ONEAPI_VERSION as build
|
||||||
|
|
||||||
|
RUN apt-get update && \
|
||||||
|
apt-get install -y git
|
||||||
|
|
||||||
|
WORKDIR /app
|
||||||
|
|
||||||
|
COPY . .
|
||||||
|
|
||||||
|
# for some reasons, "-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=Intel10_64lp -DLLAMA_NATIVE=ON" give worse performance
|
||||||
|
RUN mkdir build && \
|
||||||
|
cd build && \
|
||||||
|
cmake .. -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx && \
|
||||||
|
cmake --build . --config Release --target main server
|
||||||
|
|
||||||
|
FROM ubuntu:$UBUNTU_VERSION as runtime
|
||||||
|
|
||||||
|
COPY --from=build /app/build/bin/main /main
|
||||||
|
COPY --from=build /app/build/bin/server /server
|
||||||
|
|
||||||
|
ENV LC_ALL=C.utf8
|
||||||
|
|
||||||
|
ENTRYPOINT [ "/main" ]
|
1
.github/workflows/docker.yml
vendored
1
.github/workflows/docker.yml
vendored
|
@ -35,6 +35,7 @@ jobs:
|
||||||
- { tag: "full-cuda", dockerfile: ".devops/full-cuda.Dockerfile", platforms: "linux/amd64" }
|
- { tag: "full-cuda", dockerfile: ".devops/full-cuda.Dockerfile", platforms: "linux/amd64" }
|
||||||
- { tag: "light-rocm", dockerfile: ".devops/main-rocm.Dockerfile", platforms: "linux/amd64,linux/arm64" }
|
- { tag: "light-rocm", dockerfile: ".devops/main-rocm.Dockerfile", platforms: "linux/amd64,linux/arm64" }
|
||||||
- { tag: "full-rocm", dockerfile: ".devops/full-rocm.Dockerfile", platforms: "linux/amd64,linux/arm64" }
|
- { tag: "full-rocm", dockerfile: ".devops/full-rocm.Dockerfile", platforms: "linux/amd64,linux/arm64" }
|
||||||
|
- { tag: "light-intel", dockerfile: ".devops/main-intel.Dockerfile", platforms: "linux/amd64" }
|
||||||
steps:
|
steps:
|
||||||
- name: Check out the repo
|
- name: Check out the repo
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
|
|
|
@ -220,12 +220,10 @@ bool gpt_params_parse_ex(int argc, char ** argv, gpt_params & params) {
|
||||||
}
|
}
|
||||||
// store the external file name in params
|
// store the external file name in params
|
||||||
params.prompt_file = argv[i];
|
params.prompt_file = argv[i];
|
||||||
file.seekg(0, std::ios::end);
|
std::ostringstream ss;
|
||||||
size_t size = file.tellg();
|
ss << file.rdbuf();
|
||||||
file.seekg(0, std::ios::beg);
|
params.prompt = ss.str();
|
||||||
params.prompt.resize(size);
|
fprintf(stderr, "Read %zu bytes from binary file %s\n", params.prompt.size(), argv[i]);
|
||||||
file.read((char *)params.prompt.data(), size);
|
|
||||||
fprintf(stderr, "Read %zu bytes from binary file %s\n", size, argv[i]);
|
|
||||||
} else if (arg == "-f" || arg == "--file") {
|
} else if (arg == "-f" || arg == "--file") {
|
||||||
if (++i >= argc) {
|
if (++i >= argc) {
|
||||||
invalid_param = true;
|
invalid_param = true;
|
||||||
|
|
|
@ -6,7 +6,7 @@
|
||||||
" Similarly, you could add an insert mode keybind with
|
" Similarly, you could add an insert mode keybind with
|
||||||
" inoremap <C-B> <Cmd>call llama#doLlamaGen()<CR>
|
" inoremap <C-B> <Cmd>call llama#doLlamaGen()<CR>
|
||||||
"
|
"
|
||||||
" g:llama_api_url and g:llama_overrides can be configured in your .vimrc
|
" g:llama_api_url, g:llama_api_key and g:llama_overrides can be configured in your .vimrc
|
||||||
" let g:llama_api_url = "192.168.1.10:8080"
|
" let g:llama_api_url = "192.168.1.10:8080"
|
||||||
" llama_overrides can also be set through buffer/window scopes. For instance
|
" llama_overrides can also be set through buffer/window scopes. For instance
|
||||||
" autocmd filetype python let b:llama_overrides = {"temp": 0.2}
|
" autocmd filetype python let b:llama_overrides = {"temp": 0.2}
|
||||||
|
@ -82,6 +82,9 @@ func llama#doLlamaGen()
|
||||||
endif
|
endif
|
||||||
let l:querydata.prompt = join(l:buflines, "\n")
|
let l:querydata.prompt = join(l:buflines, "\n")
|
||||||
let l:curlcommand = copy(s:curlcommand)
|
let l:curlcommand = copy(s:curlcommand)
|
||||||
|
if exists("g:llama_api_key")
|
||||||
|
call extend(l:curlcommand, ['--header', 'Authorization: Bearer ' .. g:llama_api_key])
|
||||||
|
endif
|
||||||
let l:curlcommand[2] = json_encode(l:querydata)
|
let l:curlcommand[2] = json_encode(l:querydata)
|
||||||
let b:job = job_start(l:curlcommand, {"callback": function("s:callbackHandler", [l:cbuffer])})
|
let b:job = job_start(l:curlcommand, {"callback": function("s:callbackHandler", [l:cbuffer])})
|
||||||
endfunction
|
endfunction
|
||||||
|
|
|
@ -2,18 +2,6 @@
|
||||||
// so there might be still unnecessary artifacts hanging around
|
// so there might be still unnecessary artifacts hanging around
|
||||||
// I'll gradually clean and extend it
|
// I'll gradually clean and extend it
|
||||||
|
|
||||||
#include <cassert>
|
|
||||||
#include <cmath>
|
|
||||||
#include <cstdlib>
|
|
||||||
#include <cstring>
|
|
||||||
#include <fstream>
|
|
||||||
#include <iostream>
|
|
||||||
#include <map>
|
|
||||||
#include <regex>
|
|
||||||
#include <stdexcept>
|
|
||||||
#include <vector>
|
|
||||||
#include <sstream>
|
|
||||||
|
|
||||||
#include "clip.h"
|
#include "clip.h"
|
||||||
#include "ggml.h"
|
#include "ggml.h"
|
||||||
#include "ggml-alloc.h"
|
#include "ggml-alloc.h"
|
||||||
|
@ -30,6 +18,19 @@
|
||||||
#define STB_IMAGE_IMPLEMENTATION
|
#define STB_IMAGE_IMPLEMENTATION
|
||||||
#include "stb_image.h"
|
#include "stb_image.h"
|
||||||
|
|
||||||
|
#include <cassert>
|
||||||
|
#include <cmath>
|
||||||
|
#include <cstdlib>
|
||||||
|
#include <cstring>
|
||||||
|
#include <fstream>
|
||||||
|
#include <iostream>
|
||||||
|
#include <map>
|
||||||
|
#include <regex>
|
||||||
|
#include <stdexcept>
|
||||||
|
#include <vector>
|
||||||
|
#include <sstream>
|
||||||
|
#include <cinttypes>
|
||||||
|
|
||||||
static std::string format(const char * fmt, ...) {
|
static std::string format(const char * fmt, ...) {
|
||||||
va_list ap;
|
va_list ap;
|
||||||
va_list ap2;
|
va_list ap2;
|
||||||
|
@ -217,9 +218,9 @@ static std::string gguf_kv_to_str(const struct gguf_context * ctx_gguf, int i) {
|
||||||
|
|
||||||
static void print_tensor_info(const ggml_tensor* tensor, const char* prefix = "") {
|
static void print_tensor_info(const ggml_tensor* tensor, const char* prefix = "") {
|
||||||
size_t tensor_size = ggml_nbytes(tensor);
|
size_t tensor_size = ggml_nbytes(tensor);
|
||||||
printf("%s: n_dims = %d, name = %s, tensor_size=%zu, shape:[%d, %d, %d, %d], type: %d\n",
|
printf("%s: n_dims = %d, name = %s, tensor_size=%zu, shape:[%" PRId64 ", %" PRId64 ", %" PRId64 ", %" PRId64 "], type = %s\n",
|
||||||
prefix, ggml_n_dims(tensor), tensor->name, tensor_size,
|
prefix, ggml_n_dims(tensor), tensor->name, tensor_size,
|
||||||
tensor->ne[0], tensor->ne[1], tensor->ne[2], tensor->ne[3], tensor->type);
|
tensor->ne[0], tensor->ne[1], tensor->ne[2], tensor->ne[3], ggml_type_name(tensor->type));
|
||||||
}
|
}
|
||||||
|
|
||||||
static projector_type clip_projector_type_from_string(const std::string & name) {
|
static projector_type clip_projector_type_from_string(const std::string & name) {
|
||||||
|
@ -592,7 +593,7 @@ static ggml_cgraph * clip_image_build_graph(clip_ctx * ctx, const clip_image_f32
|
||||||
mlp_3 = ggml_cont(ctx0, ggml_permute(ctx0, mlp_3, 1, 0, 2, 3));
|
mlp_3 = ggml_cont(ctx0, ggml_permute(ctx0, mlp_3, 1, 0, 2, 3));
|
||||||
mlp_3 = ggml_reshape_4d(ctx0, mlp_3, n_patch, n_patch, mlp_3->ne[1], mlp_3->ne[2]);
|
mlp_3 = ggml_reshape_4d(ctx0, mlp_3, n_patch, n_patch, mlp_3->ne[1], mlp_3->ne[2]);
|
||||||
// stride = 1, padding = 1, bias is nullptr
|
// stride = 1, padding = 1, bias is nullptr
|
||||||
block_1 = ggml_conv_depthwise_2d(ctx0, model.mm_model_block_1_block_0_0_w, mlp_3, nullptr, 1, 1, 1, 1, 1, 1);
|
block_1 = ggml_conv_depthwise_2d(ctx0, model.mm_model_block_1_block_0_0_w, mlp_3, 1, 1, 1, 1, 1, 1);
|
||||||
|
|
||||||
// layer norm
|
// layer norm
|
||||||
// // block_1 shape = [1, 2048, 24, 24], ne = [24, 24, 2048, 1]
|
// // block_1 shape = [1, 2048, 24, 24], ne = [24, 24, 2048, 1]
|
||||||
|
@ -640,7 +641,7 @@ static ggml_cgraph * clip_image_build_graph(clip_ctx * ctx, const clip_image_f32
|
||||||
// block_2
|
// block_2
|
||||||
{
|
{
|
||||||
// stride = 2
|
// stride = 2
|
||||||
block_1 = ggml_conv_depthwise_2d(ctx0, model.mm_model_block_2_block_0_0_w, block_1, nullptr, 2, 2, 1, 1, 1, 1);
|
block_1 = ggml_conv_depthwise_2d(ctx0, model.mm_model_block_2_block_0_0_w, block_1, 2, 2, 1, 1, 1, 1);
|
||||||
|
|
||||||
// block_1 shape = [1, 2048, 12, 12], ne = [12, 12, 2048, 1]
|
// block_1 shape = [1, 2048, 12, 12], ne = [12, 12, 2048, 1]
|
||||||
// layer norm
|
// layer norm
|
||||||
|
@ -741,18 +742,10 @@ struct clip_ctx * clip_model_load(const char * fname, const int verbosity = 1) {
|
||||||
{
|
{
|
||||||
std::map<enum ggml_type, uint32_t> n_type;
|
std::map<enum ggml_type, uint32_t> n_type;
|
||||||
|
|
||||||
uint32_t n_type_max = 0;
|
|
||||||
enum ggml_type type_max = GGML_TYPE_F32;
|
|
||||||
|
|
||||||
for (int i = 0; i < n_tensors; i++) {
|
for (int i = 0; i < n_tensors; i++) {
|
||||||
enum ggml_type type = gguf_get_tensor_type(ctx, i);
|
enum ggml_type type = gguf_get_tensor_type(ctx, i);
|
||||||
|
|
||||||
n_type[type]++;
|
n_type[type]++;
|
||||||
|
|
||||||
if (n_type_max < n_type[type]) {
|
|
||||||
n_type_max = n_type[type];
|
|
||||||
type_max = type;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
printf("%s: Dumping metadata keys/values. Note: KV overrides do not apply in this output.\n", __func__);
|
printf("%s: Dumping metadata keys/values. Note: KV overrides do not apply in this output.\n", __func__);
|
||||||
|
@ -795,14 +788,12 @@ struct clip_ctx * clip_model_load(const char * fname, const int verbosity = 1) {
|
||||||
size_t tensor_size = ggml_nbytes(cur);
|
size_t tensor_size = ggml_nbytes(cur);
|
||||||
buffer_size += tensor_size;
|
buffer_size += tensor_size;
|
||||||
if (verbosity >= 3) {
|
if (verbosity >= 3) {
|
||||||
printf("%s: tensor[%d]: n_dims = %d, name = %s, tensor_size=%zu, offset=%zu, shape:[%d, %d, %d, %d], type: %d\n", __func__, i,
|
printf("%s: tensor[%d]: n_dims = %d, name = %s, tensor_size=%zu, offset=%zu, shape:[%" PRIu64 ", %" PRIu64 ", %" PRIu64 ", %" PRIu64 "], type = %s\n",
|
||||||
ggml_n_dims(cur), cur->name, tensor_size, offset, cur->ne[0], cur->ne[1], cur->ne[2], cur->ne[3], type);
|
__func__, i, ggml_n_dims(cur), cur->name, tensor_size, offset, cur->ne[0], cur->ne[1], cur->ne[2], cur->ne[3], ggml_type_name(type));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
buffer_size += n_tensors * 128 /* CLIP PADDING */;
|
buffer_size += n_tensors * 128 /* CLIP PADDING */;
|
||||||
|
|
||||||
clip_ctx * new_clip = new clip_ctx;
|
clip_ctx * new_clip = new clip_ctx;
|
||||||
|
|
|
@ -222,13 +222,18 @@ struct kl_divergence_result {
|
||||||
double sum_kld2 = 0;
|
double sum_kld2 = 0;
|
||||||
double sum_nll_diff = 0;
|
double sum_nll_diff = 0;
|
||||||
double sum_nll_diff2 = 0;
|
double sum_nll_diff2 = 0;
|
||||||
|
size_t n_same_top = 0;
|
||||||
size_t count = 0;
|
size_t count = 0;
|
||||||
};
|
};
|
||||||
|
|
||||||
static void log_softmax(int n_vocab, const float * logits, const uint16_t * base_log_prob, int tok, kl_divergence_result & kld) {
|
static double log_softmax(int n_vocab, const float * logits, const uint16_t * base_log_prob, int tok, kl_divergence_result & kld) {
|
||||||
float max_logit = logits[0];
|
float max_logit = logits[0];
|
||||||
|
int imax = 0;
|
||||||
for (int i = 1; i < n_vocab; ++i) {
|
for (int i = 1; i < n_vocab; ++i) {
|
||||||
max_logit = std::max(max_logit, logits[i]);
|
if (logits[i] > max_logit) {
|
||||||
|
max_logit = logits[i];
|
||||||
|
imax = i;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
double sum_exp = 0.0;
|
double sum_exp = 0.0;
|
||||||
for (int i = 0; i < n_vocab; ++i) {
|
for (int i = 0; i < n_vocab; ++i) {
|
||||||
|
@ -247,8 +252,14 @@ static void log_softmax(int n_vocab, const float * logits, const uint16_t * base
|
||||||
kld.sum_nll_diff2 += nll*nll;
|
kld.sum_nll_diff2 += nll*nll;
|
||||||
max_logit += log_sum_exp;
|
max_logit += log_sum_exp;
|
||||||
double sum = 0;
|
double sum = 0;
|
||||||
|
int imax_base = -1;
|
||||||
|
float p_log_base_max = 0;
|
||||||
for (int i = 0; i < n_vocab; ++i) {
|
for (int i = 0; i < n_vocab; ++i) {
|
||||||
const float p_log_base = scale*base_log_prob[i] + min_log_prob;
|
const float p_log_base = scale*base_log_prob[i] + min_log_prob;
|
||||||
|
if (i == 0 || p_log_base > p_log_base_max) {
|
||||||
|
p_log_base_max = p_log_base;
|
||||||
|
imax_base = i;
|
||||||
|
}
|
||||||
if (p_log_base > -16.f) {
|
if (p_log_base > -16.f) {
|
||||||
const float p_base = expf(p_log_base);
|
const float p_base = expf(p_log_base);
|
||||||
sum += p_base * (p_log_base - logits[i] + max_logit);
|
sum += p_base * (p_log_base - logits[i] + max_logit);
|
||||||
|
@ -257,14 +268,17 @@ static void log_softmax(int n_vocab, const float * logits, const uint16_t * base
|
||||||
kld.sum_kld += sum;
|
kld.sum_kld += sum;
|
||||||
kld.sum_kld2 += sum*sum;
|
kld.sum_kld2 += sum*sum;
|
||||||
++kld.count;
|
++kld.count;
|
||||||
|
if (imax == imax_base) ++kld.n_same_top;
|
||||||
|
return sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void process_logits(int n_vocab, const float * logits, const int * tokens, int n_token,
|
static void process_logits(int n_vocab, const float * logits, const int * tokens, int n_token,
|
||||||
std::vector<std::thread> & workers, const std::vector<uint16_t> & base_log_probs, kl_divergence_result & kld) {
|
std::vector<std::thread> & workers, const std::vector<uint16_t> & base_log_probs, kl_divergence_result & kld,
|
||||||
|
float * kld_values) {
|
||||||
std::mutex mutex;
|
std::mutex mutex;
|
||||||
const int nv = 2*((n_vocab + 1)/2) + 4;
|
const int nv = 2*((n_vocab + 1)/2) + 4;
|
||||||
int counter = 0;
|
int counter = 0;
|
||||||
auto compute = [&mutex, &counter, &base_log_probs, &kld, n_vocab, logits, tokens, n_token, nv] () {
|
auto compute = [&mutex, &counter, &base_log_probs, &kld, n_vocab, logits, tokens, n_token, nv, kld_values] () {
|
||||||
kl_divergence_result local_kld;
|
kl_divergence_result local_kld;
|
||||||
while (true) {
|
while (true) {
|
||||||
std::unique_lock<std::mutex> lock(mutex);
|
std::unique_lock<std::mutex> lock(mutex);
|
||||||
|
@ -276,11 +290,13 @@ static void process_logits(int n_vocab, const float * logits, const int * tokens
|
||||||
kld.sum_kld2 += local_kld.sum_kld2;
|
kld.sum_kld2 += local_kld.sum_kld2;
|
||||||
kld.sum_nll_diff += local_kld.sum_nll_diff;
|
kld.sum_nll_diff += local_kld.sum_nll_diff;
|
||||||
kld.sum_nll_diff2 += local_kld.sum_nll_diff2;
|
kld.sum_nll_diff2 += local_kld.sum_nll_diff2;
|
||||||
|
kld.n_same_top += local_kld.n_same_top;
|
||||||
kld.count += local_kld.count;
|
kld.count += local_kld.count;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
lock.unlock();
|
lock.unlock();
|
||||||
log_softmax(n_vocab, logits + i*n_vocab, base_log_probs.data() + i*nv, tokens[i+1], local_kld);
|
double v = log_softmax(n_vocab, logits + i*n_vocab, base_log_probs.data() + i*nv, tokens[i+1], local_kld);
|
||||||
|
kld_values[i] = (float)v;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
for (auto & w : workers) {
|
for (auto & w : workers) {
|
||||||
|
@ -1202,11 +1218,11 @@ static void winogrande_score(llama_context * ctx, const gpt_params & params) {
|
||||||
printf("Final Winogrande score(%d tasks): %.4lf +/- %.4lf\n", n_done, 100*p, sigma);
|
printf("Final Winogrande score(%d tasks): %.4lf +/- %.4lf\n", n_done, 100*p, sigma);
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool deserialize_string(std::istream& in, std::string& str) {
|
static bool deserialize_string(std::istream & in, std::string & str) {
|
||||||
uint32_t size;
|
uint32_t size;
|
||||||
if (!in.read((char *)&size, sizeof(size)).fail()) {
|
if (!in.read((char *)&size, sizeof(size)).fail()) {
|
||||||
str.resize(size);
|
str.resize(size);
|
||||||
if (!in.read((char *)str.data(), size).fail()) return true;
|
if (!in.read((char *)&str[0], size).fail()) return true;
|
||||||
}
|
}
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -1615,7 +1631,7 @@ static void kl_divergence(llama_context * ctx, const gpt_params & params) {
|
||||||
in.read((char *)&n_vocab, sizeof(n_vocab));
|
in.read((char *)&n_vocab, sizeof(n_vocab));
|
||||||
in.read((char *)&n_chunk, sizeof(n_chunk));
|
in.read((char *)&n_chunk, sizeof(n_chunk));
|
||||||
if (in.fail()) {
|
if (in.fail()) {
|
||||||
fprintf(stderr, "%s: failed rwading n_vocab, n_chunk from %s\n", __func__, params.logits_file.c_str());
|
fprintf(stderr, "%s: failed reading n_vocab, n_chunk from %s\n", __func__, params.logits_file.c_str());
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
if (n_vocab != llama_n_vocab(llama_get_model(ctx))) {
|
if (n_vocab != llama_n_vocab(llama_get_model(ctx))) {
|
||||||
|
@ -1634,6 +1650,7 @@ static void kl_divergence(llama_context * ctx, const gpt_params & params) {
|
||||||
const bool add_bos = llama_should_add_bos_token(llama_get_model(ctx));
|
const bool add_bos = llama_should_add_bos_token(llama_get_model(ctx));
|
||||||
|
|
||||||
std::vector<uint16_t> log_probs_uint16(size_t(n_ctx - 1 - n_ctx/2) * nv);
|
std::vector<uint16_t> log_probs_uint16(size_t(n_ctx - 1 - n_ctx/2) * nv);
|
||||||
|
std::vector<float> kld_values(size_t(n_ctx - 1 - n_ctx/2)*n_chunk);
|
||||||
std::vector<float> logits;
|
std::vector<float> logits;
|
||||||
if (num_batches > 1) {
|
if (num_batches > 1) {
|
||||||
logits.reserve(n_ctx * n_vocab);
|
logits.reserve(n_ctx * n_vocab);
|
||||||
|
@ -1652,6 +1669,7 @@ static void kl_divergence(llama_context * ctx, const gpt_params & params) {
|
||||||
};
|
};
|
||||||
|
|
||||||
kl_divergence_result kld;
|
kl_divergence_result kld;
|
||||||
|
auto kld_ptr = kld_values.data();
|
||||||
|
|
||||||
for (int i = 0; i < n_chunk; ++i) {
|
for (int i = 0; i < n_chunk; ++i) {
|
||||||
const int start = i * n_ctx;
|
const int start = i * n_ctx;
|
||||||
|
@ -1705,20 +1723,24 @@ static void kl_divergence(llama_context * ctx, const gpt_params & params) {
|
||||||
}
|
}
|
||||||
fprintf(stderr, "%.2f minutes\n", total_seconds / 60.0);
|
fprintf(stderr, "%.2f minutes\n", total_seconds / 60.0);
|
||||||
|
|
||||||
printf("\nchunk PPL ln(PPL(Q)/PPL(base)) KL-Divergence\n");
|
printf("\nchunk PPL ln(PPL(Q)/PPL(base)) KL-Divergence Same top\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
const int first = n_ctx/2;
|
const int first = n_ctx/2;
|
||||||
const float * all_logits = num_batches > 1 ? logits.data() : llama_get_logits(ctx);
|
const float * all_logits = num_batches > 1 ? logits.data() : llama_get_logits(ctx);
|
||||||
process_logits(n_vocab, all_logits + first*n_vocab, tokens.data() + start + first, n_ctx - 1 - first,
|
process_logits(n_vocab, all_logits + first*n_vocab, tokens.data() + start + first, n_ctx - 1 - first,
|
||||||
workers, log_probs_uint16, kld);
|
workers, log_probs_uint16, kld, kld_ptr);
|
||||||
|
kld_ptr += n_ctx - 1 - first;
|
||||||
|
|
||||||
auto ppl = mean_and_uncertainty(kld.sum_nll, kld.sum_nll2, kld.count);
|
auto ppl = mean_and_uncertainty(kld.sum_nll, kld.sum_nll2, kld.count);
|
||||||
auto log_ppl_ratio = mean_and_uncertainty(kld.sum_nll_diff, kld.sum_nll_diff2, kld.count);
|
auto log_ppl_ratio = mean_and_uncertainty(kld.sum_nll_diff, kld.sum_nll_diff2, kld.count);
|
||||||
auto kl_div = mean_and_uncertainty(kld.sum_kld, kld.sum_kld2, kld.count);
|
auto kl_div = mean_and_uncertainty(kld.sum_kld, kld.sum_kld2, kld.count);
|
||||||
|
auto p_top = 1.*kld.n_same_top/kld.count;
|
||||||
|
auto d_p_top = sqrt(p_top*(1 - p_top)/(kld.count - 1));
|
||||||
|
|
||||||
printf("%4d %10.4lf %10.5lf ± %10.5f %10.5f ± %10.5lf\n", i+1, exp(ppl.first),
|
printf("%4d %10.4lf %10.5lf ± %10.5f %10.5f ± %10.5lf %.5f ± %.5f\n", i+1, exp(ppl.first),
|
||||||
log_ppl_ratio.first, log_ppl_ratio.second, kl_div.first, kl_div.second);
|
log_ppl_ratio.first, log_ppl_ratio.second, kl_div.first, kl_div.second,
|
||||||
|
p_top, d_p_top);
|
||||||
|
|
||||||
fflush(stdout);
|
fflush(stdout);
|
||||||
|
|
||||||
|
@ -1726,6 +1748,35 @@ static void kl_divergence(llama_context * ctx, const gpt_params & params) {
|
||||||
}
|
}
|
||||||
printf("\n");
|
printf("\n");
|
||||||
|
|
||||||
|
if (kld.count < 100) return; // we do not wish to do statistics on so few values
|
||||||
|
|
||||||
|
std::sort(kld_values.begin(), kld_values.end());
|
||||||
|
|
||||||
|
printf("===== KL-divergence statistics\n");
|
||||||
|
auto kl_div = mean_and_uncertainty(kld.sum_kld, kld.sum_kld2, kld.count);
|
||||||
|
printf("Average: %10.6f ±%10.6lf\n", kl_div.first, kl_div.second);
|
||||||
|
auto kld_median = kld_values.size()%2 == 0 ? 0.5f*(kld_values[kld_values.size()/2] + kld_values[kld_values.size()/2-1])
|
||||||
|
: kld_values[kld_values.size()/2];
|
||||||
|
printf("Median : %10.6f\n", kld_median);
|
||||||
|
|
||||||
|
auto percentile = [&kld_values] (float fraction) {
|
||||||
|
if (fraction <= 0) return kld_values.front();
|
||||||
|
if (fraction >= 1) return kld_values.back();
|
||||||
|
float p = fraction*(kld_values.size() - 1);
|
||||||
|
size_t ip = size_t(p); p -= ip;
|
||||||
|
return (1 - p)*kld_values[ip] + p*kld_values[std::min(ip+1, kld_values.size()-1)];
|
||||||
|
};
|
||||||
|
|
||||||
|
printf("Maximum: %10.6f\n", kld_values.back());
|
||||||
|
printf("KLD_99 : %10.6f\n", percentile(0.99f));
|
||||||
|
printf("KLD_95 : %10.6f\n", percentile(0.95f));
|
||||||
|
printf("KLD_90 : %10.6f\n", percentile(0.90f));
|
||||||
|
|
||||||
|
printf("Minimum: %10.6f\n", kld_values.front());
|
||||||
|
printf("KLD_01 : %10.6f\n", percentile(0.01f));
|
||||||
|
printf("KLD_05 : %10.6f\n", percentile(0.05f));
|
||||||
|
printf("KLD_10 : %10.6f\n", percentile(0.10f));
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
int main(int argc, char ** argv) {
|
int main(int argc, char ** argv) {
|
||||||
|
|
89
ggml-cuda.cu
89
ggml-cuda.cu
|
@ -13,6 +13,10 @@
|
||||||
#include <map>
|
#include <map>
|
||||||
#include <array>
|
#include <array>
|
||||||
|
|
||||||
|
// stringize macro for converting __CUDA_ARCH_LIST__ (list of integers) to string
|
||||||
|
#define STRINGIZE_IMPL(...) #__VA_ARGS__
|
||||||
|
#define STRINGIZE(...) STRINGIZE_IMPL(__VA_ARGS__)
|
||||||
|
|
||||||
#if defined(GGML_USE_HIPBLAS)
|
#if defined(GGML_USE_HIPBLAS)
|
||||||
#include <hip/hip_runtime.h>
|
#include <hip/hip_runtime.h>
|
||||||
#include <hipblas/hipblas.h>
|
#include <hipblas/hipblas.h>
|
||||||
|
@ -584,13 +588,28 @@ static cuda_device_capabilities g_device_caps[GGML_CUDA_MAX_DEVICES] = { {0, 0,
|
||||||
static cublasHandle_t g_cublas_handles[GGML_CUDA_MAX_DEVICES] = {nullptr};
|
static cublasHandle_t g_cublas_handles[GGML_CUDA_MAX_DEVICES] = {nullptr};
|
||||||
|
|
||||||
[[noreturn]]
|
[[noreturn]]
|
||||||
static __device__ void bad_arch() {
|
static __device__ void no_device_code(
|
||||||
printf("ERROR: ggml-cuda was compiled without support for the current GPU architecture.\n");
|
const char * file_name, const int line, const char * function_name, const int arch, const char * arch_list) {
|
||||||
|
|
||||||
|
#if defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)
|
||||||
|
printf("%s:%d: ERROR: HIP kernel %s has no device code compatible with HIP arch %d.\n",
|
||||||
|
file_name, line, function_name, arch);
|
||||||
|
(void) arch_list;
|
||||||
|
#else
|
||||||
|
printf("%s:%d: ERROR: CUDA kernel %s has no device code compatible with CUDA arch %d. ggml-cuda.cu was compiled for: %s\n",
|
||||||
|
file_name, line, function_name, arch, arch_list);
|
||||||
|
#endif // defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)
|
||||||
__trap();
|
__trap();
|
||||||
|
|
||||||
(void) bad_arch; // suppress unused function warning
|
(void) no_device_code; // suppress unused function warning
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#ifdef __CUDA_ARCH__
|
||||||
|
#define NO_DEVICE_CODE no_device_code(__FILE__, __LINE__, __FUNCTION__, __CUDA_ARCH__, STRINGIZE(__CUDA_ARCH_LIST__))
|
||||||
|
#else
|
||||||
|
#define NO_DEVICE_CODE GGML_ASSERT(false && "NO_DEVICE_CODE not valid in host code.")
|
||||||
|
#endif // __CUDA_ARCH__
|
||||||
|
|
||||||
static __device__ __forceinline__ float warp_reduce_sum(float x) {
|
static __device__ __forceinline__ float warp_reduce_sum(float x) {
|
||||||
#pragma unroll
|
#pragma unroll
|
||||||
for (int mask = 16; mask > 0; mask >>= 1) {
|
for (int mask = 16; mask > 0; mask >>= 1) {
|
||||||
|
@ -617,7 +636,7 @@ static __device__ __forceinline__ half2 warp_reduce_sum(half2 a) {
|
||||||
return a;
|
return a;
|
||||||
#else
|
#else
|
||||||
(void) a;
|
(void) a;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // !(defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)) && __CUDA_ARCH__ >= CC_PASCAL
|
#endif // !(defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)) && __CUDA_ARCH__ >= CC_PASCAL
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -638,7 +657,7 @@ static __device__ __forceinline__ half2 warp_reduce_max(half2 x) {
|
||||||
return x;
|
return x;
|
||||||
#else
|
#else
|
||||||
(void) x;
|
(void) x;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // !(defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)) && __CUDA_ARCH__ >= CC_PASCAL && CUDART_VERSION >= CUDART_HMAX
|
#endif // !(defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)) && __CUDA_ARCH__ >= CC_PASCAL && CUDART_VERSION >= CUDART_HMAX
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2421,7 +2440,7 @@ static __global__ void dequantize_block_q8_0_f16(const void * __restrict__ vx, h
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
(void) vx; (void) y; (void) k;
|
(void) vx; (void) y; (void) k;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_PASCAL
|
#endif // __CUDA_ARCH__ >= CC_PASCAL
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2452,7 +2471,7 @@ template <int vdr> static __device__ __forceinline__ float vec_dot_q4_0_q8_1_imp
|
||||||
// second part effectively subtracts 8 from each quant value
|
// second part effectively subtracts 8 from each quant value
|
||||||
return d4 * (sumi * ds8f.x - (8*vdr/QI4_0) * ds8f.y);
|
return d4 * (sumi * ds8f.x - (8*vdr/QI4_0) * ds8f.y);
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2489,7 +2508,7 @@ template <int vdr> static __device__ __forceinline__ float vec_dot_q4_1_q8_1_imp
|
||||||
// scale second part of sum by QI8_1/(vdr * QR4_1) to compensate for multiple threads adding it
|
// scale second part of sum by QI8_1/(vdr * QR4_1) to compensate for multiple threads adding it
|
||||||
return sumi * d4d8 + m4s8 / (QI8_1 / (vdr * QR4_1));
|
return sumi * d4d8 + m4s8 / (QI8_1 / (vdr * QR4_1));
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2524,7 +2543,7 @@ template <int vdr> static __device__ __forceinline__ float vec_dot_q5_0_q8_1_imp
|
||||||
// second part effectively subtracts 16 from each quant value
|
// second part effectively subtracts 16 from each quant value
|
||||||
return d5 * (sumi * ds8f.x - (16*vdr/QI5_0) * ds8f.y);
|
return d5 * (sumi * ds8f.x - (16*vdr/QI5_0) * ds8f.y);
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2569,7 +2588,7 @@ template <int vdr> static __device__ __forceinline__ float vec_dot_q5_1_q8_1_imp
|
||||||
return sumi*d5d8 + m5s8 / (QI5_1 / vdr);
|
return sumi*d5d8 + m5s8 / (QI5_1 / vdr);
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2590,7 +2609,7 @@ template <int vdr> static __device__ __forceinline__ float vec_dot_q8_0_q8_1_imp
|
||||||
|
|
||||||
return d8_0*d8_1 * sumi;
|
return d8_0*d8_1 * sumi;
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2620,7 +2639,7 @@ template <int vdr> static __device__ __forceinline__ float vec_dot_q8_1_q8_1_imp
|
||||||
// scale second part of sum by QI8_1/ vdr to compensate for multiple threads adding it
|
// scale second part of sum by QI8_1/ vdr to compensate for multiple threads adding it
|
||||||
return sumi*d8d8 + m8s8 / (QI8_1 / vdr);
|
return sumi*d8d8 + m8s8 / (QI8_1 / vdr);
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2655,7 +2674,7 @@ static __device__ __forceinline__ float vec_dot_q2_K_q8_1_impl_mmvq(
|
||||||
|
|
||||||
return dm2f.x*sumf_d - dm2f.y*sumf_m;
|
return dm2f.x*sumf_d - dm2f.y*sumf_m;
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2692,7 +2711,7 @@ static __device__ __forceinline__ float vec_dot_q2_K_q8_1_impl_mmq(
|
||||||
|
|
||||||
return d8 * (dm2f.x*sumi_d - dm2f.y*sumi_m);
|
return d8 * (dm2f.x*sumi_d - dm2f.y*sumi_m);
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2732,7 +2751,7 @@ static __device__ __forceinline__ float vec_dot_q3_K_q8_1_impl_mmvq(
|
||||||
|
|
||||||
return d3 * sumf;
|
return d3 * sumf;
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2757,7 +2776,7 @@ static __device__ __forceinline__ float vec_dot_q3_K_q8_1_impl_mmq(
|
||||||
|
|
||||||
return d3*d8 * sumi;
|
return d3*d8 * sumi;
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2790,7 +2809,7 @@ static __device__ __forceinline__ float vec_dot_q4_K_q8_1_impl_vmmq(
|
||||||
return dm4f.x*sumf_d - dm4f.y*sumf_m;
|
return dm4f.x*sumf_d - dm4f.y*sumf_m;
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2823,7 +2842,7 @@ static __device__ __forceinline__ float vec_dot_q4_K_q8_1_impl_mmq(
|
||||||
return dm4f.x*sumf_d - dm4f.y*sumf_m;
|
return dm4f.x*sumf_d - dm4f.y*sumf_m;
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2863,7 +2882,7 @@ static __device__ __forceinline__ float vec_dot_q5_K_q8_1_impl_vmmq(
|
||||||
return dm5f.x*sumf_d - dm5f.y*sumf_m;
|
return dm5f.x*sumf_d - dm5f.y*sumf_m;
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2896,7 +2915,7 @@ static __device__ __forceinline__ float vec_dot_q5_K_q8_1_impl_mmq(
|
||||||
return dm4f.x*sumf_d - dm4f.y*sumf_m;
|
return dm4f.x*sumf_d - dm4f.y*sumf_m;
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2926,7 +2945,7 @@ static __device__ __forceinline__ float vec_dot_q6_K_q8_1_impl_mmvq(
|
||||||
|
|
||||||
return d*sumf;
|
return d*sumf;
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2957,7 +2976,7 @@ static __device__ __forceinline__ float vec_dot_q6_K_q8_1_impl_mmq(
|
||||||
return d6 * sumf_d;
|
return d6 * sumf_d;
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -3823,7 +3842,7 @@ static __device__ __forceinline__ float vec_dot_q4_K_q8_1(
|
||||||
return dall * sumf_d - dmin * sumf_m;
|
return dall * sumf_d - dmin * sumf_m;
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@ -4006,7 +4025,7 @@ static __device__ __forceinline__ float vec_dot_q5_K_q8_1(
|
||||||
return d * sumf_d;
|
return d * sumf_d;
|
||||||
|
|
||||||
#else
|
#else
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
#endif // __CUDA_ARCH__ >= MIN_CC_DP4A
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@ -4501,7 +4520,7 @@ template <bool need_check> static __global__ void
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q4_0_q8_1_mul_mat;
|
(void) vec_dot_q4_0_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4570,7 +4589,7 @@ template <bool need_check> static __global__ void
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q4_1_q8_1_mul_mat;
|
(void) vec_dot_q4_1_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4637,7 +4656,7 @@ template <bool need_check> static __global__ void
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q5_0_q8_1_mul_mat;
|
(void) vec_dot_q5_0_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4704,7 +4723,7 @@ mul_mat_q5_1(
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q5_1_q8_1_mul_mat;
|
(void) vec_dot_q5_1_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4771,7 +4790,7 @@ template <bool need_check> static __global__ void
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q8_0_q8_1_mul_mat;
|
(void) vec_dot_q8_0_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4838,7 +4857,7 @@ mul_mat_q2_K(
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q2_K_q8_1_mul_mat;
|
(void) vec_dot_q2_K_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4907,7 +4926,7 @@ template <bool need_check> static __global__ void
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q3_K_q8_1_mul_mat;
|
(void) vec_dot_q3_K_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -4976,7 +4995,7 @@ template <bool need_check> static __global__ void
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q4_K_q8_1_mul_mat;
|
(void) vec_dot_q4_K_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -5043,7 +5062,7 @@ mul_mat_q5_K(
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q5_K_q8_1_mul_mat;
|
(void) vec_dot_q5_K_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -5112,7 +5131,7 @@ template <bool need_check> static __global__ void
|
||||||
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
(vx, vy, dst, ncols_x, nrows_x, ncols_y, nrows_y, nrows_dst);
|
||||||
#else
|
#else
|
||||||
(void) vec_dot_q6_K_q8_1_mul_mat;
|
(void) vec_dot_q6_K_q8_1_mul_mat;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
#endif // __CUDA_ARCH__ >= CC_VOLTA
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -5835,7 +5854,7 @@ static __global__ void soft_max_f16(const float * x, const float * y, float * ds
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
(void) x; (void) y; (void) dst; (void) ncols_par; (void) nrows_y; (void) scale;
|
(void) x; (void) y; (void) dst; (void) ncols_par; (void) nrows_y; (void) scale;
|
||||||
bad_arch();
|
NO_DEVICE_CODE;
|
||||||
#endif // !(defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)) && __CUDA_ARCH__ >= CC_PASCAL && CUDART_VERSION >= CUDART_HMAX
|
#endif // !(defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)) && __CUDA_ARCH__ >= CC_PASCAL && CUDART_VERSION >= CUDART_HMAX
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -668,7 +668,8 @@ static bool ggml_metal_supports_op(const struct ggml_metal_context * ctx, const
|
||||||
return true;
|
return true;
|
||||||
case GGML_OP_MUL_MAT:
|
case GGML_OP_MUL_MAT:
|
||||||
case GGML_OP_MUL_MAT_ID:
|
case GGML_OP_MUL_MAT_ID:
|
||||||
return ctx->support_simdgroup_reduction;
|
return ctx->support_simdgroup_reduction &&
|
||||||
|
(op->src[0]->type != GGML_TYPE_F32 || op->src[1]->type == GGML_TYPE_F32);
|
||||||
case GGML_OP_CPY:
|
case GGML_OP_CPY:
|
||||||
case GGML_OP_DUP:
|
case GGML_OP_DUP:
|
||||||
case GGML_OP_CONT:
|
case GGML_OP_CONT:
|
||||||
|
|
9
ggml.c
9
ggml.c
|
@ -5372,14 +5372,12 @@ struct ggml_tensor * ggml_conv_depthwise_2d(
|
||||||
struct ggml_context * ctx,
|
struct ggml_context * ctx,
|
||||||
struct ggml_tensor * a,
|
struct ggml_tensor * a,
|
||||||
struct ggml_tensor * b,
|
struct ggml_tensor * b,
|
||||||
struct ggml_tensor * c,
|
|
||||||
int s0,
|
int s0,
|
||||||
int s1,
|
int s1,
|
||||||
int p0,
|
int p0,
|
||||||
int p1,
|
int p1,
|
||||||
int d0,
|
int d0,
|
||||||
int d1) {
|
int d1) {
|
||||||
|
|
||||||
struct ggml_tensor * new_a = ggml_reshape_4d(ctx, a, a->ne[0], a->ne[1], 1, a->ne[2] * a->ne[3]);
|
struct ggml_tensor * new_a = ggml_reshape_4d(ctx, a, a->ne[0], a->ne[1], 1, a->ne[2] * a->ne[3]);
|
||||||
struct ggml_tensor * im2col = ggml_im2col(ctx, new_a,
|
struct ggml_tensor * im2col = ggml_im2col(ctx, new_a,
|
||||||
ggml_reshape_4d(ctx, b, b->ne[0], b->ne[1], 1, b->ne[2] * b->ne[3]),
|
ggml_reshape_4d(ctx, b, b->ne[0], b->ne[1], 1, b->ne[2] * b->ne[3]),
|
||||||
|
@ -9995,7 +9993,7 @@ static void ggml_compute_forward_mul_mat(
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const int64_t tgemm0 = ggml_perf_time_us();
|
//const int64_t tgemm0 = ggml_perf_time_us();
|
||||||
for (int64_t i13 = 0; i13 < ne13; i13++) {
|
for (int64_t i13 = 0; i13 < ne13; i13++) {
|
||||||
for (int64_t i12 = 0; i12 < ne12; i12++) {
|
for (int64_t i12 = 0; i12 < ne12; i12++) {
|
||||||
const int64_t i03 = i13/r3;
|
const int64_t i03 = i13/r3;
|
||||||
|
@ -16944,7 +16942,10 @@ struct ggml_cplan ggml_graph_plan(const struct ggml_cgraph * cgraph, int n_threa
|
||||||
if (ggml_compute_forward_mul_mat_use_blas(node)) {
|
if (ggml_compute_forward_mul_mat_use_blas(node)) {
|
||||||
if (node->src[0]->type != GGML_TYPE_F32) {
|
if (node->src[0]->type != GGML_TYPE_F32) {
|
||||||
// here we need memory for fully dequantized matrix from src0
|
// here we need memory for fully dequantized matrix from src0
|
||||||
cur = ggml_type_size(GGML_TYPE_F32)*ggml_nelements(node->src[0]);
|
// take into account that src0 can be broadcasted into src1[2,3]
|
||||||
|
cur = ggml_type_size(GGML_TYPE_F32)
|
||||||
|
* node->src[0]->ne[0]*node->src[0]->ne[1]
|
||||||
|
* node->src[1]->ne[2]*node->src[1]->ne[3];
|
||||||
}
|
}
|
||||||
} else
|
} else
|
||||||
#endif
|
#endif
|
||||||
|
|
1
ggml.h
1
ggml.h
|
@ -1499,7 +1499,6 @@ extern "C" {
|
||||||
struct ggml_context * ctx,
|
struct ggml_context * ctx,
|
||||||
struct ggml_tensor * a,
|
struct ggml_tensor * a,
|
||||||
struct ggml_tensor * b,
|
struct ggml_tensor * b,
|
||||||
struct ggml_tensor * c,
|
|
||||||
int s0,
|
int s0,
|
||||||
int s1,
|
int s1,
|
||||||
int p0,
|
int p0,
|
||||||
|
|
24
llama.cpp
24
llama.cpp
|
@ -2306,18 +2306,18 @@ struct llama_model_loader {
|
||||||
}
|
}
|
||||||
|
|
||||||
switch (type_max) {
|
switch (type_max) {
|
||||||
case GGML_TYPE_F32: ftype = LLAMA_FTYPE_ALL_F32; break;
|
case GGML_TYPE_F32: ftype = LLAMA_FTYPE_ALL_F32; break;
|
||||||
case GGML_TYPE_F16: ftype = LLAMA_FTYPE_MOSTLY_F16; break;
|
case GGML_TYPE_F16: ftype = LLAMA_FTYPE_MOSTLY_F16; break;
|
||||||
case GGML_TYPE_Q4_0: ftype = LLAMA_FTYPE_MOSTLY_Q4_0; break;
|
case GGML_TYPE_Q4_0: ftype = LLAMA_FTYPE_MOSTLY_Q4_0; break;
|
||||||
case GGML_TYPE_Q4_1: ftype = LLAMA_FTYPE_MOSTLY_Q4_1; break;
|
case GGML_TYPE_Q4_1: ftype = LLAMA_FTYPE_MOSTLY_Q4_1; break;
|
||||||
case GGML_TYPE_Q5_0: ftype = LLAMA_FTYPE_MOSTLY_Q5_0; break;
|
case GGML_TYPE_Q5_0: ftype = LLAMA_FTYPE_MOSTLY_Q5_0; break;
|
||||||
case GGML_TYPE_Q5_1: ftype = LLAMA_FTYPE_MOSTLY_Q5_1; break;
|
case GGML_TYPE_Q5_1: ftype = LLAMA_FTYPE_MOSTLY_Q5_1; break;
|
||||||
case GGML_TYPE_Q8_0: ftype = LLAMA_FTYPE_MOSTLY_Q8_0; break;
|
case GGML_TYPE_Q8_0: ftype = LLAMA_FTYPE_MOSTLY_Q8_0; break;
|
||||||
case GGML_TYPE_Q2_K: ftype = LLAMA_FTYPE_MOSTLY_Q2_K; break;
|
case GGML_TYPE_Q2_K: ftype = LLAMA_FTYPE_MOSTLY_Q2_K; break;
|
||||||
case GGML_TYPE_Q3_K: ftype = LLAMA_FTYPE_MOSTLY_Q3_K_M; break;
|
case GGML_TYPE_Q3_K: ftype = LLAMA_FTYPE_MOSTLY_Q3_K_M; break;
|
||||||
case GGML_TYPE_Q4_K: ftype = LLAMA_FTYPE_MOSTLY_Q4_K_M; break;
|
case GGML_TYPE_Q4_K: ftype = LLAMA_FTYPE_MOSTLY_Q4_K_M; break;
|
||||||
case GGML_TYPE_Q5_K: ftype = LLAMA_FTYPE_MOSTLY_Q5_K_M; break;
|
case GGML_TYPE_Q5_K: ftype = LLAMA_FTYPE_MOSTLY_Q5_K_M; break;
|
||||||
case GGML_TYPE_Q6_K: ftype = LLAMA_FTYPE_MOSTLY_Q6_K; break;
|
case GGML_TYPE_Q6_K: ftype = LLAMA_FTYPE_MOSTLY_Q6_K; break;
|
||||||
case GGML_TYPE_IQ2_XXS: ftype = LLAMA_FTYPE_MOSTLY_IQ2_XXS; break;
|
case GGML_TYPE_IQ2_XXS: ftype = LLAMA_FTYPE_MOSTLY_IQ2_XXS; break;
|
||||||
case GGML_TYPE_IQ2_XS: ftype = LLAMA_FTYPE_MOSTLY_IQ2_XS; break;
|
case GGML_TYPE_IQ2_XS: ftype = LLAMA_FTYPE_MOSTLY_IQ2_XS; break;
|
||||||
default:
|
default:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue