Merge remote-tracking branch 'upstream/ik/context_extend' into concedo_experimental
# Conflicts: # CMakeLists.txt # Makefile
This commit is contained in:
commit
10a2bdfaf1
4 changed files with 26 additions and 3 deletions
|
@ -2220,10 +2220,13 @@ inline void ggml_cuda_op_rope(
|
|||
const int n_past = ((int32_t *) src1->data)[0];
|
||||
const int n_dims = ((int32_t *) src1->data)[1];
|
||||
const int mode = ((int32_t *) src1->data)[2];
|
||||
const int n_ctx = ((int32_t *) src1->data)[3];
|
||||
GGML_ASSERT(mode == 0);
|
||||
|
||||
const float theta_scale = powf(10000.0, -2.0f/n_dims);
|
||||
const float p = ((mode & 1) == 0 ? n_past + i02 : i02);
|
||||
const float p0 = ((mode & 1) == 0 ? n_past + i02 : i02);
|
||||
|
||||
const float p = n_ctx <= GGML_TRAINING_CTX ? p0 : p0 * GGML_TRAINING_CTX / n_ctx;
|
||||
|
||||
// compute
|
||||
rope_f32_cuda(src0_ddf_i, dst_ddf_i, ne00, i01_diff, p, theta_scale, cudaStream_main);
|
||||
|
|
14
ggml.c
14
ggml.c
|
@ -12535,6 +12535,9 @@ static void ggml_compute_forward_rope_f32(
|
|||
dst_data[n_dims/2*3] = x2*sin_block_theta + x3*cos_block_theta;
|
||||
}
|
||||
} else if (!is_neox) {
|
||||
if (n_ctx > GGML_TRAINING_CTX) {
|
||||
theta = theta * GGML_TRAINING_CTX / n_ctx;
|
||||
}
|
||||
for (int64_t i0 = 0; i0 < ne0; i0 += 2) {
|
||||
const float cos_theta = cosf(theta);
|
||||
const float sin_theta = sinf(theta);
|
||||
|
@ -12675,6 +12678,9 @@ static void ggml_compute_forward_rope_f16(
|
|||
dst_data[n_dims/2*3] = GGML_FP32_TO_FP16(x2*sin_block_theta + x3*cos_block_theta);
|
||||
}
|
||||
} if (!is_neox) {
|
||||
if (n_ctx > GGML_TRAINING_CTX) {
|
||||
theta = theta * GGML_TRAINING_CTX / n_ctx;
|
||||
}
|
||||
for (int64_t i0 = 0; i0 < ne0; i0 += 2) {
|
||||
const float cos_theta = cosf(theta);
|
||||
const float sin_theta = sinf(theta);
|
||||
|
@ -12760,6 +12766,7 @@ static void ggml_compute_forward_rope_back_f32(
|
|||
const int n_past = ((int32_t *) src1->data)[0];
|
||||
const int n_dims = ((int32_t *) src1->data)[1];
|
||||
const int mode = ((int32_t *) src1->data)[2];
|
||||
const int n_ctx = ((int32_t *) src1->data)[3];
|
||||
|
||||
assert(n_past >= 0);
|
||||
|
||||
|
@ -12813,6 +12820,9 @@ static void ggml_compute_forward_rope_back_f32(
|
|||
float theta = (float)p;
|
||||
|
||||
if (!is_neox) {
|
||||
if (n_ctx > GGML_TRAINING_CTX) {
|
||||
theta = theta * GGML_TRAINING_CTX / n_ctx;
|
||||
}
|
||||
for (int64_t i0 = 0; i0 < ne0; i0 += 2) {
|
||||
const float cos_theta = cosf(theta);
|
||||
const float sin_theta = sinf(theta);
|
||||
|
@ -12873,6 +12883,7 @@ static void ggml_compute_forward_rope_back_f16(
|
|||
const int n_past = ((int32_t *) src1->data)[0];
|
||||
const int n_dims = ((int32_t *) src1->data)[1];
|
||||
const int mode = ((int32_t *) src1->data)[2];
|
||||
const int n_ctx = ((int32_t *) src1->data)[3];
|
||||
|
||||
assert(n_past >= 0);
|
||||
|
||||
|
@ -12926,6 +12937,9 @@ static void ggml_compute_forward_rope_back_f16(
|
|||
float theta = (float)p;
|
||||
|
||||
if (!is_neox) {
|
||||
if (n_ctx > GGML_TRAINING_CTX) {
|
||||
theta = theta * GGML_TRAINING_CTX / n_ctx;
|
||||
}
|
||||
for (int64_t i0 = 0; i0 < ne0; i0 += 2) {
|
||||
const float cos_theta = cosf(theta);
|
||||
const float sin_theta = sinf(theta);
|
||||
|
|
6
ggml.h
6
ggml.h
|
@ -201,6 +201,12 @@
|
|||
#define GGML_MAX_NAME 48
|
||||
#define GGML_DEFAULT_N_THREADS 4
|
||||
|
||||
// Maximum training context of the model in use
|
||||
// For the LLaMA models this is normally 2048, but somehow "stepping out" by 128 gives better results (tested at 7B and 13B)
|
||||
#ifndef GGML_TRAINING_CTX
|
||||
#define GGML_TRAINING_CTX 2176
|
||||
#endif
|
||||
|
||||
#define GGML_ASSERT(x) \
|
||||
do { \
|
||||
if (!(x)) { \
|
||||
|
|
|
@ -1338,11 +1338,11 @@ static bool llama_eval_internal(
|
|||
offload_func_kq(tmpq);
|
||||
ggml_set_name(tmpq, "tmpq");
|
||||
|
||||
struct ggml_tensor * Kcur = ggml_rope_inplace(ctx0, ggml_reshape_3d(ctx0, tmpk, n_embd/n_head, n_head, N), n_past, n_rot, 0, 0);
|
||||
struct ggml_tensor * Kcur = ggml_rope_inplace(ctx0, ggml_reshape_3d(ctx0, tmpk, n_embd/n_head, n_head, N), n_past, n_rot, 0, n_ctx);
|
||||
offload_func_kq(Kcur);
|
||||
ggml_set_name(Kcur, "Kcur");
|
||||
|
||||
struct ggml_tensor * Qcur = ggml_rope_inplace(ctx0, ggml_reshape_3d(ctx0, tmpq, n_embd/n_head, n_head, N), n_past, n_rot, 0, 0);
|
||||
struct ggml_tensor * Qcur = ggml_rope_inplace(ctx0, ggml_reshape_3d(ctx0, tmpq, n_embd/n_head, n_head, N), n_past, n_rot, 0, n_ctx);
|
||||
offload_func_kq(Qcur);
|
||||
ggml_set_name(Qcur, "Qcur");
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue