mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-10-30 06:30:15 +01:00
ec893798b7
* tests : verify that RoPE is "additive" * llama : replace ggml_diag_mask_inf with ggml_add (custom -inf mask) * ggml : ggml_rope now takes a vector with positions instead of n_past * metal : add rope_f16 kernel + optimize cpy kernels * llama : unified KV cache + batch inference API * llama : add new llama_decode() API that works with llama_batch * llama : add cell_max heuristic for more efficient kv_cache * llama : extend llama_kv_cache API * llama : more robust cell_max heuristic + wip shift * metal : disable concurrency optimization * llama : add llama_kv_cache_shift_seq + no more context swaps * llama : apply K-cache roping for Falcon and Baichuan * speculative : fix KV cache management * parallel : example for serving multiple users in parallel * parallel : disable hot-plug to avoid cache fragmentation * fixes : speculative KV cache + llama worst-case graph * llama : extend batch API to select which logits to output * llama : fix worst case graph build * ggml-cuda : update rope implementation for parallel decoding (#3254) * ggml-cuda : update rope implementation for parallel decoding * better solution for p0 computation * fix rope * simpler rope implementation --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * make : add parallel to build + fix static functions in llama.cpp * simple : fix token counting * parallel : various improvements * llama : fix cell_max logic + rename functions * parallel : try smaller batches when the KV cache is fragmented * parallel : fix sequence termination criteria * llama : silence errors KV cache errors * parallel : remove new line from prompt * parallel : process system prompt once + configurable paramters + llama API * parallel : remove question with short answers * parallel : count cache misses * parallel : print misses on each request * parallel : minor * llama : fix n_kv to never become 0 * parallel : rename hot-plug to continuous-batching * llama : improve llama_batch API + simplify parallel example * simple : add parallel decoding support * simple : improve comments + free batch * ggml-cuda : add rope f16, restore performance with parallel decoding (#3272) * ggml-cuda : add rope f16, restore performance * offload KQ_mask with all models * fix rope shift --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * llama : disable MPI for now ggml-ci * train : make KQ_pos memory buffer permanent via dummy scale op * ggml : revert change to ggml_cpy, add ggml_cont_Nd instead (#3275) ggml-ci * parallel : fix bug (extra BOS) + smaller token_prev array * parallel : fix cases where the input prompts can overflow the batch * parallel : add disabled experimental batch chunking in powers of two * llama : llama.h formatting + comments * simple : add README.md * llama : fix kv cache heuristic when context is less than 32 * parallel : fix crash when `-n -1` * llama : simplify returns if/else branches * metal : use mm kernels for batch size > 2 * examples : utilize new llama_get_logits_ith() * examples : add example for batched decoding * examples : do not eval prompt 2 times (close #3348) * server : clear the KV cache beyond n_past before llama_decode * server : avoid context swaps by shifting the KV cache --------- Co-authored-by: slaren <slarengh@gmail.com>
222 lines
6.1 KiB
C++
222 lines
6.1 KiB
C++
#include "ggml.h"
|
|
|
|
#include <cmath>
|
|
#include <cstdio>
|
|
#include <cstdlib>
|
|
#include <cassert>
|
|
#include <vector>
|
|
|
|
#if defined(_MSC_VER)
|
|
#pragma warning(disable: 4244 4267) // possible loss of data
|
|
#endif
|
|
|
|
#if defined(__GNUC__)
|
|
#pragma GCC diagnostic ignored "-Wdouble-promotion"
|
|
#endif
|
|
|
|
#define MAX_NARGS 3
|
|
|
|
#undef MIN
|
|
#undef MAX
|
|
#define MIN(a, b) ((a) < (b) ? (a) : (b))
|
|
#define MAX(a, b) ((a) > (b) ? (a) : (b))
|
|
|
|
#define GGML_SILU_FP16
|
|
|
|
//
|
|
// logging
|
|
//
|
|
|
|
#if (GGML_DEBUG >= 1)
|
|
#define GGML_PRINT_DEBUG(...) printf(__VA_ARGS__)
|
|
#else
|
|
#define GGML_PRINT_DEBUG(...)
|
|
#endif
|
|
|
|
#if (GGML_DEBUG >= 5)
|
|
#define GGML_PRINT_DEBUG_5(...) printf(__VA_ARGS__)
|
|
#else
|
|
#define GGML_PRINT_DEBUG_5(...)
|
|
#endif
|
|
|
|
#if (GGML_DEBUG >= 10)
|
|
#define GGML_PRINT_DEBUG_10(...) printf(__VA_ARGS__)
|
|
#else
|
|
#define GGML_PRINT_DEBUG_10(...)
|
|
#endif
|
|
|
|
#define GGML_PRINT(...) printf(__VA_ARGS__)
|
|
|
|
static float frand(void) {
|
|
return (float)rand()/(float)RAND_MAX;
|
|
}
|
|
|
|
static int irand(int n) {
|
|
if (n == 0) return 0;
|
|
return rand()%n;
|
|
}
|
|
|
|
static void get_random_dims(int64_t * dims, int ndims) {
|
|
dims[0] = dims[1] = dims[2] = dims[3] = 1;
|
|
|
|
for (int i = 0; i < ndims; i++) {
|
|
dims[i] = 1 + irand(4);
|
|
}
|
|
}
|
|
|
|
static struct ggml_tensor * get_random_tensor_f32(
|
|
struct ggml_context * ctx0,
|
|
int ndims,
|
|
const int64_t ne[],
|
|
float fmin,
|
|
float fmax) {
|
|
struct ggml_tensor * result = ggml_new_tensor(ctx0, GGML_TYPE_F32, ndims, ne);
|
|
|
|
switch (ndims) {
|
|
case 1:
|
|
for (int i0 = 0; i0 < ne[0]; i0++) {
|
|
((float *)result->data)[i0] = frand()*(fmax - fmin) + fmin;
|
|
}
|
|
break;
|
|
case 2:
|
|
for (int i1 = 0; i1 < ne[1]; i1++) {
|
|
for (int i0 = 0; i0 < ne[0]; i0++) {
|
|
((float *)result->data)[i1*ne[0] + i0] = frand()*(fmax - fmin) + fmin;
|
|
}
|
|
}
|
|
break;
|
|
case 3:
|
|
for (int i2 = 0; i2 < ne[2]; i2++) {
|
|
for (int i1 = 0; i1 < ne[1]; i1++) {
|
|
for (int i0 = 0; i0 < ne[0]; i0++) {
|
|
((float *)result->data)[i2*ne[1]*ne[0] + i1*ne[0] + i0] = frand()*(fmax - fmin) + fmin;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
case 4:
|
|
for (int i3 = 0; i3 < ne[3]; i3++) {
|
|
for (int i2 = 0; i2 < ne[2]; i2++) {
|
|
for (int i1 = 0; i1 < ne[1]; i1++) {
|
|
for (int i0 = 0; i0 < ne[0]; i0++) {
|
|
((float *)result->data)[i3*ne[2]*ne[1]*ne[0] + i2*ne[1]*ne[0] + i1*ne[0] + i0] = frand()*(fmax - fmin) + fmin;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
default:
|
|
assert(false);
|
|
};
|
|
|
|
return result;
|
|
}
|
|
|
|
static void ggml_graph_compute_helper(std::vector<uint8_t> & buf, ggml_cgraph * graph, int n_threads) {
|
|
struct ggml_cplan plan = ggml_graph_plan(graph, n_threads);
|
|
|
|
if (plan.work_size > 0) {
|
|
buf.resize(plan.work_size);
|
|
plan.work_data = buf.data();
|
|
}
|
|
|
|
ggml_graph_compute(graph, &plan);
|
|
}
|
|
|
|
int main(int /*argc*/, const char ** /*argv*/) {
|
|
struct ggml_init_params params = {
|
|
/* .mem_size = */ 128*1024*1024,
|
|
/* .mem_buffer = */ NULL,
|
|
/* .no_alloc = */ false,
|
|
};
|
|
|
|
std::vector<uint8_t> work_buffer;
|
|
|
|
struct ggml_context * ctx0 = ggml_init(params);
|
|
|
|
struct ggml_tensor * x;
|
|
|
|
// rope f32
|
|
for (int m = 0; m < 3; ++m) {
|
|
const int ndims = 4;
|
|
|
|
const int64_t n_rot = 128;
|
|
const int64_t ne[4] = { 2*n_rot, 32, 73, 1 };
|
|
|
|
const int n_past_0 = 100;
|
|
const int n_past_2 = 33;
|
|
|
|
struct ggml_tensor * p0 = ggml_new_tensor_1d(ctx0, GGML_TYPE_I32, ne[2]);
|
|
struct ggml_tensor * p1 = ggml_new_tensor_1d(ctx0, GGML_TYPE_I32, ne[2]);
|
|
struct ggml_tensor * p2 = ggml_new_tensor_1d(ctx0, GGML_TYPE_I32, ne[2]);
|
|
|
|
for (int i = 0; i < ne[2]; ++i) {
|
|
((int32_t *) p0->data)[i] = n_past_0 + i;
|
|
((int32_t *) p1->data)[i] = n_past_2 - n_past_0;
|
|
((int32_t *) p2->data)[i] = n_past_2 + i;
|
|
}
|
|
|
|
// test mode 0, 2, 4 (standard, GPT-NeoX, GLM)
|
|
const int mode = m == 0 ? 0 : m == 1 ? 2 : 4;
|
|
|
|
x = get_random_tensor_f32(ctx0, ndims, ne, -1.0f, 1.0f);
|
|
|
|
// 100, 101, 102, ..., 172
|
|
struct ggml_tensor * r0 = ggml_rope(ctx0, x, p0, n_rot, mode, 1024);
|
|
// -67, -67, -67, ..., -67
|
|
struct ggml_tensor * r1 = ggml_rope(ctx0, r0, p1, n_rot, mode, 1024); // "context swap", i.e. forget n_past_0 - n_past_2 tokens
|
|
|
|
// 33, 34, 35, ..., 105
|
|
struct ggml_tensor * r2 = ggml_rope(ctx0, x, p2, n_rot, mode, 1024);
|
|
|
|
ggml_cgraph * gf = ggml_new_graph(ctx0);
|
|
|
|
ggml_build_forward_expand(gf, r0);
|
|
ggml_build_forward_expand(gf, r1);
|
|
ggml_build_forward_expand(gf, r2);
|
|
|
|
ggml_graph_compute_helper(work_buffer, gf, 4);
|
|
|
|
// check that r1 and r2 are the same
|
|
{
|
|
double sum0 = 0.0f;
|
|
double sum1 = 0.0f;
|
|
double diff = 0.0f;
|
|
|
|
const float * r1_data = (float *) r1->data;
|
|
const float * r2_data = (float *) r2->data;
|
|
|
|
const int n_elements = ggml_nelements(r1);
|
|
|
|
for (int i = 0; i < n_elements; ++i) {
|
|
sum0 += fabs(r1_data[i]);
|
|
sum1 += fabs(r2_data[i]);
|
|
diff += fabs(r1_data[i] - r2_data[i]);
|
|
//if (fabs(r1_data[i] - r2_data[i]) > 0.0001f) {
|
|
// printf("%d: %f %f\n", i, r1_data[i], r2_data[i]);
|
|
// printf("diff: %f\n", fabs(r1_data[i] - r2_data[i]));
|
|
//}
|
|
}
|
|
|
|
//for (int i = 4096; i < 4096 + 128; ++i) {
|
|
// printf("%f %f\n", r1_data[i], r2_data[i]);
|
|
//}
|
|
|
|
printf("mode: %d\n", mode);
|
|
printf("sum0: %f\n", sum0);
|
|
printf("sum1: %f\n", sum1);
|
|
printf("diff: %f\n", diff);
|
|
printf("rel err: %f\n", diff / sum0);
|
|
printf("rel err: %f\n", diff / sum1);
|
|
|
|
GGML_ASSERT(diff / sum0 < 0.0001f);
|
|
GGML_ASSERT(diff / sum1 < 0.0001f);
|
|
}
|
|
}
|
|
|
|
ggml_free(ctx0);
|
|
|
|
return 0;
|
|
}
|
|
|