1
0
mirror of https://github.com/ggerganov/llama.cpp.git synced 2025-01-23 09:59:18 +01:00
llama.cpp/tests
slaren 6744dbe924
ggml : use ggml_row_size where possible ()
* ggml : use ggml_row_size where possible

ggml-ci

* ggml : move ggml_nbytes_split to ggml-cuda.cu
2023-12-14 20:05:21 +01:00
..
CMakeLists.txt sync : ggml (new ops, tests, backend, etc.) () 2023-12-07 22:26:54 +02:00
test-backend-ops.cpp ggml : use ggml_row_size where possible () 2023-12-14 20:05:21 +01:00
test-c.c tests : add a C compliance test () 2023-08-30 09:20:26 +03:00
test-double-float.cpp ggml : move FP16 <-> FP32 code to ggml-impl.h () 2023-10-30 19:19:15 +02:00
test-grad0.cpp english : use typos to fix comments and logs () 2023-12-12 11:53:36 +02:00
test-grammar-parser.cpp gguf : new file format with flexible meta data (beta) () 2023-08-21 23:07:43 +03:00
test-llama-grammar.cpp gguf : new file format with flexible meta data (beta) () 2023-08-21 23:07:43 +03:00
test-opt.cpp sync : ggml (backend v2) () 2023-11-13 14:16:23 +02:00
test-quantize-fns.cpp ggml : move FP16 <-> FP32 code to ggml-impl.h () 2023-10-30 19:19:15 +02:00
test-quantize-perf.cpp ggml : use ggml_row_size where possible () 2023-12-14 20:05:21 +01:00
test-rope.cpp llama : custom attention mask + parallel decoding + no context swaps () 2023-09-28 19:04:36 +03:00
test-sampling.cpp sampling : refactor init to use llama_sampling_params () 2023-10-20 21:07:23 +03:00
test-tokenizer-0-falcon.cpp Minor improvements in GPT2 tokenizer () 2023-10-10 18:59:52 +02:00
test-tokenizer-0-falcon.py ci : add flake8 to github actions (python linting) () 2023-11-20 11:35:47 +01:00
test-tokenizer-0-llama.cpp Minor improvements in GPT2 tokenizer () 2023-10-10 18:59:52 +02:00
test-tokenizer-0-llama.py ci : add flake8 to github actions (python linting) () 2023-11-20 11:35:47 +01:00
test-tokenizer-1-bpe.cpp Add more tokenizer tests () 2023-10-24 09:17:17 +02:00
test-tokenizer-1-llama.cpp Work on the BPE tokenizer () 2023-10-03 09:16:26 +02:00