1
0
mirror of https://github.com/ggerganov/llama.cpp.git synced 2025-01-19 00:18:57 +01:00
llama.cpp/tests
2024-01-04 09:43:23 +01:00
..
CMakeLists.txt cmake : fix ld warning duplicate libraries libllama.a () 2023-12-29 16:39:15 +02:00
test-backend-ops.cpp Print backend name on test-backend-ops failure () 2024-01-04 09:43:23 +01:00
test-c.c tests : add a C compliance test () 2023-08-30 09:20:26 +03:00
test-double-float.cpp ggml : move FP16 <-> FP32 code to ggml-impl.h () 2023-10-30 19:19:15 +02:00
test-grad0.cpp cuda : improve cuda pool efficiency using virtual memory () 2023-12-24 14:34:22 +01:00
test-grammar-parser.cpp gguf : new file format with flexible meta data (beta) () 2023-08-21 23:07:43 +03:00
test-llama-grammar.cpp gguf : new file format with flexible meta data (beta) () 2023-08-21 23:07:43 +03:00
test-opt.cpp sync : ggml (backend v2) () 2023-11-13 14:16:23 +02:00
test-quantize-fns.cpp ggml : move FP16 <-> FP32 code to ggml-impl.h () 2023-10-30 19:19:15 +02:00
test-quantize-perf.cpp ggml : use ggml_row_size where possible () 2023-12-14 20:05:21 +01:00
test-rope.cpp llama : custom attention mask + parallel decoding + no context swaps () 2023-09-28 19:04:36 +03:00
test-sampling.cpp sampling : refactor init to use llama_sampling_params () 2023-10-20 21:07:23 +03:00
test-tokenizer-0-falcon.cpp Minor improvements in GPT2 tokenizer () 2023-10-10 18:59:52 +02:00
test-tokenizer-0-falcon.py ci : add flake8 to github actions (python linting) () 2023-11-20 11:35:47 +01:00
test-tokenizer-0-llama.cpp Minor improvements in GPT2 tokenizer () 2023-10-10 18:59:52 +02:00
test-tokenizer-0-llama.py ci : add flake8 to github actions (python linting) () 2023-11-20 11:35:47 +01:00
test-tokenizer-1-bpe.cpp Add more tokenizer tests () 2023-10-24 09:17:17 +02:00
test-tokenizer-1-llama.cpp Work on the BPE tokenizer () 2023-10-03 09:16:26 +02:00