Commit Graph

  • d83cfbad0c
    Merge 'origin/master' into hipblas Henri Vasserman 2023-05-04 11:31:16 +0300
  • 76692c90cd q4_0c: avoid _mm512_loadu_epi64 instruction Håkon H. Hitland 2023-05-04 09:53:55 +0200
  • b63654c8df load pretrained vocab alex 2023-05-04 09:21:24 +0200
  • d53f76760d q4_0c: disable prefetching on M1 Håkon H. Hitland 2023-04-27 22:48:46 +0200
  • 2949725fea q4_0c: prefetch on AVX-512 and ARM Håkon H. Hitland 2023-04-24 18:17:31 +0200
  • 1b49d26f8a q4_0c: Arm Neon acceleration Håkon H. Hitland 2023-04-21 00:11:49 +0200
  • ab543dc1a4 q4_0c: AVX512 vec_dot and quantize impl Håkon H. Hitland 2023-04-18 23:07:03 +0200
  • 4bd781cd25 q4_0c: quantize support Håkon H. Hitland 2023-04-18 00:57:30 +0200
  • a1e6fb9281 q4_0c continous row layout Håkon H. Hitland 2023-04-17 23:36:29 +0200
  • 221946777c test-quantize: fix for q8_0 intermediates Håkon H. Hitland 2023-04-16 00:37:16 +0200
  • c8f7eeb7fd update kobold lite Concedo 2023-05-04 14:43:35 +0800
  • 981d71b281 Only escape prompts when used with -e Danny Daemonic 2023-05-03 23:23:24 -0700
  • e01dc631f7 Merge branch 'master' into concedo_experimental Concedo 2023-05-04 14:04:41 +0800
  • 7c129305f5 derp (+1 squashed commits) Concedo 2023-05-04 12:10:19 +0800
  • 3f30da38ad llama, main: save state incrementally Evan Jones 2023-05-03 02:09:19 -0400
  • 866fd3f3cb
    save a token CRD716 2023-05-03 21:19:27 -0500
  • c47b349281 Support DT_BF16 tensors Ivan Stepanov 2023-05-04 04:09:45 +0300
  • 932e616cf4
    Code Formatting Tomas 2023-05-04 07:27:52 +0700
  • 2b7cf9f32b fix too relaxed model glob (breaking multifile) alex 2023-05-04 00:16:12 +0200
  • aebb5d46ff
    fix typo in ggml.c Ron Jailall 2023-05-03 18:08:54 -0400
  • 286efed05c
    conditional def of intrin.h Ron Jailall 2023-05-03 18:06:40 -0400
  • b59c371035 add support for ByteStorage, relax model glob alex 2023-05-03 23:57:08 +0200
  • ca0a3e78d9
    change immintrin.h to intrin.h for compatibility Ron Jailall 2023-05-03 17:40:33 -0400
  • 31ff9e2e83
    ci : add cublas to windows release ci_cublas-31ff9e2 ci_cublas Green Sky 2023-05-01 12:41:46 +0200
  • 9f4505a0c6 fixed some bugs FSSRepo 2023-05-03 14:25:14 -0600
  • 799fdc1b5d
    ggml : vectorize Q8_0 quantization master-799fdc1 Georgi Gerganov 2023-05-03 23:24:20 +0300
  • 8dc342c069
    quick readme update CRD716 2023-05-03 15:08:02 -0500
  • f11c0f9aa1
    add model-agnostic dan prompt CRD716 2023-05-03 15:06:24 -0500
  • 45d94c8f6f
    ci : add cublas to windows release ci_cublas-45d94c8 Green Sky 2023-05-01 12:41:46 +0200
  • 44286d3bc5
    ci : add cublas to windows release ci_cublas-44286d3 Green Sky 2023-05-01 12:41:46 +0200
  • 6daa09d879
    examples : read chat prompts from a template file (#1196) khimaros 2023-05-03 10:58:11 -0700
  • cad6ff5d36 scripts : add ppl-run-all.sh Georgi Gerganov 2023-05-03 20:53:11 +0300
  • c2aa88189c read chat prompts from a template file khimaros 2023-04-18 14:48:23 -0700
  • 0652b4209f
    llama : require first token to be BOS Georgi Gerganov 2023-05-03 20:25:55 +0300
  • 3f870c55f8 quantize: make output filename optional, default to ggml-model-<ftype>.bin slaren 2023-05-03 18:43:11 +0200
  • bca9ad938a
    minor : fix whitespaces (#1302) Georgi Gerganov 2023-05-03 20:09:42 +0300
  • 32d8b3ff24
    minor : fix whitespaces Georgi Gerganov 2023-05-03 19:54:57 +0300
  • f684c4d414 Merge branch 'master' of https://github.com/FSSRepo/llama.cpp FSSRepo 2023-05-03 10:47:06 -0600
  • 197bb66339 Added readme for server example FSSRepo 2023-05-03 10:38:35 -0600
  • 3baa706a19
    Merge branch 'ggerganov:master' into master Steward Garcia 2023-05-03 10:35:19 -0600
  • e2a937ca6a
    minor : fix trailing whitespaces Georgi Gerganov 2023-05-03 18:43:23 +0300
  • ede8e4edbb Merge branch 'master' into concedo_experimental Concedo 2023-05-03 23:34:50 +0800
  • b0c71c7b6d
    scripts : platform independent script to verify sha256 checksums (#1203) KASR 2023-05-03 17:31:28 +0200
  • a8a2efdc81
    examples : various prompt and example fixes (#1298) CRD716 2023-05-03 10:26:47 -0500
  • 105f818d45 integrated new version of rwkv from upstream Concedo 2023-05-03 23:26:39 +0800
  • 773455084c
    use common characters CRD716 2023-05-03 08:41:01 -0500
  • c14ac96c2c
    miku prompt improvements CRD716 2023-05-03 08:37:24 -0500
  • 1abe47c8d9
    fix dan.txt CRD716 2023-05-03 08:33:27 -0500
  • 4857739ab5 allow specifying a different thread count for GPU blas Concedo 2023-05-03 21:19:59 +0800
  • b67cc50dad
    Merge 'origin/master' into hipblas Henri Vasserman 2023-05-03 15:04:51 +0300
  • b78af37cd2
    fix reverse prompt and multi line Tomas 2023-05-03 18:47:11 +0700
  • 9f788b9478
    Update verify-checksum-models.py KASR 2023-05-03 13:38:49 +0200
  • fe5d0d5a3b Update main's README.md with new features Danny Daemonic 2023-05-03 03:32:25 -0700
  • df69e4e43e
    Merge 902075752a into e216aa0463 Amir Eslampanah 2023-05-03 10:51:59 +0200
  • 3bdecc2aca
    Merge branch 'ggerganov:master' into master KASR 2023-05-03 08:51:58 +0200
  • f01c6cbc7e Added httplib support FSSRepo 2023-05-02 23:55:29 -0600
  • 89044502fe just use RT Concedo 2023-05-03 11:07:36 +0800
  • e216aa0463
    llama : only copy used KV cache in get / set state (#1272) master-e216aa0 Evan Jones 2023-05-02 22:26:13 -0400
  • f43a63235b priority adjustment for linux fixed Concedo 2023-05-03 10:16:43 +0800
  • 2485d7a4d3
    Process escape sequences given in prompts (#1173) master-2485d7a DannyDaemonic 2023-05-02 18:46:20 -0700
  • 3c249a4c74 Process escape sequences given in prompts Danny Daemonic 2023-04-25 07:39:08 -0700
  • 13b0c68ed7
    Handle signals properly on Windows (#1123) master-13b0c68 DannyDaemonic 2023-05-02 18:01:57 -0700
  • 55bc5f0900
    Call sh on build-info.sh (#1294) master-55bc5f0 DannyDaemonic 2023-05-02 17:52:35 -0700
  • 9daff419f6
    fix build-info.h for git submodules (#1289) master-9daff41 kuvaus 2023-05-03 03:43:43 +0300
  • 574f730cdb
    Merge branch 'ggerganov:master' into build_info kuvaus 2023-05-03 03:32:58 +0300
  • da9ec4c06a Call sh on build-info.sh Danny Daemonic 2023-05-02 17:00:30 -0700
  • 7cac587f94 fix missing parameters in llama_init_from_gpt_params (#1293) slaren 2023-05-03 01:36:45 +0200
  • 4123652cb6 examples : add llama_init_from_gpt_params() common function (#1290) Ron Evans 2023-05-02 22:39:51 +0200
  • 924df1a9a9 llama : fix compile warnings Georgi Gerganov 2023-05-02 23:09:08 +0300
  • 44d4f1ed0e ggml : fix 32-bit ARM Georgi Gerganov 2023-05-02 22:14:50 +0300
  • 3e7d5fcba0 examples : improve vertical alignment of a few variables (#1286) Ron Evans 2023-05-02 19:53:52 +0200
  • 704c4e8c69 ggml : fix ppc64le build error and make cmake detect Power processors (#1284) Marvin Gießing 2023-05-02 18:42:16 +0200
  • 30318b795d llama : allow 0 as a seed number. (#1275) Robert Brisita 2023-05-02 12:23:44 -0400
  • 806cbeae2b main : switch input_noecho to input_echo to remove negation (#979) Ron Evans 2023-05-02 18:13:26 +0200
  • bad9913de4 ggml: add names to tensors (#1268) slaren 2023-05-02 16:03:00 +0200
  • bf4b22ffe4
    fix missing parameters in llama_init_from_gpt_params (#1293) master-bf4b22f slaren 2023-05-03 01:36:45 +0200
  • f903cfab2a fix missing parameters in llama_init_from_gpt_params Slaren 2023-05-03 01:27:40 +0200
  • 7dffb0da73
    make git build info work with submodules Green Sky 2023-05-03 01:17:11 +0200
  • e06f9b86ec
    Remove custom command kuvaus 2023-05-03 01:03:36 +0300
  • 431f1c5630
    Require .git/ to be a folder for build-info.h kuvaus 2023-05-02 22:40:44 +0300
  • 53a87937cf Call sh on build script directly Danny Daemonic 2023-05-02 15:54:56 -0700
  • 4aa91a230a Improve handling of buffer Danny Daemonic 2023-05-02 15:38:00 -0700
  • fa82549c60 new/delete and fprintf Danny Daemonic 2023-05-02 08:48:36 -0700
  • 458df7413c avoid designated initializers Evan Jones 2023-05-02 17:43:50 -0400
  • 0bf20fe5f3 switch to ggml for copying k, v Evan Jones 2023-05-02 17:37:26 -0400
  • 970547e71a llama : only copy used KV cache in get / set state Evan Jones 2023-05-01 23:36:31 -0400
  • 67c77799e0
    examples : add llama_init_from_gpt_params() common function (#1290) master-67c7779 Ron Evans 2023-05-02 22:39:51 +0200
  • a032ac5eac examples : add llama_init_from_gpt_params() common function and modify all examples to use it, in order to reduce duplicate code deadprogram 2023-05-02 21:49:28 +0200
  • 0e6cbff1b7
    llama : fix compile warnings master-0e6cbff Georgi Gerganov 2023-05-02 23:09:08 +0300
  • 5d5817ca60
    ggml : fix 32-bit ARM master-5d5817c Georgi Gerganov 2023-05-02 22:14:50 +0300
  • 8c9be35ff9
    examples : improve vertical alignment of a few variables (#1286) master-8c9be35 Ron Evans 2023-05-02 19:53:52 +0200
  • 6899e69ebc example : improve vertical alignment of a few variables deadprogram 2023-05-02 19:46:12 +0200
  • cc0bb7235c
    ggml : fix ppc64le build error and make cmake detect Power processors (#1284) master-cc0bb72 Marvin Gießing 2023-05-02 18:42:16 +0200
  • 2bb992f034
    llama : allow 0 as a seed number. (#1275) master-2bb992f Robert Brisita 2023-05-02 12:23:44 -0400
  • f3e212361d
    Added support to detect ppc64* processors Marvin Gießing 2023-05-02 18:13:41 +0200
  • e2cd506999
    main : switch input_noecho to input_echo to remove negation (#979) master-e2cd506 Ron Evans 2023-05-02 18:13:26 +0200
  • a063f2e4d3
    Fix ppc64le build issue Marvin Gießing 2023-05-02 18:02:42 +0200
  • 966cd2ce91 Merge remote-tracking branch 'temp/concedo' into concedo_experimental Concedo 2023-05-02 22:43:34 +0800
  • 58f25dce86 added flag to increase processs priority Concedo 2023-05-02 22:26:55 +0800
  • 9cdbbeeee2 Implement get_num_physical_cores() for Windows Danny Daemonic 2023-05-02 07:07:21 -0700